Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
622,324
19,621,403,909
IssuesEvent
2022-01-07 07:14:26
tgpethan/EUS
https://api.github.com/repos/tgpethan/EUS
closed
Migrate to using database.
enhancement High priority
Currently as it stands EUS stores every single image in a json file in it's module folder. This is pretty bad ***json is not a database*** and it shouldn't be used as such therefore support for mysql and sqlite should be implemented and a migration path for old installs should be provided.
1.0
Migrate to using database. - Currently as it stands EUS stores every single image in a json file in it's module folder. This is pretty bad ***json is not a database*** and it shouldn't be used as such therefore support for mysql and sqlite should be implemented and a migration path for old installs should be provided.
priority
migrate to using database currently as it stands eus stores every single image in a json file in it s module folder this is pretty bad json is not a database and it shouldn t be used as such therefore support for mysql and sqlite should be implemented and a migration path for old installs should be provided
1
652,196
21,525,055,478
IssuesEvent
2022-04-28 17:32:56
ooni/explorer
https://api.github.com/repos/ooni/explorer
closed
Help text for MAT
priority/high MAT
To enable use of the MAT, we can include the following copy under the "Help" section of the MAT: ``` # What is the MAT? OONI's Measurement Aggregation Toolkit (MAT) is a tool that enables you to generate your own custom charts based on **aggregate views of real-time OONI data** collected from around the world. OONI data consists of network measurements collected by [OONI Probe](https://ooni.org/install/) users around the world. These measurements contain information about various types of **internet censorship**, such as the [blocking of websites and apps](https://ooni.org/nettest/) around the world. # Who is the MAT for? The MAT was built for researchers, journalists, and human rights defenders interested in examining internet censorship around the world. # Why use the MAT? When examining cases of internet censorship, it's important to **look at many measurements at once** ("in aggregate") in order to answer key questions like the following: * Does the testing of a service (e.g. Facebook) present **signs of blocking every time that it is tested** in a country? This can be helpful for ruling out [false positives](https://ooni.org/support/faq/#what-are-false-positives). * What types of websites (e.g. human rights websites) are blocked in each country? * In which countries is a specific website (e.g. `bbc.com`) blocked? * How does the blocking of different apps (e.g. WhatsApp or Telegram) vary across countries? * How does the blocking of a service vary across countries and [ASNs](https://ooni.org/support/glossary/#asn)? * How does the blocking of a service change over time? When trying to answer questions like the above, we normally perform relevant data analysis (instead of inspecting measurements one by one). The MAT incorporates our data analysis techniques, enabling you to answer such questions without any data analysis skills, and with the click of a button! # How to use the MAT? Through the filters at the start of the page, select the parameters you care about in order to plot charts based on aggregate views of OONI data. The MAT includes the following filters: * **Countries:** Select a country through the drop-down menu (the "All Countries" option will show global coverage) * **Test Name:** Select an [OONI Probe test](https://ooni.org/nettest/) based on which you would like to get measurements (for example, select `Web Connectivity` to view the testing of websites) * **Domain:** Type the domain for the website you would like to get measurements (e.g. `twitter.com`) * **Website categories:** Select the [website category](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv) for which you would like to get measurements (e.g. `News Media` for news media websites) * **ASN:** Type the [ASN](https://ooni.org/support/glossary/#asn) of the network for which you would like to get measurements (e.g. `AS30722` for Vodafone Italia) * **Date range:** Select the date range of the measurements by adjusting the `Since` and `Until` filters * **X axis:** Select the values that you would like to appear on the horizontal axis of your chart * **Y axis:** Select the values that you would like to appear on the vertical axis of your chart Depending on what you would like to explore, adjust the MAT filters accordingly and click `Submit`. For example, if you would like to check the testing of BBC in all countries around the world: * Type `www.bbc.com` under `Domain` * Select `Countries` under the `Y axis` * Click `Submit` This will plot numerous charts based on the OONI Probe testing of `www.bbc.com` worldwide. # Interpreting MAT charts The MAT charts (and associated tables) include the following values: * **OK count:** Successful measurements (i.e. NO sign of internet censorship) * **Confirmed count:** Measurements from automatically **confirmed blocked websites** (e.g. a [block page](https://ooni.org/support/glossary/#block-page) was served) * **Anomaly count:** Measurements that provided **signs of potential blocking** (however, [false positives](https://ooni.org/support/faq/#what-are-false-positives) can occur) * **Failure count:** Failed experiments that should be discarded * **Measurement count:** Total volume of OONI measurements (pertaining to the selected country, resource, etc.) When trying to identify the blocking of a service (e.g. `twitter.com`), it's useful to check whether: * Measurements are annotated as `confirmed`, automatically confirming the blocking of websites * A large volume of measurements (in comparison to the overall measurement count) present `anomalies` (i.e. signs of potential censorship) You can access the raw data by clicking on the bars of charts, and subsequently clicking on the relevant measurement links. # Website categories [OONI Probe](https://ooni.org/install/) users test a wide range of [websites](https://ooni.org/support/faq/#which-websites-will-i-test-for-censorship-with-ooni-probe) that fall under the following [30 standardized categories](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv). ```
1.0
Help text for MAT - To enable use of the MAT, we can include the following copy under the "Help" section of the MAT: ``` # What is the MAT? OONI's Measurement Aggregation Toolkit (MAT) is a tool that enables you to generate your own custom charts based on **aggregate views of real-time OONI data** collected from around the world. OONI data consists of network measurements collected by [OONI Probe](https://ooni.org/install/) users around the world. These measurements contain information about various types of **internet censorship**, such as the [blocking of websites and apps](https://ooni.org/nettest/) around the world. # Who is the MAT for? The MAT was built for researchers, journalists, and human rights defenders interested in examining internet censorship around the world. # Why use the MAT? When examining cases of internet censorship, it's important to **look at many measurements at once** ("in aggregate") in order to answer key questions like the following: * Does the testing of a service (e.g. Facebook) present **signs of blocking every time that it is tested** in a country? This can be helpful for ruling out [false positives](https://ooni.org/support/faq/#what-are-false-positives). * What types of websites (e.g. human rights websites) are blocked in each country? * In which countries is a specific website (e.g. `bbc.com`) blocked? * How does the blocking of different apps (e.g. WhatsApp or Telegram) vary across countries? * How does the blocking of a service vary across countries and [ASNs](https://ooni.org/support/glossary/#asn)? * How does the blocking of a service change over time? When trying to answer questions like the above, we normally perform relevant data analysis (instead of inspecting measurements one by one). The MAT incorporates our data analysis techniques, enabling you to answer such questions without any data analysis skills, and with the click of a button! # How to use the MAT? Through the filters at the start of the page, select the parameters you care about in order to plot charts based on aggregate views of OONI data. The MAT includes the following filters: * **Countries:** Select a country through the drop-down menu (the "All Countries" option will show global coverage) * **Test Name:** Select an [OONI Probe test](https://ooni.org/nettest/) based on which you would like to get measurements (for example, select `Web Connectivity` to view the testing of websites) * **Domain:** Type the domain for the website you would like to get measurements (e.g. `twitter.com`) * **Website categories:** Select the [website category](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv) for which you would like to get measurements (e.g. `News Media` for news media websites) * **ASN:** Type the [ASN](https://ooni.org/support/glossary/#asn) of the network for which you would like to get measurements (e.g. `AS30722` for Vodafone Italia) * **Date range:** Select the date range of the measurements by adjusting the `Since` and `Until` filters * **X axis:** Select the values that you would like to appear on the horizontal axis of your chart * **Y axis:** Select the values that you would like to appear on the vertical axis of your chart Depending on what you would like to explore, adjust the MAT filters accordingly and click `Submit`. For example, if you would like to check the testing of BBC in all countries around the world: * Type `www.bbc.com` under `Domain` * Select `Countries` under the `Y axis` * Click `Submit` This will plot numerous charts based on the OONI Probe testing of `www.bbc.com` worldwide. # Interpreting MAT charts The MAT charts (and associated tables) include the following values: * **OK count:** Successful measurements (i.e. NO sign of internet censorship) * **Confirmed count:** Measurements from automatically **confirmed blocked websites** (e.g. a [block page](https://ooni.org/support/glossary/#block-page) was served) * **Anomaly count:** Measurements that provided **signs of potential blocking** (however, [false positives](https://ooni.org/support/faq/#what-are-false-positives) can occur) * **Failure count:** Failed experiments that should be discarded * **Measurement count:** Total volume of OONI measurements (pertaining to the selected country, resource, etc.) When trying to identify the blocking of a service (e.g. `twitter.com`), it's useful to check whether: * Measurements are annotated as `confirmed`, automatically confirming the blocking of websites * A large volume of measurements (in comparison to the overall measurement count) present `anomalies` (i.e. signs of potential censorship) You can access the raw data by clicking on the bars of charts, and subsequently clicking on the relevant measurement links. # Website categories [OONI Probe](https://ooni.org/install/) users test a wide range of [websites](https://ooni.org/support/faq/#which-websites-will-i-test-for-censorship-with-ooni-probe) that fall under the following [30 standardized categories](https://github.com/citizenlab/test-lists/blob/master/lists/00-LEGEND-new_category_codes.csv). ```
priority
help text for mat to enable use of the mat we can include the following copy under the help section of the mat what is the mat ooni s measurement aggregation toolkit mat is a tool that enables you to generate your own custom charts based on aggregate views of real time ooni data collected from around the world ooni data consists of network measurements collected by users around the world these measurements contain information about various types of internet censorship such as the around the world who is the mat for the mat was built for researchers journalists and human rights defenders interested in examining internet censorship around the world why use the mat when examining cases of internet censorship it s important to look at many measurements at once in aggregate in order to answer key questions like the following does the testing of a service e g facebook present signs of blocking every time that it is tested in a country this can be helpful for ruling out what types of websites e g human rights websites are blocked in each country in which countries is a specific website e g bbc com blocked how does the blocking of different apps e g whatsapp or telegram vary across countries how does the blocking of a service vary across countries and how does the blocking of a service change over time when trying to answer questions like the above we normally perform relevant data analysis instead of inspecting measurements one by one the mat incorporates our data analysis techniques enabling you to answer such questions without any data analysis skills and with the click of a button how to use the mat through the filters at the start of the page select the parameters you care about in order to plot charts based on aggregate views of ooni data the mat includes the following filters countries select a country through the drop down menu the all countries option will show global coverage test name select an based on which you would like to get measurements for example select web connectivity to view the testing of websites domain type the domain for the website you would like to get measurements e g twitter com website categories select the for which you would like to get measurements e g news media for news media websites asn type the of the network for which you would like to get measurements e g for vodafone italia date range select the date range of the measurements by adjusting the since and until filters x axis select the values that you would like to appear on the horizontal axis of your chart y axis select the values that you would like to appear on the vertical axis of your chart depending on what you would like to explore adjust the mat filters accordingly and click submit for example if you would like to check the testing of bbc in all countries around the world type under domain select countries under the y axis click submit this will plot numerous charts based on the ooni probe testing of worldwide interpreting mat charts the mat charts and associated tables include the following values ok count successful measurements i e no sign of internet censorship confirmed count measurements from automatically confirmed blocked websites e g a was served anomaly count measurements that provided signs of potential blocking however can occur failure count failed experiments that should be discarded measurement count total volume of ooni measurements pertaining to the selected country resource etc when trying to identify the blocking of a service e g twitter com it s useful to check whether measurements are annotated as confirmed automatically confirming the blocking of websites a large volume of measurements in comparison to the overall measurement count present anomalies i e signs of potential censorship you can access the raw data by clicking on the bars of charts and subsequently clicking on the relevant measurement links website categories users test a wide range of that fall under the following
1
518,736
15,033,774,767
IssuesEvent
2021-02-02 11:58:10
opentargets/platform
https://api.github.com/repos/opentargets/platform
closed
I would like to have an example for Reactome for the new JSON schema
Kind: Data Priority: High
Provide an example evidence from Reactome of how it looks with the current JSON schema and how it should look like when they use the new schema
1.0
I would like to have an example for Reactome for the new JSON schema - Provide an example evidence from Reactome of how it looks with the current JSON schema and how it should look like when they use the new schema
priority
i would like to have an example for reactome for the new json schema provide an example evidence from reactome of how it looks with the current json schema and how it should look like when they use the new schema
1
146,910
5,630,412,514
IssuesEvent
2017-04-05 12:12:48
CS2103JAN2017-T11-B2/main
https://api.github.com/repos/CS2103JAN2017-T11-B2/main
closed
V0.5rc Documentation
priority.high type.task
All .md files need to be updated, including UserGuide, DeveloperGuide, AboutUs, and README
1.0
V0.5rc Documentation - All .md files need to be updated, including UserGuide, DeveloperGuide, AboutUs, and README
priority
documentation all md files need to be updated including userguide developerguide aboutus and readme
1
470,009
13,529,607,776
IssuesEvent
2020-09-15 18:34:12
Kedyn/fusliez-notes
https://api.github.com/repos/Kedyn/fusliez-notes
closed
Add title attribute to h1 input so the user knows its editable
Priority: High Status: Pending Type: Maintenance
In my fork I had a title attribute on the h1 input that read "Click to edit". it's not obvious that it can be edited.
1.0
Add title attribute to h1 input so the user knows its editable - In my fork I had a title attribute on the h1 input that read "Click to edit". it's not obvious that it can be edited.
priority
add title attribute to input so the user knows its editable in my fork i had a title attribute on the input that read click to edit it s not obvious that it can be edited
1
803,779
29,189,021,261
IssuesEvent
2023-05-19 18:03:43
minio/docs
https://api.github.com/repos/minio/docs
opened
[RELEASE] MinIO RELEASE.2023-05-18T00-05-36Z doc changes
priority: high
**Summary** See https://github.com/minio/minio/releases/tag/RELEASE.2023-05-18T00-05-36Z for full changelog ** ToDo - [ ] Persistent Queue Store for system/audit logs - [PR 17121](https://github.com/minio/minio/pull/17121) - [ ] Max policy size of 2KiB for Service Account / STS policies (clarify w/ engineer) - [PR 17161](https://github.com/minio/minio/pull/17167) - [ ] Webhook usage metrics [PR 17179)(https://github.com/minio/minio/pull/17179) - [ ] healing updates parity based on current Storage Class - clarify w/ engineer [PR 17187](https://github.com/minio/minio/pull/17187) - [ ] Improved support for topology changes during decomm - clarify w/ engineer [PR 17221](https://github.com/minio/minio/pull/17221) **Additional context** Add any other context or screenshots about the feature request here.
1.0
[RELEASE] MinIO RELEASE.2023-05-18T00-05-36Z doc changes - **Summary** See https://github.com/minio/minio/releases/tag/RELEASE.2023-05-18T00-05-36Z for full changelog ** ToDo - [ ] Persistent Queue Store for system/audit logs - [PR 17121](https://github.com/minio/minio/pull/17121) - [ ] Max policy size of 2KiB for Service Account / STS policies (clarify w/ engineer) - [PR 17161](https://github.com/minio/minio/pull/17167) - [ ] Webhook usage metrics [PR 17179)(https://github.com/minio/minio/pull/17179) - [ ] healing updates parity based on current Storage Class - clarify w/ engineer [PR 17187](https://github.com/minio/minio/pull/17187) - [ ] Improved support for topology changes during decomm - clarify w/ engineer [PR 17221](https://github.com/minio/minio/pull/17221) **Additional context** Add any other context or screenshots about the feature request here.
priority
minio release doc changes summary see for full changelog todo persistent queue store for system audit logs max policy size of for service account sts policies clarify w engineer webhook usage metrics pr healing updates parity based on current storage class clarify w engineer improved support for topology changes during decomm clarify w engineer additional context add any other context or screenshots about the feature request here
1
751,708
26,254,527,449
IssuesEvent
2023-01-05 22:47:30
lambdaclass/cairo-rs
https://api.github.com/repos/lambdaclass/cairo-rs
closed
Abstract the representation of field elements
high-priority
We're currently using `BigInt` explicitly, which forces us to compute the `.mod_floor` of results after every operation, use costly divisions, etc. The first step to fix arithmetic performance issues is to isolate the field properties into a new type and let the rest of the code operate on fields in a more abstract way.
1.0
Abstract the representation of field elements - We're currently using `BigInt` explicitly, which forces us to compute the `.mod_floor` of results after every operation, use costly divisions, etc. The first step to fix arithmetic performance issues is to isolate the field properties into a new type and let the rest of the code operate on fields in a more abstract way.
priority
abstract the representation of field elements we re currently using bigint explicitly which forces us to compute the mod floor of results after every operation use costly divisions etc the first step to fix arithmetic performance issues is to isolate the field properties into a new type and let the rest of the code operate on fields in a more abstract way
1
479,386
13,795,650,187
IssuesEvent
2020-10-09 18:23:48
vanjarosoftware/Vanjaro.Platform
https://api.github.com/repos/vanjarosoftware/Vanjaro.Platform
closed
Add resource file in Authentication package
Area: Backend Priority: High Release: Minor
extract in file in below path DesktopModules\AuthenticationServices\Vanjaro\App_LocalResources
1.0
Add resource file in Authentication package - extract in file in below path DesktopModules\AuthenticationServices\Vanjaro\App_LocalResources
priority
add resource file in authentication package extract in file in below path desktopmodules authenticationservices vanjaro app localresources
1
565,991
16,777,734,309
IssuesEvent
2021-06-15 00:52:57
myConsciousness/twitter-bot-j
https://api.github.com/repos/myConsciousness/twitter-bot-j
opened
不要な差分シンボルの追加処理がある
Priority: high Problem: bug
# High Priotity Bug Report ## 1. Bug Details 数値が負数であった場合にその値を文字列に変換すると既に負数を表す「-」が付いている状態だが、 現在の処理の流れではこの負数にも負数のシンボルを付与してしまい「-」が2つ付いている状態で値が出力される。 ```java private String toReportCount(@NonNull final Difference difference) { return switch (difference.getDifferenceType()) { case NONE -> DifferenceSymbolUtils.toNoneString(difference.getValue()); case INCREASE -> DifferenceSymbolUtils.toIncreaseString(difference.getValue()); case DECREASE -> DifferenceSymbolUtils.toDecreaseString(difference.getValue()); }; } ``` ## 2. What you did caused that bug 運用時の出力結果で確認。 ## 3. How it should be 上記の処理で負数の場合にはシンボルを付与せずただ文字列として変換すべき。 ## 4. References
1.0
不要な差分シンボルの追加処理がある - # High Priotity Bug Report ## 1. Bug Details 数値が負数であった場合にその値を文字列に変換すると既に負数を表す「-」が付いている状態だが、 現在の処理の流れではこの負数にも負数のシンボルを付与してしまい「-」が2つ付いている状態で値が出力される。 ```java private String toReportCount(@NonNull final Difference difference) { return switch (difference.getDifferenceType()) { case NONE -> DifferenceSymbolUtils.toNoneString(difference.getValue()); case INCREASE -> DifferenceSymbolUtils.toIncreaseString(difference.getValue()); case DECREASE -> DifferenceSymbolUtils.toDecreaseString(difference.getValue()); }; } ``` ## 2. What you did caused that bug 運用時の出力結果で確認。 ## 3. How it should be 上記の処理で負数の場合にはシンボルを付与せずただ文字列として変換すべき。 ## 4. References
priority
不要な差分シンボルの追加処理がある high priotity bug report bug details 数値が負数であった場合にその値を文字列に変換すると既に負数を表す「 」が付いている状態だが、 現在の処理の流れではこの負数にも負数のシンボルを付与してしまい「 」 。 java private string toreportcount nonnull final difference difference return switch difference getdifferencetype case none differencesymbolutils tononestring difference getvalue case increase differencesymbolutils toincreasestring difference getvalue case decrease differencesymbolutils todecreasestring difference getvalue what you did caused that bug 運用時の出力結果で確認。 how it should be 上記の処理で負数の場合にはシンボルを付与せずただ文字列として変換すべき。 references
1
363,605
10,745,079,350
IssuesEvent
2019-10-30 08:11:59
execom-eu/hawaii
https://api.github.com/repos/execom-eu/hawaii
closed
Manually changing leave profile for employee crashes the application
fix frontend high priority
Steps to reproduce (must have HR_MANAGER role): - Go to administration and select Employees tab - Find your profile information and change leave profile - Save the change, and go to the dashboard Expected result: - New leave profile is set and the corresponding allowance values are set Actual result: - Application crashes with an error message (image below) ![Screenshot from 2019-10-23 13-35-11](https://user-images.githubusercontent.com/52447076/67389049-1a81e580-f59a-11e9-8b82-6a9269927099.png)
1.0
Manually changing leave profile for employee crashes the application - Steps to reproduce (must have HR_MANAGER role): - Go to administration and select Employees tab - Find your profile information and change leave profile - Save the change, and go to the dashboard Expected result: - New leave profile is set and the corresponding allowance values are set Actual result: - Application crashes with an error message (image below) ![Screenshot from 2019-10-23 13-35-11](https://user-images.githubusercontent.com/52447076/67389049-1a81e580-f59a-11e9-8b82-6a9269927099.png)
priority
manually changing leave profile for employee crashes the application steps to reproduce must have hr manager role go to administration and select employees tab find your profile information and change leave profile save the change and go to the dashboard expected result new leave profile is set and the corresponding allowance values are set actual result application crashes with an error message image below
1
811,732
30,297,940,309
IssuesEvent
2023-07-10 01:48:30
steedos/steedos-platform
https://api.github.com/repos/steedos/steedos-platform
closed
[Bug]: 导入数据,lookup字段的“关联失败时保存key”功能无效
bug done priority: High
### Description 底层代码问题,当关联失败时,判断数据条数的地方已经通过 .length 转换成数值了,后面的if里面又来了一次 .length 造成判断无效,全部进入最后的else ![image](https://github.com/steedos/steedos-platform/assets/25192306/ece66c63-0cd2-48d1-ac38-02e23f9c9ca2) ### Steps To Reproduce 重现步骤 1. 配置数据导入 2. 配置关联表字段的导入 3. 配置对应的 “关联失败时保存key” 为打勾状态 4. 保存 5. 到对应的对象上导入数据 ### Version 版本 所有版本
1.0
[Bug]: 导入数据,lookup字段的“关联失败时保存key”功能无效 - ### Description 底层代码问题,当关联失败时,判断数据条数的地方已经通过 .length 转换成数值了,后面的if里面又来了一次 .length 造成判断无效,全部进入最后的else ![image](https://github.com/steedos/steedos-platform/assets/25192306/ece66c63-0cd2-48d1-ac38-02e23f9c9ca2) ### Steps To Reproduce 重现步骤 1. 配置数据导入 2. 配置关联表字段的导入 3. 配置对应的 “关联失败时保存key” 为打勾状态 4. 保存 5. 到对应的对象上导入数据 ### Version 版本 所有版本
priority
导入数据,lookup字段的“关联失败时保存key”功能无效 description 底层代码问题,当关联失败时,判断数据条数的地方已经通过 length 转换成数值了,后面的if里面又来了一次 length 造成判断无效,全部进入最后的else steps to reproduce 重现步骤 配置数据导入 配置关联表字段的导入 配置对应的 “关联失败时保存key” 为打勾状态 保存 到对应的对象上导入数据 version 版本 所有版本
1
340,390
10,271,858,700
IssuesEvent
2019-08-23 15:02:23
storybookjs/storybook
https://api.github.com/repos/storybookjs/storybook
closed
Addon-docs: User #root styles breaks Docs tab
addon: docs bug high priority ui
### Problem User applied the following global style to the preview iframe: ```css #root { height: 100vh; display: flex; flex-direction: column; } ``` This breaks docs in the following way. When the user click on the `Docs` tab, Storybook applies the `hidden` attribute to `#root`, which triggers the following browser CSS: ```css [hidden] { display: none; } ``` However, the `#root` CSS is more specific, so the Story renders on top of the docs. ### Solution When we added the following CSS to the user's code, it fixed it. ```css #root[hidden] { display: none; } ``` We can add this to Storybook itself to avoid this issue for users that style `#root` (which is a completely reasonable thing to do). However, I'm not sure this is enough...
1.0
Addon-docs: User #root styles breaks Docs tab - ### Problem User applied the following global style to the preview iframe: ```css #root { height: 100vh; display: flex; flex-direction: column; } ``` This breaks docs in the following way. When the user click on the `Docs` tab, Storybook applies the `hidden` attribute to `#root`, which triggers the following browser CSS: ```css [hidden] { display: none; } ``` However, the `#root` CSS is more specific, so the Story renders on top of the docs. ### Solution When we added the following CSS to the user's code, it fixed it. ```css #root[hidden] { display: none; } ``` We can add this to Storybook itself to avoid this issue for users that style `#root` (which is a completely reasonable thing to do). However, I'm not sure this is enough...
priority
addon docs user root styles breaks docs tab problem user applied the following global style to the preview iframe css root height display flex flex direction column this breaks docs in the following way when the user click on the docs tab storybook applies the hidden attribute to root which triggers the following browser css css display none however the root css is more specific so the story renders on top of the docs solution when we added the following css to the user s code it fixed it css root display none we can add this to storybook itself to avoid this issue for users that style root which is a completely reasonable thing to do however i m not sure this is enough
1
430,587
12,463,497,772
IssuesEvent
2020-05-28 10:43:25
UTRS2/utrs
https://api.github.com/repos/UTRS2/utrs
closed
Unable to view appeals
Priority: High bug
Users w/ the global flag in users.wikis or maybe with multiple wikis are not seeing appeals at all. Current workaround: Manual override of DB field to "enwiki" until fixed as there are no global appeals yet. Done for these users so far, replace them when done. Xoasflux JJMC89 ST47 TonyBallioni TheSandDoctor
1.0
Unable to view appeals - Users w/ the global flag in users.wikis or maybe with multiple wikis are not seeing appeals at all. Current workaround: Manual override of DB field to "enwiki" until fixed as there are no global appeals yet. Done for these users so far, replace them when done. Xoasflux JJMC89 ST47 TonyBallioni TheSandDoctor
priority
unable to view appeals users w the global flag in users wikis or maybe with multiple wikis are not seeing appeals at all current workaround manual override of db field to enwiki until fixed as there are no global appeals yet done for these users so far replace them when done xoasflux tonyballioni thesanddoctor
1
745,862
26,004,398,832
IssuesEvent
2022-12-20 17:55:57
vyper-protocol/vyper-otc-ui
https://api.github.com/repos/vyper-protocol/vyper-otc-ui
closed
add support for featured
high priority
add a menu in the topbar with a dropdown with featured products. These can be different depending on the cluster. #364 also relevant for this
1.0
add support for featured - add a menu in the topbar with a dropdown with featured products. These can be different depending on the cluster. #364 also relevant for this
priority
add support for featured add a menu in the topbar with a dropdown with featured products these can be different depending on the cluster also relevant for this
1
517,500
15,014,950,055
IssuesEvent
2021-02-01 07:32:55
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
connect.garmin.com - site is not usable
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal
<!-- @browser: Firefox 86.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:86.0) Gecko/20100101 Firefox/86.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/66526 --> **URL**: https://connect.garmin.com/modern/ **Browser / Version**: Firefox 86.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Tab loads with the page but page remains blank <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/9f3fc4ee-7863-46dd-98e4-136ef8be4ac3.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210128185743</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/4457943e-bc80-48c5-a650-bcb7d7736133) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
connect.garmin.com - site is not usable - <!-- @browser: Firefox 86.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:86.0) Gecko/20100101 Firefox/86.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/66526 --> **URL**: https://connect.garmin.com/modern/ **Browser / Version**: Firefox 86.0 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Tab loads with the page but page remains blank <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/9f3fc4ee-7863-46dd-98e4-136ef8be4ac3.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210128185743</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/4457943e-bc80-48c5-a650-bcb7d7736133) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
connect garmin com site is not usable url browser version firefox operating system windows tested another browser no problem type site is not usable description page not loading correctly steps to reproduce tab loads with the page but page remains blank view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
805,896
29,736,130,298
IssuesEvent
2023-06-14 01:10:08
GSM-MSG/SMS-BackEnd
https://api.github.com/repos/GSM-MSG/SMS-BackEnd
closed
Spring Security에서 예외 핸들링 반대로 되어져 있음
1️⃣ Priority: High
### Describe 인증에 403이 뜨고, 인가에 401이 뜨게 잘못 핸들링 해서 수정하겠습니다. ### Additional _No response_
1.0
Spring Security에서 예외 핸들링 반대로 되어져 있음 - ### Describe 인증에 403이 뜨고, 인가에 401이 뜨게 잘못 핸들링 해서 수정하겠습니다. ### Additional _No response_
priority
spring security에서 예외 핸들링 반대로 되어져 있음 describe 인증에 뜨고 인가에 뜨게 잘못 핸들링 해서 수정하겠습니다 additional no response
1
387,720
11,467,130,334
IssuesEvent
2020-02-08 02:20:38
allenai/scholar-reader
https://api.github.com/repos/allenai/scholar-reader
closed
Fault-tolerant processing of one entity at a time
high-priority pipeline
One challenge with colorizing multiple entities at a time is cascading failures---an error in colorizing one entity may change the location of all after it, or cause an entire batch of entities not to be colorized. This task includes: - [x] Add option to colorizing commands to process entities one at a time - [x] Add option to full pipeline for processing entities one at a time - [x] Add visual validation to check for any black pixels in image diffs. If there are black pixels, then do not attempt to detect bounding boxes - [x] Documentation that conveys that running one entity at a time will result in much greater usage of storage Follow-up analysis includes (put this in a separate issue later): - [ ] Characterize the time it takes to process the 'typical' paper one entity at a time - [ ] Compare to the relative costs and benefits of reworking the TeX engine to handle colorizing without affecting paper layout - [ ] Characterize the number of entities that will be left out during one-at-a-time entity colorizing
1.0
Fault-tolerant processing of one entity at a time - One challenge with colorizing multiple entities at a time is cascading failures---an error in colorizing one entity may change the location of all after it, or cause an entire batch of entities not to be colorized. This task includes: - [x] Add option to colorizing commands to process entities one at a time - [x] Add option to full pipeline for processing entities one at a time - [x] Add visual validation to check for any black pixels in image diffs. If there are black pixels, then do not attempt to detect bounding boxes - [x] Documentation that conveys that running one entity at a time will result in much greater usage of storage Follow-up analysis includes (put this in a separate issue later): - [ ] Characterize the time it takes to process the 'typical' paper one entity at a time - [ ] Compare to the relative costs and benefits of reworking the TeX engine to handle colorizing without affecting paper layout - [ ] Characterize the number of entities that will be left out during one-at-a-time entity colorizing
priority
fault tolerant processing of one entity at a time one challenge with colorizing multiple entities at a time is cascading failures an error in colorizing one entity may change the location of all after it or cause an entire batch of entities not to be colorized this task includes add option to colorizing commands to process entities one at a time add option to full pipeline for processing entities one at a time add visual validation to check for any black pixels in image diffs if there are black pixels then do not attempt to detect bounding boxes documentation that conveys that running one entity at a time will result in much greater usage of storage follow up analysis includes put this in a separate issue later characterize the time it takes to process the typical paper one entity at a time compare to the relative costs and benefits of reworking the tex engine to handle colorizing without affecting paper layout characterize the number of entities that will be left out during one at a time entity colorizing
1
355,374
10,579,943,329
IssuesEvent
2019-10-08 04:50:44
CalNourish/ucbfpa-webapp
https://api.github.com/repos/CalNourish/ucbfpa-webapp
opened
Update Add/Edit UI
high priority 🔥
- [ ] Remove points - [ ] Remove images - [ ] Introduce some UI for easily adding and subtracting counts without having to math (and be able to edit the total amount directly somehow)
1.0
Update Add/Edit UI - - [ ] Remove points - [ ] Remove images - [ ] Introduce some UI for easily adding and subtracting counts without having to math (and be able to edit the total amount directly somehow)
priority
update add edit ui remove points remove images introduce some ui for easily adding and subtracting counts without having to math and be able to edit the total amount directly somehow
1
166,221
6,300,163,869
IssuesEvent
2017-07-21 02:21:05
minio/minio
https://api.github.com/repos/minio/minio
closed
[mint] Tests fail in Azure/GCS gateway mode
priority: high working as intended
## Expected Behavior Mint test cases should pass. ## Current Behavior When Mint is run against the `minio gateway azure` or `minio gateway gcs`, requests fail with server log as below ``` $ ./minio gateway azure Endpoint: http://192.168.86.129:9000 http://172.17.0.1:9000 http://172.18.0.1:9000 http://127.0.0.1:9000 ERRO[0007] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-10234","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=ac91ba1d601e7b2d5136e531a3111770db0603705d9f388aee05c36f53fc6f65"],"Aws-Sdk-Invocation-Id":["8ccba5715ed1842e5862c4ec36a94c8c"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()] ERRO[0007] {"method":"DELETE","reqURI":"/aws-sdk-php-bucket-10234/obj1","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=7f448c75c4f825a397624b1fc1836841dffd8ecbf402a474c1a28486c026b9ae"],"Aws-Sdk-Invocation-Id":["908b9e424ad67ce48c2d5cde6c2659ec"],"Aws-Sdk-Retry":["0/0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()] ``` ``` $ ./minio gateway gcs peak-essence-171622 *** Warning: Not Ready for Production *** ERRO[0025] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-43021","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=2297da513a5dabc3230d17adc6713c54a5af97f4f0573ee86c3486add17d36c9"],"Aws-Sdk-Invocation-Id":["3ac3bd8d67bb14aedf64ee263ae42d6d"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T102637Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()] ``` Note that region is set using `export MINIO_REGION="us-east-1"` ## Steps to Reproduce (for bugs) 1. Start Minio server in azure or gcs gateway mode. 2. Run Mint against the Minio server instance.
1.0
[mint] Tests fail in Azure/GCS gateway mode - ## Expected Behavior Mint test cases should pass. ## Current Behavior When Mint is run against the `minio gateway azure` or `minio gateway gcs`, requests fail with server log as below ``` $ ./minio gateway azure Endpoint: http://192.168.86.129:9000 http://172.17.0.1:9000 http://172.18.0.1:9000 http://127.0.0.1:9000 ERRO[0007] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-10234","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=ac91ba1d601e7b2d5136e531a3111770db0603705d9f388aee05c36f53fc6f65"],"Aws-Sdk-Invocation-Id":["8ccba5715ed1842e5862c4ec36a94c8c"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()] ERRO[0007] {"method":"DELETE","reqURI":"/aws-sdk-php-bucket-10234/obj1","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=7f448c75c4f825a397624b1fc1836841dffd8ecbf402a474c1a28486c026b9ae"],"Aws-Sdk-Invocation-Id":["908b9e424ad67ce48c2d5cde6c2659ec"],"Aws-Sdk-Retry":["0/0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T094407Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()] ``` ``` $ ./minio gateway gcs peak-essence-171622 *** Warning: Not Ready for Production *** ERRO[0025] {"method":"PUT","reqURI":"/aws-sdk-php-bucket-43021","header":{"Authorization":["AWS4-HMAC-SHA256 Credential=minio/20170720/us-east-1/s3/aws4_request, SignedHeaders=aws-sdk-invocation-id;aws-sdk-retry;host;x-amz-content-sha256;x-amz-date, Signature=2297da513a5dabc3230d17adc6713c54a5af97f4f0573ee86c3486add17d36c9"],"Aws-Sdk-Invocation-Id":["3ac3bd8d67bb14aedf64ee263ae42d6d"],"Aws-Sdk-Retry":["0/0"],"Content-Length":["0"],"Host":["127.0.0.1:9000"],"User-Agent":["aws-sdk-php/3.31.7 GuzzleHttp/6.2.1 curl/7.47.0 PHP/7.0.18-0ubuntu0.16.04.1"],"X-Amz-Content-Sha256":["e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855"],"X-Amz-Date":["20170720T102637Z"]}} cause=Signature does not match source=[auth-handler.go:122:checkRequestAuthType()] ``` Note that region is set using `export MINIO_REGION="us-east-1"` ## Steps to Reproduce (for bugs) 1. Start Minio server in azure or gcs gateway mode. 2. Run Mint against the Minio server instance.
priority
tests fail in azure gcs gateway mode expected behavior mint test cases should pass current behavior when mint is run against the minio gateway azure or minio gateway gcs requests fail with server log as below minio gateway azure endpoint erro method put requri aws sdk php bucket header authorization aws sdk invocation id aws sdk retry content length host user agent x amz content x amz date cause signature does not match source erro method delete requri aws sdk php bucket header authorization aws sdk invocation id aws sdk retry host user agent x amz content x amz date cause signature does not match source minio gateway gcs peak essence warning not ready for production erro method put requri aws sdk php bucket header authorization aws sdk invocation id aws sdk retry content length host user agent x amz content x amz date cause signature does not match source note that region is set using export minio region us east steps to reproduce for bugs start minio server in azure or gcs gateway mode run mint against the minio server instance
1
520,042
15,077,759,202
IssuesEvent
2021-02-05 07:34:09
wso2/cellery
https://api.github.com/repos/wso2/cellery
closed
Dry run mode should be implemented for running cellery component tests
Priority/High Resolution/Won’t Fix Type/Improvement
**Description:** After the CLI refactoring, the kubectl commands are run in a ballerina native function. Therefore the component tests are required to be run on dry run mode to `cellery run` command perform a kubectl dry run to stop communication to the api server. As a workaround for the moment, the exceptions thrown by kubectl apply command are skipped after checking.
1.0
Dry run mode should be implemented for running cellery component tests - **Description:** After the CLI refactoring, the kubectl commands are run in a ballerina native function. Therefore the component tests are required to be run on dry run mode to `cellery run` command perform a kubectl dry run to stop communication to the api server. As a workaround for the moment, the exceptions thrown by kubectl apply command are skipped after checking.
priority
dry run mode should be implemented for running cellery component tests description after the cli refactoring the kubectl commands are run in a ballerina native function therefore the component tests are required to be run on dry run mode to cellery run command perform a kubectl dry run to stop communication to the api server as a workaround for the moment the exceptions thrown by kubectl apply command are skipped after checking
1
362,260
10,724,513,707
IssuesEvent
2019-10-28 02:02:55
LuanKovacs/LittleMatchGirlGame
https://api.github.com/repos/LuanKovacs/LittleMatchGirlGame
closed
[UPDATE, NOW WITH CAMERA PROBLEMS] I've fallen over and can get back up
Priority: High bug
**Describe the bug** you can fall over if you die near the end of the game **To Reproduce** Steps to reproduce the behavior: 1. Go to the final area 2. die 3. if the spawn in falling stutters a little then the player can now fall over when walking **Expected behavior** you shouldn't be able to fall over like that, the player while running should stay upright **Additional context** Add any other context about the problem here.
1.0
[UPDATE, NOW WITH CAMERA PROBLEMS] I've fallen over and can get back up - **Describe the bug** you can fall over if you die near the end of the game **To Reproduce** Steps to reproduce the behavior: 1. Go to the final area 2. die 3. if the spawn in falling stutters a little then the player can now fall over when walking **Expected behavior** you shouldn't be able to fall over like that, the player while running should stay upright **Additional context** Add any other context about the problem here.
priority
i ve fallen over and can get back up describe the bug you can fall over if you die near the end of the game to reproduce steps to reproduce the behavior go to the final area die if the spawn in falling stutters a little then the player can now fall over when walking expected behavior you shouldn t be able to fall over like that the player while running should stay upright additional context add any other context about the problem here
1
192,217
6,847,705,995
IssuesEvent
2017-11-13 16:11:06
cceh/capitularia
https://api.github.com/repos/cceh/capitularia
closed
Implementierung des Ortsregisters in den Handschriftenfilter
High Priority
Das Ortsregister sollte zeitnah in den Handschriftenfilter implementiert werden.
1.0
Implementierung des Ortsregisters in den Handschriftenfilter - Das Ortsregister sollte zeitnah in den Handschriftenfilter implementiert werden.
priority
implementierung des ortsregisters in den handschriftenfilter das ortsregister sollte zeitnah in den handschriftenfilter implementiert werden
1
656,227
21,724,004,263
IssuesEvent
2022-05-11 05:21:29
ever-co/ever-gauzy
https://api.github.com/repos/ever-co/ever-gauzy
closed
Fix: Theme Card Scrollbars (Firefox)
priority: highest Firefox Chorme Safari
The theme "scrollbars" in Firefox so they will be "thin" like in Chrome. ![image (16)](https://user-images.githubusercontent.com/41804588/167560809-67c3c171-991b-490e-b611-bbaab3052512.png)
1.0
Fix: Theme Card Scrollbars (Firefox) - The theme "scrollbars" in Firefox so they will be "thin" like in Chrome. ![image (16)](https://user-images.githubusercontent.com/41804588/167560809-67c3c171-991b-490e-b611-bbaab3052512.png)
priority
fix theme card scrollbars firefox the theme scrollbars in firefox so they will be thin like in chrome
1
151,026
5,795,538,621
IssuesEvent
2017-05-02 17:20:06
JiscRDSS/rdss-canonical-data-model
https://api.github.com/repos/JiscRDSS/rdss-canonical-data-model
closed
UC87 Metadata fields
alpha priority:High use case
UC no.: 87 Theme: Metadata fields As a Data creator I want Widespread use of unique IDs for people and organisations So that I don't have to keep up with correct/canonical names changing all the time, particularly for organisations and individuals' email addresses Comments Identifiers are a key component of the MVP (and data model)
1.0
UC87 Metadata fields - UC no.: 87 Theme: Metadata fields As a Data creator I want Widespread use of unique IDs for people and organisations So that I don't have to keep up with correct/canonical names changing all the time, particularly for organisations and individuals' email addresses Comments Identifiers are a key component of the MVP (and data model)
priority
metadata fields uc no theme metadata fields as a data creator i want widespread use of unique ids for people and organisations so that i don t have to keep up with correct canonical names changing all the time particularly for organisations and individuals email addresses comments identifiers are a key component of the mvp and data model
1
432,023
12,488,173,186
IssuesEvent
2020-05-31 13:01:57
STAMACODING/RSA-App
https://api.github.com/repos/STAMACODING/RSA-App
opened
Switch to OpenJDK
high priority meeting relevant organization
Auf meiner Raspberry Pi 4B gibt es leider nur schweren Support für das kommerzielle Oracle JDK. Für die neuste Version (14) sogar gar keine. Der Standard ist [OpenJDK](https://openjdk.java.net/), eine Open-Source-Variante des JDKs. Fürs Programmieren würde das quasi keine Unterschiede machen. Es müsste nur jeder bei sich einrichten. Dafür könnte ich auch ein Tutorial schreiben. Am besten besprechen wir das ganze in einem **Meeting**.
1.0
Switch to OpenJDK - Auf meiner Raspberry Pi 4B gibt es leider nur schweren Support für das kommerzielle Oracle JDK. Für die neuste Version (14) sogar gar keine. Der Standard ist [OpenJDK](https://openjdk.java.net/), eine Open-Source-Variante des JDKs. Fürs Programmieren würde das quasi keine Unterschiede machen. Es müsste nur jeder bei sich einrichten. Dafür könnte ich auch ein Tutorial schreiben. Am besten besprechen wir das ganze in einem **Meeting**.
priority
switch to openjdk auf meiner raspberry pi gibt es leider nur schweren support für das kommerzielle oracle jdk für die neuste version sogar gar keine der standard ist eine open source variante des jdks fürs programmieren würde das quasi keine unterschiede machen es müsste nur jeder bei sich einrichten dafür könnte ich auch ein tutorial schreiben am besten besprechen wir das ganze in einem meeting
1
413,005
12,059,178,805
IssuesEvent
2020-04-15 18:48:16
tern-tools/tern
https://api.github.com/repos/tern-tools/tern
closed
Update SPDX format to include file level analysis
high-priority
**Description** Update the SPDX report format to include situations where there is file level data. Use the http://13.57.134.254/app/validate/ online tool to validate the generated SPDX document for various container images. **Background** This depends on https://github.com/vmware/tern/pull/582 to be merged. **Super Issues** #583
1.0
Update SPDX format to include file level analysis - **Description** Update the SPDX report format to include situations where there is file level data. Use the http://13.57.134.254/app/validate/ online tool to validate the generated SPDX document for various container images. **Background** This depends on https://github.com/vmware/tern/pull/582 to be merged. **Super Issues** #583
priority
update spdx format to include file level analysis description update the spdx report format to include situations where there is file level data use the online tool to validate the generated spdx document for various container images background this depends on to be merged super issues
1
296,020
9,103,469,990
IssuesEvent
2019-02-20 15:57:57
infor-design/website
https://api.github.com/repos/infor-design/website
closed
Source Sans fonts aren't included in ng7 build
for: dev priority: high
**Describe the bug** Oops. **To Reproduce** Navigate to site on a device that doesn't have Source Sans installed. **Expected behavior** Include using Google fonts as previously or you could include and serve the fonts from `ids-identity`.
1.0
Source Sans fonts aren't included in ng7 build - **Describe the bug** Oops. **To Reproduce** Navigate to site on a device that doesn't have Source Sans installed. **Expected behavior** Include using Google fonts as previously or you could include and serve the fonts from `ids-identity`.
priority
source sans fonts aren t included in build describe the bug oops to reproduce navigate to site on a device that doesn t have source sans installed expected behavior include using google fonts as previously or you could include and serve the fonts from ids identity
1
727,707
25,044,734,986
IssuesEvent
2022-11-05 04:50:12
WFP-VAM/prism-app
https://api.github.com/repos/WFP-VAM/prism-app
closed
PRISM reports module
enhancement help wanted idea priority:high
Most users of information accessible through PRISM will not frequently visit the application. In addition, more advanced users such as analysts and GIS officers produce outputs that are then shared to a wider audience. Recognizing this, PRISM requires the ability to 1) condense various data inputs into a single output for a single snapshot view of key risk and impact factors; 2) reduce the steps involved for a user to generate a shareable output to a wide audience. A reports feature in PRISM will produce a static view of the dashboard exposing multiple datasets through a map view complemented with tabular data and a chart. The report could be triggered through the ‘Export’ button, and a direct link will be accessible via a URL such as https://prism.org?view=report&date=2021-05-17&hazard=tropical-storm Google doc to comment on here: https://docs.google.com/document/d/1FaV2jp526Xa134j32iXPZ_U4lhxFanlr/edit?usp=sharing&ouid=105953411765103382631&rtpof=true&sd=true Mockups: ![image](https://user-images.githubusercontent.com/3343536/123584040-87994580-d795-11eb-9ac6-aadd140b8e82.png) ![image](https://user-images.githubusercontent.com/3343536/123584048-8b2ccc80-d795-11eb-9822-4f3d82e20252.png)
1.0
PRISM reports module - Most users of information accessible through PRISM will not frequently visit the application. In addition, more advanced users such as analysts and GIS officers produce outputs that are then shared to a wider audience. Recognizing this, PRISM requires the ability to 1) condense various data inputs into a single output for a single snapshot view of key risk and impact factors; 2) reduce the steps involved for a user to generate a shareable output to a wide audience. A reports feature in PRISM will produce a static view of the dashboard exposing multiple datasets through a map view complemented with tabular data and a chart. The report could be triggered through the ‘Export’ button, and a direct link will be accessible via a URL such as https://prism.org?view=report&date=2021-05-17&hazard=tropical-storm Google doc to comment on here: https://docs.google.com/document/d/1FaV2jp526Xa134j32iXPZ_U4lhxFanlr/edit?usp=sharing&ouid=105953411765103382631&rtpof=true&sd=true Mockups: ![image](https://user-images.githubusercontent.com/3343536/123584040-87994580-d795-11eb-9ac6-aadd140b8e82.png) ![image](https://user-images.githubusercontent.com/3343536/123584048-8b2ccc80-d795-11eb-9822-4f3d82e20252.png)
priority
prism reports module most users of information accessible through prism will not frequently visit the application in addition more advanced users such as analysts and gis officers produce outputs that are then shared to a wider audience recognizing this prism requires the ability to condense various data inputs into a single output for a single snapshot view of key risk and impact factors reduce the steps involved for a user to generate a shareable output to a wide audience a reports feature in prism will produce a static view of the dashboard exposing multiple datasets through a map view complemented with tabular data and a chart the report could be triggered through the ‘export’ button and a direct link will be accessible via a url such as google doc to comment on here mockups
1
527,434
15,342,640,821
IssuesEvent
2021-02-27 17:02:21
getting-things-gnome/gtg
https://api.github.com/repos/getting-things-gnome/gtg
opened
Tags picker popover is empty after the v2 file format branch merge
bug priority:high reproducible-in-git
With current master, with the "screenshots" dataset, if you create a new task, in the task editor, clicking the tags button will give you an empty popover rather than showing existing tags.
1.0
Tags picker popover is empty after the v2 file format branch merge - With current master, with the "screenshots" dataset, if you create a new task, in the task editor, clicking the tags button will give you an empty popover rather than showing existing tags.
priority
tags picker popover is empty after the file format branch merge with current master with the screenshots dataset if you create a new task in the task editor clicking the tags button will give you an empty popover rather than showing existing tags
1
135,793
5,258,857,791
IssuesEvent
2017-02-03 00:57:57
ucdavis/ipa-client-angular
https://api.github.com/repos/ucdavis/ipa-client-angular
closed
TeachingCallResponse report: suggested courses not displayed correctly
bug high priority
From email: If you look at faculty member "Hanti Kao Lin" in Philosophy - his interested courses are all listed as sabbaticals in the report.
1.0
TeachingCallResponse report: suggested courses not displayed correctly - From email: If you look at faculty member "Hanti Kao Lin" in Philosophy - his interested courses are all listed as sabbaticals in the report.
priority
teachingcallresponse report suggested courses not displayed correctly from email if you look at faculty member hanti kao lin in philosophy his interested courses are all listed as sabbaticals in the report
1
274,845
8,568,542,085
IssuesEvent
2018-11-10 22:37:32
giftdibs/giftdibs-browser
https://api.github.com/repos/giftdibs/giftdibs-browser
closed
Gift Detail > Gift delivered message
priority: high
Show message on gift detail (and `gift_delivered` notification) when someone delivers a gift, to let the owner of the gift mark it as received. Clean up the "Delivered by" section on the gift detail.
1.0
Gift Detail > Gift delivered message - Show message on gift detail (and `gift_delivered` notification) when someone delivers a gift, to let the owner of the gift mark it as received. Clean up the "Delivered by" section on the gift detail.
priority
gift detail gift delivered message show message on gift detail and gift delivered notification when someone delivers a gift to let the owner of the gift mark it as received clean up the delivered by section on the gift detail
1
597,806
18,172,502,063
IssuesEvent
2021-09-27 21:45:22
StatisticsNZ/simplevis
https://api.github.com/repos/StatisticsNZ/simplevis
closed
bar: x_var date labels are not working correctly
high priority
``` library(tidyverse) library(er.helpers) library(simplevis) setup_datalake_access() no2_nzta <- er.helpers::read_from_datalake( "air/2021/tidy/no2_nzta.RDS") sitecheck_data <- no2_nzta %>% select(site, "value" = concentration, month, year) %>% mutate(len = str_length(site)) %>% mutate(temp_id = as.character(substring(site, 1,6))) %>% group_by(temp_id) %>% filter(any(str_length(site) > 6)) %>% mutate(measurement_date = lubridate::my(paste0(month, year)) %>% lubridate::as_date()) %>% mutate(site = as.character(site)) p <- sitecheck_data %>% filter(temp_id == "AUC004") %>% simplevis::gg_bar_col(x_var = measurement_date, y_var = value, col_var = site, x_pretty_n = 10, x_labels = scales::date_format("%y")) p sitecheck_data %>% filter(temp_id == "AUC004") %>% ggplot(aes(x = measurement_date, y = value, fill = site)) + geom_col() plotly::ggplotly(p) ```
1.0
bar: x_var date labels are not working correctly - ``` library(tidyverse) library(er.helpers) library(simplevis) setup_datalake_access() no2_nzta <- er.helpers::read_from_datalake( "air/2021/tidy/no2_nzta.RDS") sitecheck_data <- no2_nzta %>% select(site, "value" = concentration, month, year) %>% mutate(len = str_length(site)) %>% mutate(temp_id = as.character(substring(site, 1,6))) %>% group_by(temp_id) %>% filter(any(str_length(site) > 6)) %>% mutate(measurement_date = lubridate::my(paste0(month, year)) %>% lubridate::as_date()) %>% mutate(site = as.character(site)) p <- sitecheck_data %>% filter(temp_id == "AUC004") %>% simplevis::gg_bar_col(x_var = measurement_date, y_var = value, col_var = site, x_pretty_n = 10, x_labels = scales::date_format("%y")) p sitecheck_data %>% filter(temp_id == "AUC004") %>% ggplot(aes(x = measurement_date, y = value, fill = site)) + geom_col() plotly::ggplotly(p) ```
priority
bar x var date labels are not working correctly library tidyverse library er helpers library simplevis setup datalake access nzta er helpers read from datalake air tidy nzta rds sitecheck data select site value concentration month year mutate len str length site mutate temp id as character substring site group by temp id filter any str length site mutate measurement date lubridate my month year lubridate as date mutate site as character site p filter temp id simplevis gg bar col x var measurement date y var value col var site x pretty n x labels scales date format y p sitecheck data filter temp id ggplot aes x measurement date y value fill site geom col plotly ggplotly p
1
478,897
13,787,839,858
IssuesEvent
2020-10-09 05:58:08
wso2/streaming-integrator
https://api.github.com/repos/wso2/streaming-integrator
opened
Improvement for Siddhi Aggregation process
Priority/High Severity/Major Type/Improvement
**Description:** We need to have a check before purging tables. Ex. when purging the "days" table, we need to check whether the "months" table aggregations have happened with the relevant data which is going to be purged. And if not, we need to log a warning or an error. So that if any error happens, there will be data in the tables without purging so after providing a fix, it will resume aggregate data from where it left off. This will result in no data loss. **Affected Product Version:** SI-1.1.0
1.0
Improvement for Siddhi Aggregation process - **Description:** We need to have a check before purging tables. Ex. when purging the "days" table, we need to check whether the "months" table aggregations have happened with the relevant data which is going to be purged. And if not, we need to log a warning or an error. So that if any error happens, there will be data in the tables without purging so after providing a fix, it will resume aggregate data from where it left off. This will result in no data loss. **Affected Product Version:** SI-1.1.0
priority
improvement for siddhi aggregation process description we need to have a check before purging tables ex when purging the days table we need to check whether the months table aggregations have happened with the relevant data which is going to be purged and if not we need to log a warning or an error so that if any error happens there will be data in the tables without purging so after providing a fix it will resume aggregate data from where it left off this will result in no data loss affected product version si
1
45,343
2,928,232,999
IssuesEvent
2015-06-27 00:55:54
EFForg/privacybadgerchrome
https://api.github.com/repos/EFForg/privacybadgerchrome
closed
One click whitelist is broken for youtube.com
bug High priority
It should pop up when a user tries to comment on youtube but it doesn't. We should check disqus as well.
1.0
One click whitelist is broken for youtube.com - It should pop up when a user tries to comment on youtube but it doesn't. We should check disqus as well.
priority
one click whitelist is broken for youtube com it should pop up when a user tries to comment on youtube but it doesn t we should check disqus as well
1
106,641
4,281,570,843
IssuesEvent
2016-07-15 03:57:55
fflewddur/archivo
https://api.github.com/repos/fflewddur/archivo
closed
Pixelization/macroblocking in archived videos
bug high priority
Using Windows 10 and PrivateInternetAccess VPN, PC is unable to find TiVo device. They are on the same network - disabling firewall for private connections didn't help. If I disable the VPN it works fine. Any ideas on how to use without disabling the VPN?
1.0
Pixelization/macroblocking in archived videos - Using Windows 10 and PrivateInternetAccess VPN, PC is unable to find TiVo device. They are on the same network - disabling firewall for private connections didn't help. If I disable the VPN it works fine. Any ideas on how to use without disabling the VPN?
priority
pixelization macroblocking in archived videos using windows and privateinternetaccess vpn pc is unable to find tivo device they are on the same network disabling firewall for private connections didn t help if i disable the vpn it works fine any ideas on how to use without disabling the vpn
1
624,964
19,714,774,121
IssuesEvent
2022-01-13 09:55:10
hermeznetwork/wallet-ui
https://api.github.com/repos/hermeznetwork/wallet-ui
closed
Token Swap/Implement Design Token selector
type: enhancement priority: high
- Needs API to query tokens - Needs to check valid swaps for second token - Needs way to get images for token ![image.png](https://images.zenhubusercontent.com/5f05d715b733e51358956513/ed0bc673-f646-4e16-bee0-d698b6bcc1ed)
1.0
Token Swap/Implement Design Token selector - - Needs API to query tokens - Needs to check valid swaps for second token - Needs way to get images for token ![image.png](https://images.zenhubusercontent.com/5f05d715b733e51358956513/ed0bc673-f646-4e16-bee0-d698b6bcc1ed)
priority
token swap implement design token selector needs api to query tokens needs to check valid swaps for second token needs way to get images for token
1
454,140
13,095,491,648
IssuesEvent
2020-08-03 14:12:24
firecracker-microvm/firecracker
https://api.github.com/repos/firecracker-microvm/firecracker
closed
InvalidOffset (virtio-block) error when resuming after loading a snapshot (with rootfs from firecracker-containerd)
Feature: Snapshotting Priority: High Quality: Bug
Hi, we started developing support for snapshot `pause/resume/create/load` inside our fork of firecracker-containerd (we can make the code public). While we had no problem with supporting `pause/resume/create-snapshot` methods we ran into an error inside Firecracker's virtio-block module with `load-snapshot->resume`. The workflow is startVM(boot) -> Pause -> Create-Snap -> Offload (kill the VM with SIGTERM) -> SnapshotLoad -> Resume The problem we face is at the Resume point where we get the following error from firecracker's log: ``` 2020-06-25T07:21:10.652265445 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Put request on "/snapshot/load" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\"}". 2020-06-25T07:21:10.664695532 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:21:10.665787599 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Patch request on "/vm" with body "{\"state\":\"Resumed\"}". 2020-06-25T07:21:10.665908820 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:21:14.799994834 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset) 2020-06-25T07:21:14.802628426 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset) ``` We tend to think that this can be related to the way firecracker-containerd manages block devices. The VM mounts block devices in 2 phases. First, the VM boots from a generic rootfs (with the containerd agent): ``` The API server received a Put request on "/drives/root_drive" with body "The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n". ``` Then the agent needs to mount a second block device that contains the container-specific data. To do so, containerd attaches another drive and patches the path to the drive twice: ``` The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n". <<..>> The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n". The request was executed successfully. Status code: 204 No Content. The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n". ``` Can the issue be connected to the way [PATCH drive works](https://github.com/firecracker-microvm/firecracker/blob/master/docs/api_requests/patch-block.md)? This drive is supposed to remain mounted into the restored guest. We would greatly appreciate comments and ideas on what could be the root cause from Firecracker and firecracker-containerd maintainers: for example, @acatangiu @kzys . Once we fix the issue, we would be happy to contribute our changes to firecracker-containerd upstream. Full workflow log (Firecracker's log): ``` Running Firecracker v0.21.0 2020-06-25T07:20:17.287742839 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.288144594 [anonymous-instance:INFO] The API server received a Put request on "/machine-config" with body "{\"cpu_template\":\"T2\",\"ht_enabled\":false,\"mem_size_mib\":512,\"vcpu_count\":1}\n". 2020-06-25T07:20:17.288291418 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.288521865 [anonymous-instance:INFO] The API server received a Get request on "/machine-config". 2020-06-25T07:20:17.288625819 [anonymous-instance:INFO] The request was executed successfully. Status code: 200 OK. 2020-06-25T07:20:17.288971127 [anonymous-instance:INFO] The API server received a Put request on "/boot-source" with body "{\"boot_args\":\"8250.nr_uarts=0 ip=190.128.0.2::190.128.0.1:255.192.0.0:::off::: systemd.log_color=false init=/sbin/overlay-init systemd.unit=firecracker.target quiet noapic nomodules ipv6.disable=1 ro panic=1 tsc=reliable reboot=k pci=off\",\"kernel_image_path\":\"/var/lib/firecracker-containerd/runtime/hello-vmlinux.bin\"}\n". 2020-06-25T07:20:17.289115444 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.289571124 [anonymous-instance:INFO] The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n". 2020-06-25T07:20:17.289732138 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.290046499 [anonymous-instance:INFO] The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n". 2020-06-25T07:20:17.290151860 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.290522553 [anonymous-instance:INFO] The API server received a Put request on "/network-interfaces/1" with body "{\"guest_mac\":\"02:FC:00:00:00:00\",\"host_dev_name\":\"fc-0-tap0\",\"iface_id\":\"1\"}\n". 2020-06-25T07:20:17.292898333 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.293236487 [anonymous-instance:INFO] The API server received a Put request on "/vsock" with body "{\"guest_cid\":0,\"uds_path\":\"firecracker.vsock\",\"vsock_id\":\"agent_api\"}\n". 2020-06-25T07:20:17.293472939 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.293754191 [anonymous-instance:INFO] The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n". 2020-06-25T07:20:17.305625745 [anonymous-instance:WARN] Could not add serial input event to epoll: Error during epoll call: Operation not permitted (os error 1) 2020-06-25T07:20:17.306309907 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:19.177877690 [anonymous-instance:INFO] The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n". 2020-06-25T07:20:19.178152684 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:20.104482947 [anonymous-instance:INFO] The API server received a Patch request on "/vm" with body "{\"state\":\"Paused\"}". 2020-06-25T07:20:20.104702379 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:20.106119166 [anonymous-instance:INFO] The API server received a Put request on "/snapshot/create" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\",\"snapshot_type\":\"Full\"}". 2020-06-25T07:20:20.649402863 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. ``` And the corresponding log from firecracker-containerd (containerd's log) ``` INFO[2020-06-25T07:20:17.275031352-04:00] creating new VM runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:17.275464729-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" shim_stream=stderr vmID=505 setupLogging setupMetrics ERRO[2020-06-25T07:20:17.288041835-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="VMM metrics disabled." shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.288909406-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate:T2 HtEnabled:0xc000582f43 MemSizeMib:0xc000582f38 VcpuCount:0xc000582f30}" shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289306935-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289343891-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289933498-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289952823-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0, slot MN2HE43UOVRDA, root false." shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.290347821-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.290377745-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching NIC fc-0-tap0 (hwaddr 02:FC:00:00:00:00) at index 1" shim_stream=stderr vmID=505 INFO[2020-06-25T07:20:17.306497334-04:00] calling agent runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:17.306529875-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " shim_stream=stderr vmID=505 DEBU[2020-06-25T07:20:18.004394761-04:00] garbage collected d=10.79516ms INFO[2020-06-25T07:20:18.410674791-04:00] successfully started the VM runtime=aws.firecracker vmID=505 DEBU[2020-06-25T07:20:18.411052023-04:00] event forwarded ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart DEBU[2020-06-25T07:20:18.440217583-04:00] prepare snapshot key=505 parent="sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc" DEBU[2020-06-25T07:20:18.440347865-04:00] prepare key=firecracker-containerd/17/505 parent="firecracker-containerd/16/sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc" DEBU[2020-06-25T07:20:18.440504004-04:00] creating snapshot device 'fc-dev-thinpool-snap-9' from 'fc-dev-thinpool-snap-8' DEBU[2020-06-25T07:20:18.795683399-04:00] event published ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare DEBU[2020-06-25T07:20:18.800390309-04:00] get snapshot mounts key=505 DEBU[2020-06-25T07:20:18.800469759-04:00] mounts key=firecracker-containerd/17/505 DEBU[2020-06-25T07:20:19.120516743-04:00] event published ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate DEBU[2020-06-25T07:20:19.138167895-04:00] get snapshot mounts key=505 DEBU[2020-06-25T07:20:19.138206395-04:00] mounts key=firecracker-containerd/17/505 time="2020-06-25T07:20:19.174115148-04:00" level=debug msg=StartShim runtime=aws.firecracker task_id=505 DEBU[2020-06-25T07:20:19.175656280-04:00] create VM request: VMID:"505" DEBU[2020-06-25T07:20:19.175707632-04:00] using namespace: firecracker-containerd time="2020-06-25T07:20:19.176188684-04:00" level=info msg="successfully started shim (git commit: 22b7cfceb4901b91c998a380df74d9b825614967)." runtime=aws.firecracker task_id=505 vmID=505 ERRO[2020-06-25T07:20:19.178386061-04:00] time="2020-06-25T07:20:19-04:00" level=info msg="PatchGuestDrive successful" shim_stream=stderr vmID=505 INFO[2020-06-25T07:20:19.330752483-04:00] successfully created task ExecID= TaskID=505 pid_in_vm=727 runtime=aws.firecracker vmID=505 DEBU[2020-06-25T07:20:19.331251986-04:00] event forwarded ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate DEBU[2020-06-25T07:20:19.337046160-04:00] garbage collected d=12.841932ms DEBU[2020-06-25T07:20:19.339260338-04:00] event forwarded ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart DEBU[2020-06-25T07:20:20.103637583-04:00] pause VM request: VMID:"505" DEBU[2020-06-25T07:20:20.105566980-04:00] make snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file" DEBU[2020-06-25T07:20:20.650066779-04:00] offload original request: VMID:"505" ERRO[2020-06-25T07:20:25.892253044-04:00] error waiting for exit ExecID= TaskID=505 error="ttrpc: closed" runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:25.892257191-04:00] error while forwarding events from VM agent error="ttrpc: closed" runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:25.892484333-04:00] error copying io ExecID= TaskID=505 error="reading from a closed fifo" runtime=aws.firecracker stream=stdin vmID=505 ERRO[2020-06-25T07:20:25.892580185-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=505 ERRO[2020-06-25T07:20:25.923729272-04:00] time="2020-06-25T07:20:25-04:00" level=warning msg="firecracker exited: signal: killed" shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:30.892486598-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=505 ERRO[2020-06-25T07:20:30.892525517-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=505 DEBU[2020-06-25T07:21:10.651353555-04:00] load snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file" DEBU[2020-06-25T07:21:10.665366928-04:00] resume VM request: VMID:"505" ``` Platform: Host: Ubuntu 18, v4.15. Guest: containerd's default kernel and rootfs
1.0
InvalidOffset (virtio-block) error when resuming after loading a snapshot (with rootfs from firecracker-containerd) - Hi, we started developing support for snapshot `pause/resume/create/load` inside our fork of firecracker-containerd (we can make the code public). While we had no problem with supporting `pause/resume/create-snapshot` methods we ran into an error inside Firecracker's virtio-block module with `load-snapshot->resume`. The workflow is startVM(boot) -> Pause -> Create-Snap -> Offload (kill the VM with SIGTERM) -> SnapshotLoad -> Resume The problem we face is at the Resume point where we get the following error from firecracker's log: ``` 2020-06-25T07:21:10.652265445 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Put request on "/snapshot/load" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\"}". 2020-06-25T07:21:10.664695532 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:21:10.665787599 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:124] The API server received a Patch request on "/vm" with body "{\"state\":\"Resumed\"}". 2020-06-25T07:21:10.665908820 [anonymous-instance:INFO:src/api_server/src/parsed_request.rs:89] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:21:14.799994834 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset) 2020-06-25T07:21:14.802628426 [anonymous-instance:ERROR:src/devices/src/virtio/block/device.rs:226] Failed to execute request: BadRequest(InvalidOffset) ``` We tend to think that this can be related to the way firecracker-containerd manages block devices. The VM mounts block devices in 2 phases. First, the VM boots from a generic rootfs (with the containerd agent): ``` The API server received a Put request on "/drives/root_drive" with body "The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n". ``` Then the agent needs to mount a second block device that contains the container-specific data. To do so, containerd attaches another drive and patches the path to the drive twice: ``` The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n". <<..>> The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n". The request was executed successfully. Status code: 204 No Content. The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n". ``` Can the issue be connected to the way [PATCH drive works](https://github.com/firecracker-microvm/firecracker/blob/master/docs/api_requests/patch-block.md)? This drive is supposed to remain mounted into the restored guest. We would greatly appreciate comments and ideas on what could be the root cause from Firecracker and firecracker-containerd maintainers: for example, @acatangiu @kzys . Once we fix the issue, we would be happy to contribute our changes to firecracker-containerd upstream. Full workflow log (Firecracker's log): ``` Running Firecracker v0.21.0 2020-06-25T07:20:17.287742839 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.288144594 [anonymous-instance:INFO] The API server received a Put request on "/machine-config" with body "{\"cpu_template\":\"T2\",\"ht_enabled\":false,\"mem_size_mib\":512,\"vcpu_count\":1}\n". 2020-06-25T07:20:17.288291418 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.288521865 [anonymous-instance:INFO] The API server received a Get request on "/machine-config". 2020-06-25T07:20:17.288625819 [anonymous-instance:INFO] The request was executed successfully. Status code: 200 OK. 2020-06-25T07:20:17.288971127 [anonymous-instance:INFO] The API server received a Put request on "/boot-source" with body "{\"boot_args\":\"8250.nr_uarts=0 ip=190.128.0.2::190.128.0.1:255.192.0.0:::off::: systemd.log_color=false init=/sbin/overlay-init systemd.unit=firecracker.target quiet noapic nomodules ipv6.disable=1 ro panic=1 tsc=reliable reboot=k pci=off\",\"kernel_image_path\":\"/var/lib/firecracker-containerd/runtime/hello-vmlinux.bin\"}\n". 2020-06-25T07:20:17.289115444 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.289571124 [anonymous-instance:INFO] The API server received a Put request on "/drives/root_drive" with body "{\"drive_id\":\"root_drive\",\"is_read_only\":true,\"is_root_device\":true,\"path_on_host\":\"/var/lib/firecracker-containerd/runtime/default-rootfs.img\"}\n". 2020-06-25T07:20:17.289732138 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.290046499 [anonymous-instance:INFO] The API server received a Put request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"is_read_only\":false,\"is_root_device\":false,\"path_on_host\":\"/var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0\"}\n". 2020-06-25T07:20:17.290151860 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.290522553 [anonymous-instance:INFO] The API server received a Put request on "/network-interfaces/1" with body "{\"guest_mac\":\"02:FC:00:00:00:00\",\"host_dev_name\":\"fc-0-tap0\",\"iface_id\":\"1\"}\n". 2020-06-25T07:20:17.292898333 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.293236487 [anonymous-instance:INFO] The API server received a Put request on "/vsock" with body "{\"guest_cid\":0,\"uds_path\":\"firecracker.vsock\",\"vsock_id\":\"agent_api\"}\n". 2020-06-25T07:20:17.293472939 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:17.293754191 [anonymous-instance:INFO] The API server received a Put request on "/actions" with body "{\"action_type\":\"InstanceStart\"}\n". 2020-06-25T07:20:17.305625745 [anonymous-instance:WARN] Could not add serial input event to epoll: Error during epoll call: Operation not permitted (os error 1) 2020-06-25T07:20:17.306309907 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:19.177877690 [anonymous-instance:INFO] The API server received a Patch request on "/drives/MN2HE43UOVRDA" with body "{\"drive_id\":\"MN2HE43UOVRDA\",\"path_on_host\":\"/dev/mapper/fc-dev-thinpool-snap-9\"}\n". 2020-06-25T07:20:19.178152684 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:20.104482947 [anonymous-instance:INFO] The API server received a Patch request on "/vm" with body "{\"state\":\"Paused\"}". 2020-06-25T07:20:20.104702379 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. 2020-06-25T07:20:20.106119166 [anonymous-instance:INFO] The API server received a Put request on "/snapshot/create" with body "{\"mem_file_path\":\"/tmp/mem_file\",\"snapshot_path\":\"/tmp/snapshot_file\",\"snapshot_type\":\"Full\"}". 2020-06-25T07:20:20.649402863 [anonymous-instance:INFO] The request was executed successfully. Status code: 204 No Content. ``` And the corresponding log from firecracker-containerd (containerd's log) ``` INFO[2020-06-25T07:20:17.275031352-04:00] creating new VM runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:17.275464729-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Called startVMM(), setting up a VMM on firecracker.sock" shim_stream=stderr vmID=505 setupLogging setupMetrics ERRO[2020-06-25T07:20:17.288041835-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="VMM metrics disabled." shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.288909406-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="refreshMachineConfiguration: [GET /machine-config][200] getMachineConfigurationOK &{CPUTemplate:T2 HtEnabled:0xc000582f43 MemSizeMib:0xc000582f38 VcpuCount:0xc000582f30}" shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289306935-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="PutGuestBootSource: [PUT /boot-source][204] putGuestBootSourceNoContent " shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289343891-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/runtime/default-rootfs.img, slot root_drive, root true." shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289933498-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/runtime/default-rootfs.img: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.289952823-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0, slot MN2HE43UOVRDA, root false." shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.290347821-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attached drive /var/lib/firecracker-containerd/shim-base/firecracker-containerd/505/ctrstub0: [PUT /drives/{drive_id}][204] putGuestDriveByIdNoContent " shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:17.290377745-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="Attaching NIC fc-0-tap0 (hwaddr 02:FC:00:00:00:00) at index 1" shim_stream=stderr vmID=505 INFO[2020-06-25T07:20:17.306497334-04:00] calling agent runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:17.306529875-04:00] time="2020-06-25T07:20:17-04:00" level=info msg="startInstance successful: [PUT /actions][204] createSyncActionNoContent " shim_stream=stderr vmID=505 DEBU[2020-06-25T07:20:18.004394761-04:00] garbage collected d=10.79516ms INFO[2020-06-25T07:20:18.410674791-04:00] successfully started the VM runtime=aws.firecracker vmID=505 DEBU[2020-06-25T07:20:18.411052023-04:00] event forwarded ns=firecracker-containerd topic=/firecracker-vm/start type=VMStart DEBU[2020-06-25T07:20:18.440217583-04:00] prepare snapshot key=505 parent="sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc" DEBU[2020-06-25T07:20:18.440347865-04:00] prepare key=firecracker-containerd/17/505 parent="firecracker-containerd/16/sha256:0065b13862ce3af1bf002db326cc2c0a1245165cf6ae7f22abaceae55e09f6fc" DEBU[2020-06-25T07:20:18.440504004-04:00] creating snapshot device 'fc-dev-thinpool-snap-9' from 'fc-dev-thinpool-snap-8' DEBU[2020-06-25T07:20:18.795683399-04:00] event published ns=firecracker-containerd topic=/snapshot/prepare type=containerd.events.SnapshotPrepare DEBU[2020-06-25T07:20:18.800390309-04:00] get snapshot mounts key=505 DEBU[2020-06-25T07:20:18.800469759-04:00] mounts key=firecracker-containerd/17/505 DEBU[2020-06-25T07:20:19.120516743-04:00] event published ns=firecracker-containerd topic=/containers/create type=containerd.events.ContainerCreate DEBU[2020-06-25T07:20:19.138167895-04:00] get snapshot mounts key=505 DEBU[2020-06-25T07:20:19.138206395-04:00] mounts key=firecracker-containerd/17/505 time="2020-06-25T07:20:19.174115148-04:00" level=debug msg=StartShim runtime=aws.firecracker task_id=505 DEBU[2020-06-25T07:20:19.175656280-04:00] create VM request: VMID:"505" DEBU[2020-06-25T07:20:19.175707632-04:00] using namespace: firecracker-containerd time="2020-06-25T07:20:19.176188684-04:00" level=info msg="successfully started shim (git commit: 22b7cfceb4901b91c998a380df74d9b825614967)." runtime=aws.firecracker task_id=505 vmID=505 ERRO[2020-06-25T07:20:19.178386061-04:00] time="2020-06-25T07:20:19-04:00" level=info msg="PatchGuestDrive successful" shim_stream=stderr vmID=505 INFO[2020-06-25T07:20:19.330752483-04:00] successfully created task ExecID= TaskID=505 pid_in_vm=727 runtime=aws.firecracker vmID=505 DEBU[2020-06-25T07:20:19.331251986-04:00] event forwarded ns=firecracker-containerd topic=/tasks/create type=containerd.events.TaskCreate DEBU[2020-06-25T07:20:19.337046160-04:00] garbage collected d=12.841932ms DEBU[2020-06-25T07:20:19.339260338-04:00] event forwarded ns=firecracker-containerd topic=/tasks/start type=containerd.events.TaskStart DEBU[2020-06-25T07:20:20.103637583-04:00] pause VM request: VMID:"505" DEBU[2020-06-25T07:20:20.105566980-04:00] make snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file" DEBU[2020-06-25T07:20:20.650066779-04:00] offload original request: VMID:"505" ERRO[2020-06-25T07:20:25.892253044-04:00] error waiting for exit ExecID= TaskID=505 error="ttrpc: closed" runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:25.892257191-04:00] error while forwarding events from VM agent error="ttrpc: closed" runtime=aws.firecracker vmID=505 ERRO[2020-06-25T07:20:25.892484333-04:00] error copying io ExecID= TaskID=505 error="reading from a closed fifo" runtime=aws.firecracker stream=stdin vmID=505 ERRO[2020-06-25T07:20:25.892580185-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdin vmID=505 ERRO[2020-06-25T07:20:25.923729272-04:00] time="2020-06-25T07:20:25-04:00" level=warning msg="firecracker exited: signal: killed" shim_stream=stderr vmID=505 ERRO[2020-06-25T07:20:30.892486598-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stderr vmID=505 ERRO[2020-06-25T07:20:30.892525517-04:00] error closing io stream ExecID= TaskID=505 error="1 error occurred:\n\t* close unix @->firecracker.vsock: use of closed network connection\n\n" runtime=aws.firecracker stream=stdout vmID=505 DEBU[2020-06-25T07:21:10.651353555-04:00] load snapshot request: VMID:"505" SnapshotFilePath:"/tmp/snapshot_file" MemFilePath:"/tmp/mem_file" DEBU[2020-06-25T07:21:10.665366928-04:00] resume VM request: VMID:"505" ``` Platform: Host: Ubuntu 18, v4.15. Guest: containerd's default kernel and rootfs
priority
invalidoffset virtio block error when resuming after loading a snapshot with rootfs from firecracker containerd hi we started developing support for snapshot pause resume create load inside our fork of firecracker containerd we can make the code public while we had no problem with supporting pause resume create snapshot methods we ran into an error inside firecracker s virtio block module with load snapshot resume the workflow is startvm boot pause create snap offload kill the vm with sigterm snapshotload resume the problem we face is at the resume point where we get the following error from firecracker s log the api server received a put request on snapshot load with body mem file path tmp mem file snapshot path tmp snapshot file the request was executed successfully status code no content the api server received a patch request on vm with body state resumed the request was executed successfully status code no content failed to execute request badrequest invalidoffset failed to execute request badrequest invalidoffset we tend to think that this can be related to the way firecracker containerd manages block devices the vm mounts block devices in phases first the vm boots from a generic rootfs with the containerd agent the api server received a put request on drives root drive with body the api server received a put request on drives root drive with body drive id root drive is read only true is root device true path on host var lib firecracker containerd runtime default rootfs img n then the agent needs to mount a second block device that contains the container specific data to do so containerd attaches another drive and patches the path to the drive twice the api server received a put request on drives with body drive id is read only false is root device false path on host var lib firecracker containerd shim base firecracker containerd n the api server received a put request on actions with body action type instancestart n the request was executed successfully status code no content the api server received a patch request on drives with body drive id path on host dev mapper fc dev thinpool snap n can the issue be connected to the way this drive is supposed to remain mounted into the restored guest we would greatly appreciate comments and ideas on what could be the root cause from firecracker and firecracker containerd maintainers for example acatangiu kzys once we fix the issue we would be happy to contribute our changes to firecracker containerd upstream full workflow log firecracker s log running firecracker the request was executed successfully status code no content the api server received a put request on machine config with body cpu template ht enabled false mem size mib vcpu count n the request was executed successfully status code no content the api server received a get request on machine config the request was executed successfully status code ok the api server received a put request on boot source with body boot args nr uarts ip off systemd log color false init sbin overlay init systemd unit firecracker target quiet noapic nomodules disable ro panic tsc reliable reboot k pci off kernel image path var lib firecracker containerd runtime hello vmlinux bin n the request was executed successfully status code no content the api server received a put request on drives root drive with body drive id root drive is read only true is root device true path on host var lib firecracker containerd runtime default rootfs img n the request was executed successfully status code no content the api server received a put request on drives with body drive id is read only false is root device false path on host var lib firecracker containerd shim base firecracker containerd n the request was executed successfully status code no content the api server received a put request on network interfaces with body guest mac fc host dev name fc iface id n the request was executed successfully status code no content the api server received a put request on vsock with body guest cid uds path firecracker vsock vsock id agent api n the request was executed successfully status code no content the api server received a put request on actions with body action type instancestart n could not add serial input event to epoll error during epoll call operation not permitted os error the request was executed successfully status code no content the api server received a patch request on drives with body drive id path on host dev mapper fc dev thinpool snap n the request was executed successfully status code no content the api server received a patch request on vm with body state paused the request was executed successfully status code no content the api server received a put request on snapshot create with body mem file path tmp mem file snapshot path tmp snapshot file snapshot type full the request was executed successfully status code no content and the corresponding log from firecracker containerd containerd s log info creating new vm runtime aws firecracker vmid erro time level info msg called startvmm setting up a vmm on firecracker sock shim stream stderr vmid setuplogging setupmetrics erro time level info msg vmm metrics disabled shim stream stderr vmid erro time level info msg refreshmachineconfiguration getmachineconfigurationok cputemplate htenabled memsizemib vcpucount shim stream stderr vmid erro time level info msg putguestbootsource putguestbootsourcenocontent shim stream stderr vmid erro time level info msg attaching drive var lib firecracker containerd runtime default rootfs img slot root drive root true shim stream stderr vmid erro time level info msg attached drive var lib firecracker containerd runtime default rootfs img putguestdrivebyidnocontent shim stream stderr vmid erro time level info msg attaching drive var lib firecracker containerd shim base firecracker containerd slot root false shim stream stderr vmid erro time level info msg attached drive var lib firecracker containerd shim base firecracker containerd putguestdrivebyidnocontent shim stream stderr vmid erro time level info msg attaching nic fc hwaddr fc at index shim stream stderr vmid info calling agent runtime aws firecracker vmid erro time level info msg startinstance successful createsyncactionnocontent shim stream stderr vmid debu garbage collected d info successfully started the vm runtime aws firecracker vmid debu event forwarded ns firecracker containerd topic firecracker vm start type vmstart debu prepare snapshot key parent debu prepare key firecracker containerd parent firecracker containerd debu creating snapshot device fc dev thinpool snap from fc dev thinpool snap debu event published ns firecracker containerd topic snapshot prepare type containerd events snapshotprepare debu get snapshot mounts key debu mounts key firecracker containerd debu event published ns firecracker containerd topic containers create type containerd events containercreate debu get snapshot mounts key debu mounts key firecracker containerd time level debug msg startshim runtime aws firecracker task id debu create vm request vmid debu using namespace firecracker containerd time level info msg successfully started shim git commit runtime aws firecracker task id vmid erro time level info msg patchguestdrive successful shim stream stderr vmid info successfully created task execid taskid pid in vm runtime aws firecracker vmid debu event forwarded ns firecracker containerd topic tasks create type containerd events taskcreate debu garbage collected d debu event forwarded ns firecracker containerd topic tasks start type containerd events taskstart debu pause vm request vmid debu make snapshot request vmid snapshotfilepath tmp snapshot file memfilepath tmp mem file debu offload original request vmid erro error waiting for exit execid taskid error ttrpc closed runtime aws firecracker vmid erro error while forwarding events from vm agent error ttrpc closed runtime aws firecracker vmid erro error copying io execid taskid error reading from a closed fifo runtime aws firecracker stream stdin vmid erro error closing io stream execid taskid error error occurred n t close unix firecracker vsock use of closed network connection n n runtime aws firecracker stream stdin vmid erro time level warning msg firecracker exited signal killed shim stream stderr vmid erro error closing io stream execid taskid error error occurred n t close unix firecracker vsock use of closed network connection n n runtime aws firecracker stream stderr vmid erro error closing io stream execid taskid error error occurred n t close unix firecracker vsock use of closed network connection n n runtime aws firecracker stream stdout vmid debu load snapshot request vmid snapshotfilepath tmp snapshot file memfilepath tmp mem file debu resume vm request vmid platform host ubuntu guest containerd s default kernel and rootfs
1
159,064
6,040,075,512
IssuesEvent
2017-06-10 10:34:14
zulip/zulip-electron
https://api.github.com/repos/zulip/zulip-electron
closed
Add desktop notifications support on Windows 7
help wanted Priority: High Type: Enhancement windows
According to electron documentation Notification api does not work on Windows7, it has to use Tray baloon api. Is it possible to implement baloon notifications ? It would most propably require work also on standard website.
1.0
Add desktop notifications support on Windows 7 - According to electron documentation Notification api does not work on Windows7, it has to use Tray baloon api. Is it possible to implement baloon notifications ? It would most propably require work also on standard website.
priority
add desktop notifications support on windows according to electron documentation notification api does not work on it has to use tray baloon api is it possible to implement baloon notifications it would most propably require work also on standard website
1
343,913
10,338,268,977
IssuesEvent
2019-09-03 16:31:15
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
[feature request] `select_index` for sparse tensors
feature high priority module: sparse todo triaged
## Issue description Sparse tensors do not currently support indexing (neither `gather`, nor `select_index`). This could be handy for e.g. embedding categorical features with non-contiguous vocabulary. Sidenote: a similar result can be achieved via bucketization (see #7284), without using sparse tensors. ## Code example ``` >>> import numpy as np >>> import torch >>> vocab = [102, 104, 2, 103, 0, 101, 3] >>> lookup_table = torch.sparse.IntTensor( ... torch.tensor(np.atleast_2d(vocab)), ... torch.arange(len(vocab), dtype=torch.int), ... torch.Size((max(vocab) + 1, ))) ... >>> lookup_table[102] Traceback (most recent call last): File "<stdin>", line 1, in <module> RuntimeError: Sparse tensors do not have strides. ``` ## System Info ``` PyTorch version: 0.5.0a0+3e785d5 Is debug build: No CUDA used to build PyTorch: None OS: Mac OSX 10.13.3 GCC version: Could not collect CMake version: version 3.9.4 Python version: 3.6 Is CUDA available: No CUDA runtime version: No CUDA GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA Versions of relevant libraries: [pip3] numpy (1.14.2) [pip3] torch (0.5.0a0+3e785d5) [pip3] torchvision (0.2.1) [conda] torch 0.4.0 <pip> [conda] torch 0.5.0a0+3e785d5 <pip> [conda] torchvision 0.2.1 <pip> ```
1.0
[feature request] `select_index` for sparse tensors - ## Issue description Sparse tensors do not currently support indexing (neither `gather`, nor `select_index`). This could be handy for e.g. embedding categorical features with non-contiguous vocabulary. Sidenote: a similar result can be achieved via bucketization (see #7284), without using sparse tensors. ## Code example ``` >>> import numpy as np >>> import torch >>> vocab = [102, 104, 2, 103, 0, 101, 3] >>> lookup_table = torch.sparse.IntTensor( ... torch.tensor(np.atleast_2d(vocab)), ... torch.arange(len(vocab), dtype=torch.int), ... torch.Size((max(vocab) + 1, ))) ... >>> lookup_table[102] Traceback (most recent call last): File "<stdin>", line 1, in <module> RuntimeError: Sparse tensors do not have strides. ``` ## System Info ``` PyTorch version: 0.5.0a0+3e785d5 Is debug build: No CUDA used to build PyTorch: None OS: Mac OSX 10.13.3 GCC version: Could not collect CMake version: version 3.9.4 Python version: 3.6 Is CUDA available: No CUDA runtime version: No CUDA GPU models and configuration: No CUDA Nvidia driver version: No CUDA cuDNN version: No CUDA Versions of relevant libraries: [pip3] numpy (1.14.2) [pip3] torch (0.5.0a0+3e785d5) [pip3] torchvision (0.2.1) [conda] torch 0.4.0 <pip> [conda] torch 0.5.0a0+3e785d5 <pip> [conda] torchvision 0.2.1 <pip> ```
priority
select index for sparse tensors issue description sparse tensors do not currently support indexing neither gather nor select index this could be handy for e g embedding categorical features with non contiguous vocabulary sidenote a similar result can be achieved via bucketization see without using sparse tensors code example import numpy as np import torch vocab lookup table torch sparse inttensor torch tensor np atleast vocab torch arange len vocab dtype torch int torch size max vocab lookup table traceback most recent call last file line in runtimeerror sparse tensors do not have strides system info pytorch version is debug build no cuda used to build pytorch none os mac osx gcc version could not collect cmake version version python version is cuda available no cuda runtime version no cuda gpu models and configuration no cuda nvidia driver version no cuda cudnn version no cuda versions of relevant libraries numpy torch torchvision torch torch torchvision
1
343,825
10,336,956,183
IssuesEvent
2019-09-03 13:59:29
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Cannot load multivariable netcdf files
Bug Feedback High Priority Raster Regression
Author Name: **saverio vicario** (saverio vicario) Original Redmine Issue: [21822](https://issues.qgis.org/issues/21822) Affected QGIS version: 3.4.6 Redmine category:rasters --- I upgraded QGis and I cannot load some netcdf anymore. All the file multivariable that I have cannot be read. The netcdf that I can load do not trigger the window were to choose the variable as before but directly load the single variable present in the files. When I try to load a multivariable netcdf Qgis report on the GUI that source is not valid and on terminal I am getting only warnings relative to the naming of dimensions " Warning 1: dimension #2 (x) is not a Longitude/X dimension. Warning 1: dimension #1 (y) is not a Latitude/Y dimension. Warning 1: dimension #0 (Year) is not a Time or Vertical dimension. " so no critical error. When I am using gdalinfo is all fine. when I am using gdal_translate to transform a single variable in geotiff is all fine and am getting a tiff file correctly read by qgis itself. --- - [Testy.nc](https://issues.qgis.org/attachments/download/14779/Testy.nc) (saverio vicario) --- Related issue(s): #29598 (relates) Redmine related issue(s): [21783](https://issues.qgis.org/issues/21783) ---
1.0
Cannot load multivariable netcdf files - Author Name: **saverio vicario** (saverio vicario) Original Redmine Issue: [21822](https://issues.qgis.org/issues/21822) Affected QGIS version: 3.4.6 Redmine category:rasters --- I upgraded QGis and I cannot load some netcdf anymore. All the file multivariable that I have cannot be read. The netcdf that I can load do not trigger the window were to choose the variable as before but directly load the single variable present in the files. When I try to load a multivariable netcdf Qgis report on the GUI that source is not valid and on terminal I am getting only warnings relative to the naming of dimensions " Warning 1: dimension #2 (x) is not a Longitude/X dimension. Warning 1: dimension #1 (y) is not a Latitude/Y dimension. Warning 1: dimension #0 (Year) is not a Time or Vertical dimension. " so no critical error. When I am using gdalinfo is all fine. when I am using gdal_translate to transform a single variable in geotiff is all fine and am getting a tiff file correctly read by qgis itself. --- - [Testy.nc](https://issues.qgis.org/attachments/download/14779/Testy.nc) (saverio vicario) --- Related issue(s): #29598 (relates) Redmine related issue(s): [21783](https://issues.qgis.org/issues/21783) ---
priority
cannot load multivariable netcdf files author name saverio vicario saverio vicario original redmine issue affected qgis version redmine category rasters i upgraded qgis and i cannot load some netcdf anymore all the file multivariable that i have cannot be read the netcdf that i can load do not trigger the window were to choose the variable as before but directly load the single variable present in the files when i try to load a multivariable netcdf qgis report on the gui that source is not valid and on terminal i am getting only warnings relative to the naming of dimensions warning dimension x is not a longitude x dimension warning dimension y is not a latitude y dimension warning dimension year is not a time or vertical dimension so no critical error when i am using gdalinfo is all fine when i am using gdal translate to transform a single variable in geotiff is all fine and am getting a tiff file correctly read by qgis itself saverio vicario related issue s relates redmine related issue s
1
263,120
8,273,715,368
IssuesEvent
2018-09-17 07:16:56
theblacksea/issues
https://api.github.com/repos/theblacksea/issues
closed
social media
High Priority
Social media sharing link needs to be fixed. With Facebook and Twitter, sharing doesn't work properly, it doesn't show the story, only shows the theblacksea.eu address. Please see the picture explaining ![screen shot 2017-07-27 at 14 09 49](https://user-images.githubusercontent.com/29794271/28669636-4f91b1be-72d5-11e7-99fe-e699a8c580fd.png)
1.0
social media - Social media sharing link needs to be fixed. With Facebook and Twitter, sharing doesn't work properly, it doesn't show the story, only shows the theblacksea.eu address. Please see the picture explaining ![screen shot 2017-07-27 at 14 09 49](https://user-images.githubusercontent.com/29794271/28669636-4f91b1be-72d5-11e7-99fe-e699a8c580fd.png)
priority
social media social media sharing link needs to be fixed with facebook and twitter sharing doesn t work properly it doesn t show the story only shows the theblacksea eu address please see the picture explaining
1
200,281
7,005,201,993
IssuesEvent
2017-12-19 00:37:50
python/mypy
https://api.github.com/repos/python/mypy
closed
Calling `callable` in an if-condition makes mypy think branch in unreachable
bug priority-0-high
In the following code, mypy thinking `callable(o)` always returns `False`. This might be related to #3603. ```py def f(o: object) -> None: if callable(o): # this branch is not typechecked o() 1 + 'boom' # no error from mypy o() f(lambda: print('hello')) ``` Here is the output from `mypy` and `python`. ``` $ mypy n.py $ python3 n.py hello Traceback (most recent call last): File "n.py", line 8, in <module> f(lambda: print('hello')) File "n.py", line 5, in f 1 + 'boom' TypeError: unsupported operand type(s) for +: 'int' and 'str' ```
1.0
Calling `callable` in an if-condition makes mypy think branch in unreachable - In the following code, mypy thinking `callable(o)` always returns `False`. This might be related to #3603. ```py def f(o: object) -> None: if callable(o): # this branch is not typechecked o() 1 + 'boom' # no error from mypy o() f(lambda: print('hello')) ``` Here is the output from `mypy` and `python`. ``` $ mypy n.py $ python3 n.py hello Traceback (most recent call last): File "n.py", line 8, in <module> f(lambda: print('hello')) File "n.py", line 5, in f 1 + 'boom' TypeError: unsupported operand type(s) for +: 'int' and 'str' ```
priority
calling callable in an if condition makes mypy think branch in unreachable in the following code mypy thinking callable o always returns false this might be related to py def f o object none if callable o this branch is not typechecked o boom no error from mypy o f lambda print hello here is the output from mypy and python mypy n py n py hello traceback most recent call last file n py line in f lambda print hello file n py line in f boom typeerror unsupported operand type s for int and str
1
601,855
18,437,684,507
IssuesEvent
2021-10-14 14:37:05
DXgovernance/dxvote
https://api.github.com/repos/DXgovernance/dxvote
closed
New proposal metadata page
Enhancement High Priority
**Blocked by #187** In the simplified UX flow for proposal creation we will have a first view after selecting proposal type be a metadata editor. For the scope of this issue this includes the ability to edit the title and description via a markdown editor with more features to be added later. This first implementation should be accessible upon clicking the "Contributor proposal" option. (Not CUSTOM) This page should be paramterised and reusable, multiple different types of proposal will use this to edit metadata before going on to different next pages and so at the very least proposalType must be passed through to know where to navigate the user afterwards. (For now the page should not lead anywhere) The screenshots and figma are meant more as wireframes than exact design specifications so some choice lies with the developer on how they should look, feel free to ask for feedback from everyone. Basic design should follow the screenshots having a lefthand sidebar for set inputs and a more freeform MD editor with preview on the right. <img width="905" alt="スクリーンショット 2021-09-28 9 38 04" src="https://user-images.githubusercontent.com/39137239/135053467-f745b6c8-f14d-456f-8a40-ebf905629c7b.png"> https://www.figma.com/file/N4e2YuSiv1SycIhmqhIXGB/DXgov?node-id=95%3A2
1.0
New proposal metadata page - **Blocked by #187** In the simplified UX flow for proposal creation we will have a first view after selecting proposal type be a metadata editor. For the scope of this issue this includes the ability to edit the title and description via a markdown editor with more features to be added later. This first implementation should be accessible upon clicking the "Contributor proposal" option. (Not CUSTOM) This page should be paramterised and reusable, multiple different types of proposal will use this to edit metadata before going on to different next pages and so at the very least proposalType must be passed through to know where to navigate the user afterwards. (For now the page should not lead anywhere) The screenshots and figma are meant more as wireframes than exact design specifications so some choice lies with the developer on how they should look, feel free to ask for feedback from everyone. Basic design should follow the screenshots having a lefthand sidebar for set inputs and a more freeform MD editor with preview on the right. <img width="905" alt="スクリーンショット 2021-09-28 9 38 04" src="https://user-images.githubusercontent.com/39137239/135053467-f745b6c8-f14d-456f-8a40-ebf905629c7b.png"> https://www.figma.com/file/N4e2YuSiv1SycIhmqhIXGB/DXgov?node-id=95%3A2
priority
new proposal metadata page blocked by in the simplified ux flow for proposal creation we will have a first view after selecting proposal type be a metadata editor for the scope of this issue this includes the ability to edit the title and description via a markdown editor with more features to be added later this first implementation should be accessible upon clicking the contributor proposal option not custom this page should be paramterised and reusable multiple different types of proposal will use this to edit metadata before going on to different next pages and so at the very least proposaltype must be passed through to know where to navigate the user afterwards for now the page should not lead anywhere the screenshots and figma are meant more as wireframes than exact design specifications so some choice lies with the developer on how they should look feel free to ask for feedback from everyone basic design should follow the screenshots having a lefthand sidebar for set inputs and a more freeform md editor with preview on the right img width alt スクリーンショット src
1
544,253
15,891,627,858
IssuesEvent
2021-04-10 20:02:28
wso2/streaming-integrator
https://api.github.com/repos/wso2/streaming-integrator
closed
SI Extension installer overwrites the jars in the <SI_HOME>/lib directory
Priority/Highest Severity/Major T7 Type/Bug
**Description:** SI Extension installer overwrites the jars in the <SI_HOME>/lib directory **Affected Product Version:** SI 1.1.0 **OS, DB, other environment details and versions:** N/A
1.0
SI Extension installer overwrites the jars in the <SI_HOME>/lib directory - **Description:** SI Extension installer overwrites the jars in the <SI_HOME>/lib directory **Affected Product Version:** SI 1.1.0 **OS, DB, other environment details and versions:** N/A
priority
si extension installer overwrites the jars in the lib directory description si extension installer overwrites the jars in the lib directory affected product version si os db other environment details and versions n a
1
269,403
8,435,428,556
IssuesEvent
2018-10-17 13:08:34
vasyoid/cremebrulee
https://api.github.com/repos/vasyoid/cremebrulee
opened
Ошибка в сортировке при отображении списка экскурсий
bug high importance high priority
При указании выборгского района для поиска экскурсии первый рекомендованный вариант с более низким рейтингом, чем следующие варианты.
1.0
Ошибка в сортировке при отображении списка экскурсий - При указании выборгского района для поиска экскурсии первый рекомендованный вариант с более низким рейтингом, чем следующие варианты.
priority
ошибка в сортировке при отображении списка экскурсий при указании выборгского района для поиска экскурсии первый рекомендованный вариант с более низким рейтингом чем следующие варианты
1
660,824
22,032,586,746
IssuesEvent
2022-05-28 04:21:53
sqlalchemy/sqlalchemy
https://api.github.com/repos/sqlalchemy/sqlalchemy
closed
no control of DDL sequences for indexes, FK constraints that are mutually dependent
bug duplicate high priority schema
**Migrated issue, originally created by Michael Bayer ([@zzzeek](https://github.com/zzzeek))** the whole system described at http://docs.sqlalchemy.org/en/rel_1_0/core/ddl.html#controlling-ddl-sequences only takes place for constraints that are rendered inline within the table definition via the _create_rule() callable checked in compiler.py. This callable is not consulted anywhere in ddl.py when it goes to create indexes, foreign keys, or for that matter sequences. e.g. note below they are all execute_if(postgresql) but the engine is MySQL, and they all get built ``` from sqlalchemy import * from sqlalchemy.orm import * from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() class A(Base): __tablename__ = 'a' id = Column(Integer, primary_key=True) x = Column(Integer) favorite_b = Column(Integer) class B(Base): __tablename__ = 'b' id = Column(Integer, primary_key=True) a_id = Column(Integer) from sqlalchemy.schema import CreateIndex, AddConstraint idx = Index('aidx', A.x) fk1 = ForeignKeyConstraint([B.a_id], [A.id], name='afk') fk2 = ForeignKeyConstraint([A.favorite_b], [B.id], name='bfk') from sqlalchemy import event event.listen( Base.metadata, 'after_create', CreateIndex(idx).execute_if(dialect='postgresql') ) event.listen( Base.metadata, 'after_create', AddConstraint(fk1).execute_if(dialect='postgresql') ) event.listen( Base.metadata, 'after_create', AddConstraint(fk2).execute_if(dialect='postgresql') ) e = create_engine("mysql://scott:tiger@localhost/test", echo=True) Base.metadata.drop_all(e) Base.metadata.create_all(e) ``` ---------------------------------------- Attachments: [3442.patch](../wiki/imported_issue_attachments/3442/3442.patch)
1.0
no control of DDL sequences for indexes, FK constraints that are mutually dependent - **Migrated issue, originally created by Michael Bayer ([@zzzeek](https://github.com/zzzeek))** the whole system described at http://docs.sqlalchemy.org/en/rel_1_0/core/ddl.html#controlling-ddl-sequences only takes place for constraints that are rendered inline within the table definition via the _create_rule() callable checked in compiler.py. This callable is not consulted anywhere in ddl.py when it goes to create indexes, foreign keys, or for that matter sequences. e.g. note below they are all execute_if(postgresql) but the engine is MySQL, and they all get built ``` from sqlalchemy import * from sqlalchemy.orm import * from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() class A(Base): __tablename__ = 'a' id = Column(Integer, primary_key=True) x = Column(Integer) favorite_b = Column(Integer) class B(Base): __tablename__ = 'b' id = Column(Integer, primary_key=True) a_id = Column(Integer) from sqlalchemy.schema import CreateIndex, AddConstraint idx = Index('aidx', A.x) fk1 = ForeignKeyConstraint([B.a_id], [A.id], name='afk') fk2 = ForeignKeyConstraint([A.favorite_b], [B.id], name='bfk') from sqlalchemy import event event.listen( Base.metadata, 'after_create', CreateIndex(idx).execute_if(dialect='postgresql') ) event.listen( Base.metadata, 'after_create', AddConstraint(fk1).execute_if(dialect='postgresql') ) event.listen( Base.metadata, 'after_create', AddConstraint(fk2).execute_if(dialect='postgresql') ) e = create_engine("mysql://scott:tiger@localhost/test", echo=True) Base.metadata.drop_all(e) Base.metadata.create_all(e) ``` ---------------------------------------- Attachments: [3442.patch](../wiki/imported_issue_attachments/3442/3442.patch)
priority
no control of ddl sequences for indexes fk constraints that are mutually dependent migrated issue originally created by michael bayer the whole system described at only takes place for constraints that are rendered inline within the table definition via the create rule callable checked in compiler py this callable is not consulted anywhere in ddl py when it goes to create indexes foreign keys or for that matter sequences e g note below they are all execute if postgresql but the engine is mysql and they all get built from sqlalchemy import from sqlalchemy orm import from sqlalchemy ext declarative import declarative base base declarative base class a base tablename a id column integer primary key true x column integer favorite b column integer class b base tablename b id column integer primary key true a id column integer from sqlalchemy schema import createindex addconstraint idx index aidx a x foreignkeyconstraint name afk foreignkeyconstraint name bfk from sqlalchemy import event event listen base metadata after create createindex idx execute if dialect postgresql event listen base metadata after create addconstraint execute if dialect postgresql event listen base metadata after create addconstraint execute if dialect postgresql e create engine mysql scott tiger localhost test echo true base metadata drop all e base metadata create all e attachments wiki imported issue attachments patch
1
606,048
18,753,941,060
IssuesEvent
2021-11-05 08:12:43
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Golang API
Type: Backlog Category: API Status: Accepted Priority: High
# Enhancement Request ## Description Implement official Go binding for OpenNebula. It should be based on https://github.com/OpenNebula/goca. ## Use case OpenNebula should have a new API for Golang. It should be used in a similar way as the Java API. ## Interface Changes New API. # Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Insert into release process - [ ] Documentation (including code examples) - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches - [ ] Close https://github.com/OpenNebula/goca and reference the new docs
1.0
Golang API - # Enhancement Request ## Description Implement official Go binding for OpenNebula. It should be based on https://github.com/OpenNebula/goca. ## Use case OpenNebula should have a new API for Golang. It should be used in a similar way as the Java API. ## Interface Changes New API. # Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Insert into release process - [ ] Documentation (including code examples) - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches - [ ] Close https://github.com/OpenNebula/goca and reference the new docs
priority
golang api enhancement request description implement official go binding for opennebula it should be based on use case opennebula should have a new api for golang it should be used in a similar way as the java api interface changes new api progress status branch created code committed to development branch testing qa insert into release process documentation including code examples release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches close and reference the new docs
1
442,863
12,751,964,282
IssuesEvent
2020-06-27 14:00:55
LewisGaul/minegauler
https://api.github.com/repos/LewisGaul/minegauler
closed
Changing Max mines per cell mid game
bug high-priority resolved
Starting a game with max mines per cell as 3 and then changing it to 1 makes you unable to flag multiple mines on a single cell
1.0
Changing Max mines per cell mid game - Starting a game with max mines per cell as 3 and then changing it to 1 makes you unable to flag multiple mines on a single cell
priority
changing max mines per cell mid game starting a game with max mines per cell as and then changing it to makes you unable to flag multiple mines on a single cell
1
595,452
18,067,330,395
IssuesEvent
2021-09-20 20:51:34
zapzombies/zap-community
https://api.github.com/repos/zapzombies/zap-community
opened
Mobs riding other mobs do not pathfind properly
bug feature-ai high-priority
**Describe the bug** Mobs, such as the r25 boss(pig), cannot walk up blocks. **Where did the bug occur** Hades' Dominion **To Reproduce** Spawn bosspig, see how it does not move up blocks **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** This happens because the pathfinder is using the topmost pig's bounding box for pathfinding. It should be using the lowest bounding box.
1.0
Mobs riding other mobs do not pathfind properly - **Describe the bug** Mobs, such as the r25 boss(pig), cannot walk up blocks. **Where did the bug occur** Hades' Dominion **To Reproduce** Spawn bosspig, see how it does not move up blocks **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Additional context** This happens because the pathfinder is using the topmost pig's bounding box for pathfinding. It should be using the lowest bounding box.
priority
mobs riding other mobs do not pathfind properly describe the bug mobs such as the boss pig cannot walk up blocks where did the bug occur hades dominion to reproduce spawn bosspig see how it does not move up blocks expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem additional context this happens because the pathfinder is using the topmost pig s bounding box for pathfinding it should be using the lowest bounding box
1
389,627
11,504,732,060
IssuesEvent
2020-02-13 00:09:20
fecgov/fec-cms
https://api.github.com/repos/fecgov/fec-cms
opened
Fix IE version filter to return current and unknown version when applied
Bug High priority
**What we're after:** When the current version filter is applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true and most_recent=NULL, or current and unknown versions. At this time we are only seeing most_recent=true. When the current version filter is not applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true, most_recent=false and most_recent=NULL. **Key:** - most_recent=true is current version - most_recent=false is past version - most_recent=NULL is unknown version ### Related issues - #3545 ### Completion criteria - [ ] Make sure that the version filter works according to the above description.
1.0
Fix IE version filter to return current and unknown version when applied - **What we're after:** When the current version filter is applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true and most_recent=NULL, or current and unknown versions. At this time we are only seeing most_recent=true. When the current version filter is not applied for raw (all) and processed (24- and 48-hour reports) we should see most_recent=true, most_recent=false and most_recent=NULL. **Key:** - most_recent=true is current version - most_recent=false is past version - most_recent=NULL is unknown version ### Related issues - #3545 ### Completion criteria - [ ] Make sure that the version filter works according to the above description.
priority
fix ie version filter to return current and unknown version when applied what we re after when the current version filter is applied for raw all and processed and hour reports we should see most recent true and most recent null or current and unknown versions at this time we are only seeing most recent true when the current version filter is not applied for raw all and processed and hour reports we should see most recent true most recent false and most recent null key most recent true is current version most recent false is past version most recent null is unknown version related issues completion criteria make sure that the version filter works according to the above description
1
78,272
3,509,547,465
IssuesEvent
2016-01-08 23:23:12
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Help Tavara Quest (BB #1043)
Category: Quests migrated Priority: High Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** Ezopko **Original Date:** 10.07.2015 10:38:11 GMT+0000 **Original Priority:** critical **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1043 <hr> Quest Tavara not Working.. http://www.wowhead.com/quest=9586/help-tavara https://github.com/TrinityCore/TrinityCore/issues/2744
1.0
Help Tavara Quest (BB #1043) - This issue was migrated from bitbucket. **Original Reporter:** Ezopko **Original Date:** 10.07.2015 10:38:11 GMT+0000 **Original Priority:** critical **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/1043 <hr> Quest Tavara not Working.. http://www.wowhead.com/quest=9586/help-tavara https://github.com/TrinityCore/TrinityCore/issues/2744
priority
help tavara quest bb this issue was migrated from bitbucket original reporter ezopko original date gmt original priority critical original type bug original state resolved direct link quest tavara not working
1
736,718
25,484,426,393
IssuesEvent
2022-11-26 07:09:32
xsuite/xsuite
https://api.github.com/repos/xsuite/xsuite
opened
When regenerating json for hllhc15 example test_match_tune_chroma fails
bug High priority
Old json was in the format with elements as list instead of elements as dict. Maybe a bug in Line.to_dict/Line.from_dict?
1.0
When regenerating json for hllhc15 example test_match_tune_chroma fails - Old json was in the format with elements as list instead of elements as dict. Maybe a bug in Line.to_dict/Line.from_dict?
priority
when regenerating json for example test match tune chroma fails old json was in the format with elements as list instead of elements as dict maybe a bug in line to dict line from dict
1
330,550
10,050,207,110
IssuesEvent
2019-07-20 08:59:25
WoWManiaUK/Blackwing-Lair
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
closed
[Zone] Stranglethorn Vale - Death
Confirmed Fixed Confirmed Fixed in Dev Priority-High zone 30 - 40
**Links:** https://www.wowhead.com/the-cape-of-stranglethorn from WoWHead or our Armory **What is happening:** died in stranglethorn vale and was portaled to the spirit healer in westfall **What should happen:** be portaled to stranglethorn vale spirit healer Coords : 44.01, 91.52
1.0
[Zone] Stranglethorn Vale - Death - **Links:** https://www.wowhead.com/the-cape-of-stranglethorn from WoWHead or our Armory **What is happening:** died in stranglethorn vale and was portaled to the spirit healer in westfall **What should happen:** be portaled to stranglethorn vale spirit healer Coords : 44.01, 91.52
priority
stranglethorn vale death links from wowhead or our armory what is happening died in stranglethorn vale and was portaled to the spirit healer in westfall what should happen be portaled to stranglethorn vale spirit healer coords
1
163,949
6,217,494,309
IssuesEvent
2017-07-08 15:30:01
cytoscape/cytoscape.js
https://api.github.com/repos/cytoscape/cytoscape.js
closed
Benchmark and improve collection creation performance
performance priority-1-high
New collections are created by many APIs that devs use. They are also used internally in places. It would benefit overall performance to make collection creation less expensive.
1.0
Benchmark and improve collection creation performance - New collections are created by many APIs that devs use. They are also used internally in places. It would benefit overall performance to make collection creation less expensive.
priority
benchmark and improve collection creation performance new collections are created by many apis that devs use they are also used internally in places it would benefit overall performance to make collection creation less expensive
1
502,746
14,566,183,977
IssuesEvent
2020-12-17 08:31:36
CatalogueOfLife/portal
https://api.github.com/repos/CatalogueOfLife/portal
closed
References are missing with synonyms on species page
bug data high priority
Feedback from Paul Schoolmeesters: >I regret that the original reference for a species is not used when there has been any modification, in which case the last reference is shown. The example Ammoecioides catulus (Balthasar, 1946) In 2019 reference is correct (in the book picture) Best wishes, Paul Schoolmeesters
1.0
References are missing with synonyms on species page - Feedback from Paul Schoolmeesters: >I regret that the original reference for a species is not used when there has been any modification, in which case the last reference is shown. The example Ammoecioides catulus (Balthasar, 1946) In 2019 reference is correct (in the book picture) Best wishes, Paul Schoolmeesters
priority
references are missing with synonyms on species page feedback from paul schoolmeesters i regret that the original reference for a species is not used when there has been any modification in which case the last reference is shown the example ammoecioides catulus balthasar in reference is correct in the book picture best wishes paul schoolmeesters
1
392,381
11,590,691,310
IssuesEvent
2020-02-24 07:34:28
StanfordHCI/bang
https://api.github.com/repos/StanfordHCI/bang
closed
Dynamic Team size (1 v n) Breaks the Manipulation Check
bug high priority
When the team size is dynamic, and you have team size = 1, you cannot choose any partners during the final 'manipulation check' because there are no other partners. <img width="1071" alt="Screen Shot 2020-02-11 at 11 02 21 PM" src="https://user-images.githubusercontent.com/28793641/74313299-492df280-4d28-11ea-8439-4ddb863c9091.png"> As a result, the partner-select is null. When the survey is submitted, this causes a crash. **Below is the dump of the error I got.** [reconvene-logic-issue-dev-break.txt](https://github.com/StanfordHCI/bang/files/4190452/reconvene-logic-issue-dev-break.txt) **Suggestion:** When asking the question about the same partners, check to make sure you are comparing between rounds that actually have partners.
1.0
Dynamic Team size (1 v n) Breaks the Manipulation Check - When the team size is dynamic, and you have team size = 1, you cannot choose any partners during the final 'manipulation check' because there are no other partners. <img width="1071" alt="Screen Shot 2020-02-11 at 11 02 21 PM" src="https://user-images.githubusercontent.com/28793641/74313299-492df280-4d28-11ea-8439-4ddb863c9091.png"> As a result, the partner-select is null. When the survey is submitted, this causes a crash. **Below is the dump of the error I got.** [reconvene-logic-issue-dev-break.txt](https://github.com/StanfordHCI/bang/files/4190452/reconvene-logic-issue-dev-break.txt) **Suggestion:** When asking the question about the same partners, check to make sure you are comparing between rounds that actually have partners.
priority
dynamic team size v n breaks the manipulation check when the team size is dynamic and you have team size you cannot choose any partners during the final manipulation check because there are no other partners img width alt screen shot at pm src as a result the partner select is null when the survey is submitted this causes a crash below is the dump of the error i got suggestion when asking the question about the same partners check to make sure you are comparing between rounds that actually have partners
1
538,492
15,769,961,700
IssuesEvent
2021-03-31 18:55:00
Sage-Bionetworks/dccvalidator
https://api.github.com/repos/Sage-Bionetworks/dccvalidator
closed
Update AD instructions to point to docs submission form
consortia-related high priority
The vignette for the AD instructions is [here](https://github.com/Sage-Bionetworks/dccvalidator/blob/master/inst/using-the-dccvalidator-app-amp-ad.Rmd). This needs to be updated to remove information about uploading via the documentation tab and, instead, pointing the user to the Synapse submission form [here](https://www.synapse.org/#!Synapse:syn25051271). @avanlinden and @pitviper6, due to this being an instruction set, I want to make sure it's correct for how you would like people to submit/validate data. I am a little out of the loop on the AD process other than knowing that docs are going to the submission form. Would you be able to give me the information you want in these two sections below? Alternatively, @avanlinden, this is an Rmarkdown file. If you would feel comfortable enough, feel free to start a branch with PR to update with changes. https://github.com/Sage-Bionetworks/dccvalidator/blob/b122c93a796d613cb6846fa908a1625d72f5fd9c/inst/using-the-dccvalidator-app-amp-ad.Rmd#L47-L67
1.0
Update AD instructions to point to docs submission form - The vignette for the AD instructions is [here](https://github.com/Sage-Bionetworks/dccvalidator/blob/master/inst/using-the-dccvalidator-app-amp-ad.Rmd). This needs to be updated to remove information about uploading via the documentation tab and, instead, pointing the user to the Synapse submission form [here](https://www.synapse.org/#!Synapse:syn25051271). @avanlinden and @pitviper6, due to this being an instruction set, I want to make sure it's correct for how you would like people to submit/validate data. I am a little out of the loop on the AD process other than knowing that docs are going to the submission form. Would you be able to give me the information you want in these two sections below? Alternatively, @avanlinden, this is an Rmarkdown file. If you would feel comfortable enough, feel free to start a branch with PR to update with changes. https://github.com/Sage-Bionetworks/dccvalidator/blob/b122c93a796d613cb6846fa908a1625d72f5fd9c/inst/using-the-dccvalidator-app-amp-ad.Rmd#L47-L67
priority
update ad instructions to point to docs submission form the vignette for the ad instructions is this needs to be updated to remove information about uploading via the documentation tab and instead pointing the user to the synapse submission form avanlinden and due to this being an instruction set i want to make sure it s correct for how you would like people to submit validate data i am a little out of the loop on the ad process other than knowing that docs are going to the submission form would you be able to give me the information you want in these two sections below alternatively avanlinden this is an rmarkdown file if you would feel comfortable enough feel free to start a branch with pr to update with changes
1
480,992
13,878,980,682
IssuesEvent
2020-10-17 12:22:03
sButtons/sbuttons
https://api.github.com/repos/sButtons/sbuttons
closed
Place "Snake Border" button in the correct alphabetical order under Animated
Hacktoberfest Priority: High bug good first issue help wanted up-for-grabs website
Place "Snake border" in the correct alphabetical order under Animated. Changes should be made only in `assets/js/buttons-examples.js`
1.0
Place "Snake Border" button in the correct alphabetical order under Animated - Place "Snake border" in the correct alphabetical order under Animated. Changes should be made only in `assets/js/buttons-examples.js`
priority
place snake border button in the correct alphabetical order under animated place snake border in the correct alphabetical order under animated changes should be made only in assets js buttons examples js
1
265,222
8,345,304,020
IssuesEvent
2018-10-01 01:36:07
lgou2w/ldk
https://api.github.com/repos/lgou2w/ldk
closed
0.1-rc
Enhancement Priority: Highest work in progress
### 0.0.1-beta1 ✔️ - [x] Common - [x] #2 ChatComponent - [x] #3 NBT - [x] #4 Document - [x] #5 TAG - [x] #6 I18N - [x] Bukkit - [x] #8 - [x] #9 - [x] ItemBuilder - [x] ldk-bukkit-common Optimization ### 0.0.1-beta2 ✔️ - [x] Cmd API - [x] Gui API - [x] Region API - [x] Depdend API - [x] #15 Compatibility - [x] ItemBuilder Advanced - [x] Optimization ### 0.0.1-beta3 ✔️ - [x] Hikari module - [x] Unility - [x] Optimization ### 0.1-rc1 ✔️ - [x] Particle - [x] Problem solved - [x] Optimization
1.0
0.1-rc - ### 0.0.1-beta1 ✔️ - [x] Common - [x] #2 ChatComponent - [x] #3 NBT - [x] #4 Document - [x] #5 TAG - [x] #6 I18N - [x] Bukkit - [x] #8 - [x] #9 - [x] ItemBuilder - [x] ldk-bukkit-common Optimization ### 0.0.1-beta2 ✔️ - [x] Cmd API - [x] Gui API - [x] Region API - [x] Depdend API - [x] #15 Compatibility - [x] ItemBuilder Advanced - [x] Optimization ### 0.0.1-beta3 ✔️ - [x] Hikari module - [x] Unility - [x] Optimization ### 0.1-rc1 ✔️ - [x] Particle - [x] Problem solved - [x] Optimization
priority
rc ✔️ common chatcomponent nbt document tag bukkit itembuilder ldk bukkit common optimization ✔️ cmd api gui api region api depdend api compatibility itembuilder advanced optimization ✔️ hikari module unility optimization ✔️ particle problem solved optimization
1
117,213
4,712,451,699
IssuesEvent
2016-10-14 16:48:21
genome/civic-client
https://api.github.com/repos/genome/civic-client
closed
Add funding acknowledgement section to about page
high priority reviewer feedback
We should add a section on the about page to gratefully acknowledge funding sources for CIViC. Maybe with their logos displayed as well (for NCI, NHGRI, NIH). Example acknowledgement text as a starting point: Acknowledgements and funding: First and foremost we are grateful to the community of curators, editors, domain experts, and users who make CIViC possible. Malachi Griffith was supported by the National Human Genome Research Institute (NHGRI) of the National Institutes of Health (NIH) under Award Number K99HG007940. Obi Lee Griffith was supported by the National Cancer Institute of the NIH under Award Number K22CA188163. This work was also supported by a grant to Richard K. Wilson from the NHGRI under Award Number U54HG003079. The CIViC project is also supported by the NCI under Award Number U01CA209936 to Obi Lee Griffith (with Malachi Griffith and Elaine R Mardis as co-PIs). The content of this website is solely the responsibility of the creators and does not necessarily represent the official views of the National Institutes of Health.
1.0
Add funding acknowledgement section to about page - We should add a section on the about page to gratefully acknowledge funding sources for CIViC. Maybe with their logos displayed as well (for NCI, NHGRI, NIH). Example acknowledgement text as a starting point: Acknowledgements and funding: First and foremost we are grateful to the community of curators, editors, domain experts, and users who make CIViC possible. Malachi Griffith was supported by the National Human Genome Research Institute (NHGRI) of the National Institutes of Health (NIH) under Award Number K99HG007940. Obi Lee Griffith was supported by the National Cancer Institute of the NIH under Award Number K22CA188163. This work was also supported by a grant to Richard K. Wilson from the NHGRI under Award Number U54HG003079. The CIViC project is also supported by the NCI under Award Number U01CA209936 to Obi Lee Griffith (with Malachi Griffith and Elaine R Mardis as co-PIs). The content of this website is solely the responsibility of the creators and does not necessarily represent the official views of the National Institutes of Health.
priority
add funding acknowledgement section to about page we should add a section on the about page to gratefully acknowledge funding sources for civic maybe with their logos displayed as well for nci nhgri nih example acknowledgement text as a starting point acknowledgements and funding first and foremost we are grateful to the community of curators editors domain experts and users who make civic possible malachi griffith was supported by the national human genome research institute nhgri of the national institutes of health nih under award number obi lee griffith was supported by the national cancer institute of the nih under award number this work was also supported by a grant to richard k wilson from the nhgri under award number the civic project is also supported by the nci under award number to obi lee griffith with malachi griffith and elaine r mardis as co pis the content of this website is solely the responsibility of the creators and does not necessarily represent the official views of the national institutes of health
1
261,223
8,228,136,381
IssuesEvent
2018-09-07 03:16:40
phetsims/perennial
https://api.github.com/repos/phetsims/perennial
closed
Refining the latest links for phet-io
dev:phet-io meeting:phet-io priority:2-high
@jonathanolson, @zepumph, and I discussed the possibility of changing the `/latest` links for phet-io to instead be `/latest/{{major.minor}}`. The reasoning is this will allow clients to automatically pick up non-breaking maintenance releases, but avoid potentially api-breaking minor/major releases. Here is an example of how the redirect structure would look in the .htaccess file: ``` RewriteEngine on RewriteBase /sims/chains/ RewriteRule latest/1.17(.*) 1.17.0$1 RewriteRule latest/1.16(.*) 1.16.4$1 ``` Here, the link `/sims/chains/latest/1.17` redirects to `/sims/chains/1.17.0` and `/sims/chains/latest/1.16` redirects to `/sims/chains/1.16.4`. This will be difficult to manage using our current process, where the .htaccess file is recreated on every production deploy based on the current version being deployed. We discussed three possible solutions. 1. The build-server scans the directory structure on every production deploy, determines all of the latest maintenance releases based on a parsing of the file system, and writes a new .htaccess file. 2. The build server edits the .htaccess file in place, using string parsing to determine if a line needs to be changed or a new line added. 3. We add a database backend to the build server which maintains a record of all phet-io deploys. On a production deploy, the build-server queries the database of all latest minor versions and rewrites the .htaccess from scratch. (3) is probably the most robust and scalable solution, but would require some time to figure out how to integrate a db.
1.0
Refining the latest links for phet-io - @jonathanolson, @zepumph, and I discussed the possibility of changing the `/latest` links for phet-io to instead be `/latest/{{major.minor}}`. The reasoning is this will allow clients to automatically pick up non-breaking maintenance releases, but avoid potentially api-breaking minor/major releases. Here is an example of how the redirect structure would look in the .htaccess file: ``` RewriteEngine on RewriteBase /sims/chains/ RewriteRule latest/1.17(.*) 1.17.0$1 RewriteRule latest/1.16(.*) 1.16.4$1 ``` Here, the link `/sims/chains/latest/1.17` redirects to `/sims/chains/1.17.0` and `/sims/chains/latest/1.16` redirects to `/sims/chains/1.16.4`. This will be difficult to manage using our current process, where the .htaccess file is recreated on every production deploy based on the current version being deployed. We discussed three possible solutions. 1. The build-server scans the directory structure on every production deploy, determines all of the latest maintenance releases based on a parsing of the file system, and writes a new .htaccess file. 2. The build server edits the .htaccess file in place, using string parsing to determine if a line needs to be changed or a new line added. 3. We add a database backend to the build server which maintains a record of all phet-io deploys. On a production deploy, the build-server queries the database of all latest minor versions and rewrites the .htaccess from scratch. (3) is probably the most robust and scalable solution, but would require some time to figure out how to integrate a db.
priority
refining the latest links for phet io jonathanolson zepumph and i discussed the possibility of changing the latest links for phet io to instead be latest major minor the reasoning is this will allow clients to automatically pick up non breaking maintenance releases but avoid potentially api breaking minor major releases here is an example of how the redirect structure would look in the htaccess file rewriteengine on rewritebase sims chains rewriterule latest rewriterule latest here the link sims chains latest redirects to sims chains and sims chains latest redirects to sims chains this will be difficult to manage using our current process where the htaccess file is recreated on every production deploy based on the current version being deployed we discussed three possible solutions the build server scans the directory structure on every production deploy determines all of the latest maintenance releases based on a parsing of the file system and writes a new htaccess file the build server edits the htaccess file in place using string parsing to determine if a line needs to be changed or a new line added we add a database backend to the build server which maintains a record of all phet io deploys on a production deploy the build server queries the database of all latest minor versions and rewrites the htaccess from scratch is probably the most robust and scalable solution but would require some time to figure out how to integrate a db
1
364,297
10,761,848,769
IssuesEvent
2019-10-31 21:46:54
seakers/ABSP-EOSS
https://api.github.com/repos/seakers/ABSP-EOSS
closed
planner: Add post-mortum list updates
High Priority bug
After an agent has died, it must be able to keep track of the list of winning vectors for all of the other alive agents so to not lead to discrepancies at the end of the simulation
1.0
planner: Add post-mortum list updates - After an agent has died, it must be able to keep track of the list of winning vectors for all of the other alive agents so to not lead to discrepancies at the end of the simulation
priority
planner add post mortum list updates after an agent has died it must be able to keep track of the list of winning vectors for all of the other alive agents so to not lead to discrepancies at the end of the simulation
1
445,623
12,833,895,185
IssuesEvent
2020-07-07 10:04:23
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[0.9.0 staging-1644] Civis: broken UI with set conditions deleting
Category: UI Priority: High
1. Create a law 2. In the IF section choose "Set of conditions" and then delete it ![bandicam 2020-07-07 15-00-31-775](https://user-images.githubusercontent.com/27898520/86765711-f6e8c780-c062-11ea-82ad-5877383bc194.jpg) 3. If section will be broken, PLUS won't ass any positions ![bandicam 2020-07-07 15-01-18-329](https://user-images.githubusercontent.com/27898520/86765750-0536e380-c063-11ea-9f98-373a318ced47.jpg) 4. After reopening law ui ![bandicam 2020-07-07 15-01-23-439](https://user-images.githubusercontent.com/27898520/86765771-0ec04b80-c063-11ea-9124-43e93e3e3776.jpg)
1.0
[0.9.0 staging-1644] Civis: broken UI with set conditions deleting - 1. Create a law 2. In the IF section choose "Set of conditions" and then delete it ![bandicam 2020-07-07 15-00-31-775](https://user-images.githubusercontent.com/27898520/86765711-f6e8c780-c062-11ea-82ad-5877383bc194.jpg) 3. If section will be broken, PLUS won't ass any positions ![bandicam 2020-07-07 15-01-18-329](https://user-images.githubusercontent.com/27898520/86765750-0536e380-c063-11ea-9f98-373a318ced47.jpg) 4. After reopening law ui ![bandicam 2020-07-07 15-01-23-439](https://user-images.githubusercontent.com/27898520/86765771-0ec04b80-c063-11ea-9124-43e93e3e3776.jpg)
priority
civis broken ui with set conditions deleting create a law in the if section choose set of conditions and then delete it if section will be broken plus won t ass any positions after reopening law ui
1
621,850
19,597,802,602
IssuesEvent
2022-01-05 20:10:12
DSpace/dspace-angular
https://api.github.com/repos/DSpace/dspace-angular
closed
Edit EPerson | Reset Password disabled
bug high priority authentication: general authentication: password component: administrative tools e/2
**Describe the bug** Logged on as a member of Administrator, on the Edit EPerson page, Reset Password is disabled. **To Reproduce** Steps to reproduce the behavior: 1. Log on as a member of Administrator. 2. Try to reset someone's password. **Expected behavior** A member of Administrator should be able to reset someone's password. **Related work** Link to any related tickets or PRs here.
1.0
Edit EPerson | Reset Password disabled - **Describe the bug** Logged on as a member of Administrator, on the Edit EPerson page, Reset Password is disabled. **To Reproduce** Steps to reproduce the behavior: 1. Log on as a member of Administrator. 2. Try to reset someone's password. **Expected behavior** A member of Administrator should be able to reset someone's password. **Related work** Link to any related tickets or PRs here.
priority
edit eperson reset password disabled describe the bug logged on as a member of administrator on the edit eperson page reset password is disabled to reproduce steps to reproduce the behavior log on as a member of administrator try to reset someone s password expected behavior a member of administrator should be able to reset someone s password related work link to any related tickets or prs here
1
381,912
11,297,700,876
IssuesEvent
2020-01-17 06:56:55
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Issue in editing an oauth application which is created using Application management Rest API
Affected/5.10.0-Alpha2 Priority/Highest Severity/Critical Type/Bug
I am using Application management Rest API to create oauth application in IS-5.10.0-alpha pack. Once I create an oauth app using Rest API and when I'm trying to edit this app it gives a blank page. ![image](https://user-images.githubusercontent.com/40532734/72313640-fe31a880-36b1-11ea-8f92-b635c859192f.png)
1.0
Issue in editing an oauth application which is created using Application management Rest API - I am using Application management Rest API to create oauth application in IS-5.10.0-alpha pack. Once I create an oauth app using Rest API and when I'm trying to edit this app it gives a blank page. ![image](https://user-images.githubusercontent.com/40532734/72313640-fe31a880-36b1-11ea-8f92-b635c859192f.png)
priority
issue in editing an oauth application which is created using application management rest api i am using application management rest api to create oauth application in is alpha pack once i create an oauth app using rest api and when i m trying to edit this app it gives a blank page
1
248,180
7,928,251,050
IssuesEvent
2018-07-06 10:55:28
checkorg1/test
https://api.github.com/repos/checkorg1/test
closed
Greek characters in Genes collection
Highest Priority
_From egarciarivera on 13-06-2018 00:32 IST_ deepcompute-agent the Genes collection source file contains several tokens with greek characters that are not appearing on the staging collection. Can we check what the error is here? Source file is here: https://www.dropbox.com/s/s8vuebqcpa8eo7h/Nferx_Genes.csv?dl=0 **Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/322]
1.0
Greek characters in Genes collection - _From egarciarivera on 13-06-2018 00:32 IST_ deepcompute-agent the Genes collection source file contains several tokens with greek characters that are not appearing on the staging collection. Can we check what the error is here? Source file is here: https://www.dropbox.com/s/s8vuebqcpa8eo7h/Nferx_Genes.csv?dl=0 **Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/322]
priority
greek characters in genes collection from egarciarivera on ist deepcompute agent the genes collection source file contains several tokens with greek characters that are not appearing on the staging collection can we check what the error is here source file is here copied from original issue
1
377,157
11,165,295,367
IssuesEvent
2019-12-27 08:52:33
VirtoCommerce/vc-platform
https://api.github.com/repos/VirtoCommerce/vc-platform
closed
Improved frontend search UI
Epic Feature Priority: High
**Purpose** Improve frontend search so that it will consider autocomplete, recommendation, empty results, spell checking, search pages and blogs **Requirements** - [ ] ? **Related issues** - [ ] #950 Storefront: Add relevant suggestion for search phrase **User stories** - [ ] ? **Acceptance criteria** - [ ] ? **Definition of Done** - [ ] ?
1.0
Improved frontend search UI - **Purpose** Improve frontend search so that it will consider autocomplete, recommendation, empty results, spell checking, search pages and blogs **Requirements** - [ ] ? **Related issues** - [ ] #950 Storefront: Add relevant suggestion for search phrase **User stories** - [ ] ? **Acceptance criteria** - [ ] ? **Definition of Done** - [ ] ?
priority
improved frontend search ui purpose improve frontend search so that it will consider autocomplete recommendation empty results spell checking search pages and blogs requirements related issues storefront add relevant suggestion for search phrase user stories acceptance criteria definition of done
1
325,109
9,917,071,214
IssuesEvent
2019-06-28 22:16:52
amgrubb/BloomingLeaf
https://api.github.com/repos/amgrubb/BloomingLeaf
closed
Deleted Elements remain in the List of Assignments
HIGH PRIORITY bug
I created a big model and ran some simulations. Then I deleted half the model elements and the deleted elements remained List of Assignments Table (see screenshot of model and View List of Assignments Window). Branch: Develop ![screen shot 2018-12-20 at 9 27 49 am](https://user-images.githubusercontent.com/6596937/50290560-0cbf7900-043a-11e9-8ceb-22fa2c8f548d.png)
1.0
Deleted Elements remain in the List of Assignments - I created a big model and ran some simulations. Then I deleted half the model elements and the deleted elements remained List of Assignments Table (see screenshot of model and View List of Assignments Window). Branch: Develop ![screen shot 2018-12-20 at 9 27 49 am](https://user-images.githubusercontent.com/6596937/50290560-0cbf7900-043a-11e9-8ceb-22fa2c8f548d.png)
priority
deleted elements remain in the list of assignments i created a big model and ran some simulations then i deleted half the model elements and the deleted elements remained list of assignments table see screenshot of model and view list of assignments window branch develop
1
816,551
30,602,469,033
IssuesEvent
2023-07-22 15:11:23
alex4401/mediawiki-extensions-DataMaps
https://api.github.com/repos/alex4401/mediawiki-extensions-DataMaps
closed
Change zoom range via data
1. feature 3. priority: high affects wiki.gg 1. refactoring
Replace zoom Leaflet parameters with a `zoomRange` or `zoom` property. Replace `disableZoom` with `zoom: false`. This will require a data migration bot on wiki.gg.
1.0
Change zoom range via data - Replace zoom Leaflet parameters with a `zoomRange` or `zoom` property. Replace `disableZoom` with `zoom: false`. This will require a data migration bot on wiki.gg.
priority
change zoom range via data replace zoom leaflet parameters with a zoomrange or zoom property replace disablezoom with zoom false this will require a data migration bot on wiki gg
1
220,487
7,360,332,313
IssuesEvent
2018-03-10 17:30:14
bounswe/bounswe2018group5
https://api.github.com/repos/bounswe/bounswe2018group5
opened
Revise Requirements
Effort: Medium Priority: High Status: Available Type: Wiki
Per Cihat's comment: > * I am just lazy to check the previous version of your requirements and my feedback. However, one thing just got my attention. For user requirements, it seems that you have grouped the requirements according to the "user types". It could be better if you cluster them according to the functionalities. > * One more comment: Instead of "the user shall be able to see", you can write them as system requirements, since seeing is not something like a user requirement :) It just sees it, if the system provides.
1.0
Revise Requirements - Per Cihat's comment: > * I am just lazy to check the previous version of your requirements and my feedback. However, one thing just got my attention. For user requirements, it seems that you have grouped the requirements according to the "user types". It could be better if you cluster them according to the functionalities. > * One more comment: Instead of "the user shall be able to see", you can write them as system requirements, since seeing is not something like a user requirement :) It just sees it, if the system provides.
priority
revise requirements per cihat s comment i am just lazy to check the previous version of your requirements and my feedback however one thing just got my attention for user requirements it seems that you have grouped the requirements according to the user types it could be better if you cluster them according to the functionalities one more comment instead of the user shall be able to see you can write them as system requirements since seeing is not something like a user requirement it just sees it if the system provides
1
145,069
5,558,139,994
IssuesEvent
2017-03-24 14:03:34
Valhalla-Gaming/Tracker
https://api.github.com/repos/Valhalla-Gaming/Tracker
closed
[Death Knight][Artifact] Gravitational Pull
Class-Death Knight Priority-High Type-Spell
**Describe the issue you're having**: Doesn't work at all **Explain how you expect it work**: When using death grip, the enemy should be slowed by 50% for 6 seconds. **Links to Wowhead, YouTube, etc**: http://www.wowhead.com/spell=191721/gravitational-pull
1.0
[Death Knight][Artifact] Gravitational Pull - **Describe the issue you're having**: Doesn't work at all **Explain how you expect it work**: When using death grip, the enemy should be slowed by 50% for 6 seconds. **Links to Wowhead, YouTube, etc**: http://www.wowhead.com/spell=191721/gravitational-pull
priority
gravitational pull describe the issue you re having doesn t work at all explain how you expect it work when using death grip the enemy should be slowed by for seconds links to wowhead youtube etc
1
91,983
3,864,031,990
IssuesEvent
2016-04-08 12:14:59
eXistSolutions/hsg-shell
https://api.github.com/repos/eXistSolutions/hsg-shell
closed
Propose & implement approach for secrets
help wanted high priority
where needed: - s3: used on http://localhost:8080/exist/apps/hsg-shell/historicaldocuments/ebooks and FRUS volume landing pages (the latter not yet implemented in new site) - twitter: used currently on homepage - tumblr: used currently on homepage - eXist user account info: used in our ant scripts for populating an empty database and uploading/deleting content via oXygen
1.0
Propose & implement approach for secrets - where needed: - s3: used on http://localhost:8080/exist/apps/hsg-shell/historicaldocuments/ebooks and FRUS volume landing pages (the latter not yet implemented in new site) - twitter: used currently on homepage - tumblr: used currently on homepage - eXist user account info: used in our ant scripts for populating an empty database and uploading/deleting content via oXygen
priority
propose implement approach for secrets where needed used on and frus volume landing pages the latter not yet implemented in new site twitter used currently on homepage tumblr used currently on homepage exist user account info used in our ant scripts for populating an empty database and uploading deleting content via oxygen
1
304,082
9,320,822,208
IssuesEvent
2019-03-27 01:00:55
cuappdev/ithaca-transit-backend
https://api.github.com/repos/cuappdev/ithaca-transit-backend
closed
Show delayed routes on route display page
Priority: High
- Right now, if it is past 11:15pm, we don't see the first bus route even though we should since its delayed ![Image from iOS](https://user-images.githubusercontent.com/26048121/54861112-55692580-4cfa-11e9-8589-0d18f208dae8.jpg)
1.0
Show delayed routes on route display page - - Right now, if it is past 11:15pm, we don't see the first bus route even though we should since its delayed ![Image from iOS](https://user-images.githubusercontent.com/26048121/54861112-55692580-4cfa-11e9-8589-0d18f208dae8.jpg)
priority
show delayed routes on route display page right now if it is past we don t see the first bus route even though we should since its delayed
1
504,979
14,625,257,019
IssuesEvent
2020-12-23 08:10:23
bounswe/bounswe2020group7
https://api.github.com/repos/bounswe/bounswe2020group7
closed
Android - Upload Profile Photo Implementation
Priority: High Status: In Review Subteam: Android Type: New Feature
A ui option must be provided to upload profile photo to user.
1.0
Android - Upload Profile Photo Implementation - A ui option must be provided to upload profile photo to user.
priority
android upload profile photo implementation a ui option must be provided to upload profile photo to user
1
461,976
13,239,182,225
IssuesEvent
2020-08-19 02:40:09
bazzinotti/snestracker
https://api.github.com/repos/bazzinotti/snestracker
closed
[PatternEditor] cursor bug when switching to pattern of less rows
bug closed-in-earlypaid high priority
To Reproduce: - Create one pattern (00) with default length $40. - Create another pattern (01), but this time set the length to much lower. - Now, while in pattern (01), click the cursor on a high numbered row near the bottom of the screen. The highlight bar should move there as normal. - Finally, switch back to pattern (00). The highlight bar stays floating in black space because pattern (00) doesn't have as many rows. Solution: Add some logic when switching rows. (Pardon the horrid scaling in the photos) ![Screen Shot 2020-06-06 at 9 13 13 AM](https://user-images.githubusercontent.com/2224787/83945077-20010880-a7d6-11ea-805c-d2b7c8c93df0.png) -> ![Screen Shot 2020-06-06 at 9 13 49 AM](https://user-images.githubusercontent.com/2224787/83945078-20010880-a7d6-11ea-853a-5f3fa0f078b1.png)
1.0
[PatternEditor] cursor bug when switching to pattern of less rows - To Reproduce: - Create one pattern (00) with default length $40. - Create another pattern (01), but this time set the length to much lower. - Now, while in pattern (01), click the cursor on a high numbered row near the bottom of the screen. The highlight bar should move there as normal. - Finally, switch back to pattern (00). The highlight bar stays floating in black space because pattern (00) doesn't have as many rows. Solution: Add some logic when switching rows. (Pardon the horrid scaling in the photos) ![Screen Shot 2020-06-06 at 9 13 13 AM](https://user-images.githubusercontent.com/2224787/83945077-20010880-a7d6-11ea-805c-d2b7c8c93df0.png) -> ![Screen Shot 2020-06-06 at 9 13 49 AM](https://user-images.githubusercontent.com/2224787/83945078-20010880-a7d6-11ea-853a-5f3fa0f078b1.png)
priority
cursor bug when switching to pattern of less rows to reproduce create one pattern with default length create another pattern but this time set the length to much lower now while in pattern click the cursor on a high numbered row near the bottom of the screen the highlight bar should move there as normal finally switch back to pattern the highlight bar stays floating in black space because pattern doesn t have as many rows solution add some logic when switching rows pardon the horrid scaling in the photos
1
511,305
14,857,988,571
IssuesEvent
2021-01-18 16:09:34
ooni/run
https://api.github.com/repos/ooni/run
closed
Extremely high page size of links shared
bug priority/high
Simple typical links like [this](https://run.ooni.io/nettest\?tn\=web_connectivity\&ta\=%7B%22urls%22%3A%5B%22https%3A%2F%2Ftwitter.com%22%5D%7D\&mv\=1.2.0) when opened are too heavy. Close to 15MB. ![image](https://user-images.githubusercontent.com/700829/104509979-7a5a1700-55b8-11eb-8281-2418b637eed1.png) Link to [webpage-size-checker-tool](https://www.seoptimer.com/web-page-size-check/run.ooni.io/nettest?tn=web_connectivity&ta=%7b%22urls%22%3a%5b%22https%3a%2f%2ftwitter.com%22%5d%7d&mv=1.2.0)
1.0
Extremely high page size of links shared - Simple typical links like [this](https://run.ooni.io/nettest\?tn\=web_connectivity\&ta\=%7B%22urls%22%3A%5B%22https%3A%2F%2Ftwitter.com%22%5D%7D\&mv\=1.2.0) when opened are too heavy. Close to 15MB. ![image](https://user-images.githubusercontent.com/700829/104509979-7a5a1700-55b8-11eb-8281-2418b637eed1.png) Link to [webpage-size-checker-tool](https://www.seoptimer.com/web-page-size-check/run.ooni.io/nettest?tn=web_connectivity&ta=%7b%22urls%22%3a%5b%22https%3a%2f%2ftwitter.com%22%5d%7d&mv=1.2.0)
priority
extremely high page size of links shared simple typical links like when opened are too heavy close to link to
1
618,396
19,440,711,616
IssuesEvent
2021-12-22 00:13:07
CityOfDetroit/bloom
https://api.github.com/repos/CityOfDetroit/bloom
closed
Add ADA Features Filter
size: 5 P0 epic: filtering M17 HighestPriority
Add filtering based on ADA accessibility features, on the building or on the listing. See #609 for the fields added to the database. Some notes: - We only have data on the building, not the specific unit, so that will need to be reflected in the frontend filter (see #591) - Should the filter be a checkbox, or dropdown that lets you select "some units in building have features" or "all units in building have features"?
1.0
Add ADA Features Filter - Add filtering based on ADA accessibility features, on the building or on the listing. See #609 for the fields added to the database. Some notes: - We only have data on the building, not the specific unit, so that will need to be reflected in the frontend filter (see #591) - Should the filter be a checkbox, or dropdown that lets you select "some units in building have features" or "all units in building have features"?
priority
add ada features filter add filtering based on ada accessibility features on the building or on the listing see for the fields added to the database some notes we only have data on the building not the specific unit so that will need to be reflected in the frontend filter see should the filter be a checkbox or dropdown that lets you select some units in building have features or all units in building have features
1
618,254
19,430,813,345
IssuesEvent
2021-12-21 11:43:48
turbot/steampipe-plugin-aws
https://api.github.com/repos/turbot/steampipe-plugin-aws
closed
Queries against aws_ebs_snapshot are not returning results for aws ap-southeast-2 region
bug priority:high
**Describe the bug** Queries against aws_ebs_snapshot are not returning results in the ap-southeast-2 region **Steampipe version (`steampipe -v`)** v0.10.0 **Plugin version (`steampipe plugin list`)** v0.41.0 **To reproduce** 1) Create a snapshot in ap-southeast-2 2) query `select title, region from aws_ebs_snapshot` **Expected behavior** Returns results from existing snapshots.
1.0
Queries against aws_ebs_snapshot are not returning results for aws ap-southeast-2 region - **Describe the bug** Queries against aws_ebs_snapshot are not returning results in the ap-southeast-2 region **Steampipe version (`steampipe -v`)** v0.10.0 **Plugin version (`steampipe plugin list`)** v0.41.0 **To reproduce** 1) Create a snapshot in ap-southeast-2 2) query `select title, region from aws_ebs_snapshot` **Expected behavior** Returns results from existing snapshots.
priority
queries against aws ebs snapshot are not returning results for aws ap southeast region describe the bug queries against aws ebs snapshot are not returning results in the ap southeast region steampipe version steampipe v plugin version steampipe plugin list to reproduce create a snapshot in ap southeast query select title region from aws ebs snapshot expected behavior returns results from existing snapshots
1
48,909
3,000,832,732
IssuesEvent
2015-07-24 06:33:55
jayway/powermock
https://api.github.com/repos/jayway/powermock
closed
Test methods are not executed if PowerMockRule declaration is present.
bug imported Milestone-Release1.4 Priority-High
_From [Yaroslav...@gmail.com](https://code.google.com/u/107516775877747812247/) on July 16, 2010 23:42:18_ If PowerMockRule declaration is present in a test class, test methods annotated with @Test are not executed. Example test below passes without any warnings, while one of test methods throws NPE: @PrepareForTest(Static.class) public class TestCase { @Rule public PowerMockRule rule = new PowerMockRule (); @Test public void testMethodThatThrowsNPE() { throw new NullPointerException(); } } If rule declaration is commented out, test fails as expected. Quick debug shown that Runnable created using "evaluate()" method of PowerMockStatement is never executed. It seems, like this Runnable is intended to be executed in "Object execute(Object instance, Method method, Object... arguments)" method of ClassloaderExecutor, but it does not, because there is a check based on return type, and return type is expected to be non-void (line 61). PowerMock Ver.: 1.3.9 (Mockito API, with dependencies) Java: 1.6.0_21-b06 Eclipse (Helios, Build id: 20100617-1415) plugin was used to run the tests. _Original issue: http://code.google.com/p/powermock/issues/detail?id=268_
1.0
Test methods are not executed if PowerMockRule declaration is present. - _From [Yaroslav...@gmail.com](https://code.google.com/u/107516775877747812247/) on July 16, 2010 23:42:18_ If PowerMockRule declaration is present in a test class, test methods annotated with @Test are not executed. Example test below passes without any warnings, while one of test methods throws NPE: @PrepareForTest(Static.class) public class TestCase { @Rule public PowerMockRule rule = new PowerMockRule (); @Test public void testMethodThatThrowsNPE() { throw new NullPointerException(); } } If rule declaration is commented out, test fails as expected. Quick debug shown that Runnable created using "evaluate()" method of PowerMockStatement is never executed. It seems, like this Runnable is intended to be executed in "Object execute(Object instance, Method method, Object... arguments)" method of ClassloaderExecutor, but it does not, because there is a check based on return type, and return type is expected to be non-void (line 61). PowerMock Ver.: 1.3.9 (Mockito API, with dependencies) Java: 1.6.0_21-b06 Eclipse (Helios, Build id: 20100617-1415) plugin was used to run the tests. _Original issue: http://code.google.com/p/powermock/issues/detail?id=268_
priority
test methods are not executed if powermockrule declaration is present from on july if powermockrule declaration is present in a test class test methods annotated with test are not executed example test below passes without any warnings while one of test methods throws npe preparefortest static class public class testcase rule public powermockrule rule new powermockrule test public void testmethodthatthrowsnpe throw new nullpointerexception if rule declaration is commented out test fails as expected quick debug shown that runnable created using evaluate method of powermockstatement is never executed it seems like this runnable is intended to be executed in object execute object instance method method object arguments method of classloaderexecutor but it does not because there is a check based on return type and return type is expected to be non void line powermock ver mockito api with dependencies java eclipse helios build id plugin was used to run the tests original issue
1
86,739
3,728,766,271
IssuesEvent
2016-03-07 02:34:09
Benrnz/BudgetAnalyser
https://api.github.com/repos/Benrnz/BudgetAnalyser
closed
As a reconciliation user I need to see opening balances on ledgers
enhancement Priority-high
So I can double check consistency of data and see if the ledger is increasing or decreasing.
1.0
As a reconciliation user I need to see opening balances on ledgers - So I can double check consistency of data and see if the ledger is increasing or decreasing.
priority
as a reconciliation user i need to see opening balances on ledgers so i can double check consistency of data and see if the ledger is increasing or decreasing
1
354,999
10,575,561,037
IssuesEvent
2019-10-07 15:58:02
compodoc/compodoc
https://api.github.com/repos/compodoc/compodoc
closed
[BUG] @example sections not formatted correctly
Context : JSDoc Priority: High Status: Accepted Time: ~1 hour Type: Bug wontfix
##### **Overview of the issue** @example code is incorrectly formatted. Html tags are displaying, `&quot;` and other characters have been translated and output. ``` /** * This method will display a dialog to the user to capture some arbitary text, you * supply the buttons that should be displayed with the callback that should be invoked * for each (if required). * * @param {string} title the title for the dialog * @param {string} message the messsage to be presented to the user * @param {*} buttonConfigs an array, containing a buttonName and callback, i.e. * @param {boolean} [enableBackdropDismiss=true] optional parameter - allow user to dismiss the prompt by clicking * on the backdrop (default to true)? * * @example * ```typescript * this.alertCtrl.showTextInputWithButtons( * "Pin Verification", * "You have been sent an email. Please check your email and then type the pin code that you have received below.", * [ * { buttonName: "Cancel", callback: null }, * { buttonName: "Verify", callback: (value: string) => { this.sendPin(value); } } * ]); * ``` * * @memberOf IonicAlerterService */ public abstract showTextInputWithButtons( title: string, message: string, buttonConfigs: any, enableBackdropDismiss?: boolean, ): void; ``` version: 1.0.1 ![screen shot 2017-09-08 at 16 43 29](https://user-images.githubusercontent.com/28921238/30219726-e28e139e-94b4-11e7-8e3f-101d53452d65.png) Has changed post version 1.0.0-beta.13 example below is as expected ![screen shot 2017-09-08 at 16 41 49](https://user-images.githubusercontent.com/28921238/30219662-ad262462-94b4-11e7-97db-3f04c8bebfd8.png) ##### **Operating System, Node.js, npm, compodoc version(s)** Node : v8.2.1 npm : 5.3.0 OS : macOS Sierra "@compodoc/compodoc": "1.0.1" ##### **Angular configuration, a `package.json` file in the root folder** "@angular/animations": "4.1.3", "@angular/common": "4.1.3", "@angular/compiler": "4.1.3", "@angular/compiler-cli": "4.1.3", "@angular/core": "4.1.3", "@angular/forms": "4.1.3", "@angular/http": "4.1.3", "@angular/platform-browser": "4.1.3", "@angular/platform-browser-dynamic": "4.1.3", "@angular/platform-server": "4.1.3", "@compodoc/compodoc": "1.0.1", "@types/jasmine": "2.5.41", "@types/node": "7.0.5", "codelyzer": "^3.1.2", "ionic-angular": "3.6.0", "ncp": "^2.0.0", "rexreplace": "^2.5.2", "rimraf": "^2.6.1", "rxjs": "5.4.0", "ts-node": "2.0.0", "tslint": "^5.6.0", "tslint-consistent-codestyle": "^1.6.0", "tslint-eslint-rules": "^4.1.1", "tslint-ionic-rules": "^0.0.11", "typescript": "2.3.4", "zone.js": "0.8.12" ##### **Compodoc installed globally or locally ?** locally ##### **Motivation for or Use Case** See diagrams ##### **Reproduce the error** Example above ##### **Related issues** There was some changes to @example between 1.0.0-beta.13 and 1.0.0-beta.14 ##### **Suggest a Fix**
1.0
[BUG] @example sections not formatted correctly - ##### **Overview of the issue** @example code is incorrectly formatted. Html tags are displaying, `&quot;` and other characters have been translated and output. ``` /** * This method will display a dialog to the user to capture some arbitary text, you * supply the buttons that should be displayed with the callback that should be invoked * for each (if required). * * @param {string} title the title for the dialog * @param {string} message the messsage to be presented to the user * @param {*} buttonConfigs an array, containing a buttonName and callback, i.e. * @param {boolean} [enableBackdropDismiss=true] optional parameter - allow user to dismiss the prompt by clicking * on the backdrop (default to true)? * * @example * ```typescript * this.alertCtrl.showTextInputWithButtons( * "Pin Verification", * "You have been sent an email. Please check your email and then type the pin code that you have received below.", * [ * { buttonName: "Cancel", callback: null }, * { buttonName: "Verify", callback: (value: string) => { this.sendPin(value); } } * ]); * ``` * * @memberOf IonicAlerterService */ public abstract showTextInputWithButtons( title: string, message: string, buttonConfigs: any, enableBackdropDismiss?: boolean, ): void; ``` version: 1.0.1 ![screen shot 2017-09-08 at 16 43 29](https://user-images.githubusercontent.com/28921238/30219726-e28e139e-94b4-11e7-8e3f-101d53452d65.png) Has changed post version 1.0.0-beta.13 example below is as expected ![screen shot 2017-09-08 at 16 41 49](https://user-images.githubusercontent.com/28921238/30219662-ad262462-94b4-11e7-97db-3f04c8bebfd8.png) ##### **Operating System, Node.js, npm, compodoc version(s)** Node : v8.2.1 npm : 5.3.0 OS : macOS Sierra "@compodoc/compodoc": "1.0.1" ##### **Angular configuration, a `package.json` file in the root folder** "@angular/animations": "4.1.3", "@angular/common": "4.1.3", "@angular/compiler": "4.1.3", "@angular/compiler-cli": "4.1.3", "@angular/core": "4.1.3", "@angular/forms": "4.1.3", "@angular/http": "4.1.3", "@angular/platform-browser": "4.1.3", "@angular/platform-browser-dynamic": "4.1.3", "@angular/platform-server": "4.1.3", "@compodoc/compodoc": "1.0.1", "@types/jasmine": "2.5.41", "@types/node": "7.0.5", "codelyzer": "^3.1.2", "ionic-angular": "3.6.0", "ncp": "^2.0.0", "rexreplace": "^2.5.2", "rimraf": "^2.6.1", "rxjs": "5.4.0", "ts-node": "2.0.0", "tslint": "^5.6.0", "tslint-consistent-codestyle": "^1.6.0", "tslint-eslint-rules": "^4.1.1", "tslint-ionic-rules": "^0.0.11", "typescript": "2.3.4", "zone.js": "0.8.12" ##### **Compodoc installed globally or locally ?** locally ##### **Motivation for or Use Case** See diagrams ##### **Reproduce the error** Example above ##### **Related issues** There was some changes to @example between 1.0.0-beta.13 and 1.0.0-beta.14 ##### **Suggest a Fix**
priority
example sections not formatted correctly overview of the issue example code is incorrectly formatted html tags are displaying quot and other characters have been translated and output this method will display a dialog to the user to capture some arbitary text you supply the buttons that should be displayed with the callback that should be invoked for each if required param string title the title for the dialog param string message the messsage to be presented to the user param buttonconfigs an array containing a buttonname and callback i e param boolean optional parameter allow user to dismiss the prompt by clicking on the backdrop default to true example typescript this alertctrl showtextinputwithbuttons pin verification you have been sent an email please check your email and then type the pin code that you have received below buttonname cancel callback null buttonname verify callback value string this sendpin value memberof ionicalerterservice public abstract showtextinputwithbuttons title string message string buttonconfigs any enablebackdropdismiss boolean void version has changed post version beta example below is as expected operating system node js npm compodoc version s node npm os macos sierra compodoc compodoc angular configuration a package json file in the root folder angular animations angular common angular compiler angular compiler cli angular core angular forms angular http angular platform browser angular platform browser dynamic angular platform server compodoc compodoc types jasmine types node codelyzer ionic angular ncp rexreplace rimraf rxjs ts node tslint tslint consistent codestyle tslint eslint rules tslint ionic rules typescript zone js compodoc installed globally or locally locally motivation for or use case see diagrams reproduce the error example above related issues there was some changes to example between beta and beta suggest a fix
1
423,891
12,303,412,525
IssuesEvent
2020-05-11 18:39:04
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
connect.secure.wellsfargo.com - see bug description
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52765 --> **URL**: https://connect.secure.wellsfargo.com/auth/login/present?origin=mobilebrowser **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 6.0 **Tested Another Browser**: Yes Opera **Problem type**: Something else **Description**: unable to sign in **Steps to Reproduce**: yried to sin in, said didnt reconize my username and/or password even tho i used the correct ones over and over again. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190406020101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/5/77b9bd01-4de6-4a40-b0d9-e0f9aff989b9) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
connect.secure.wellsfargo.com - see bug description - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52765 --> **URL**: https://connect.secure.wellsfargo.com/auth/login/present?origin=mobilebrowser **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 6.0 **Tested Another Browser**: Yes Opera **Problem type**: Something else **Description**: unable to sign in **Steps to Reproduce**: yried to sin in, said didnt reconize my username and/or password even tho i used the correct ones over and over again. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190406020101</li><li>channel: alpha</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/5/77b9bd01-4de6-4a40-b0d9-e0f9aff989b9) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
connect secure wellsfargo com see bug description url browser version firefox mobile operating system android tested another browser yes opera problem type something else description unable to sign in steps to reproduce yried to sin in said didnt reconize my username and or password even tho i used the correct ones over and over again browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel alpha hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
274,224
8,558,677,568
IssuesEvent
2018-11-08 18:58:22
spacejam/sled
https://api.github.com/repos/spacejam/sled
closed
multithread write seems to lost data
bug high priority
Bug reports must include: 1. sled version 0.16.2 Hi,I wrote a simple multi-threaded write / read test for sled in [https://github.com/WenjunHuang/SledMultiThreadTest](https://github.com/WenjunHuang/SledMultiThreadTest) I found that when i first used multi-threaded to write some data to sled and then __restart__ the program to start as multi-thread read, some data will lost. But if i first write then read(without restart the program),everything was ok. And If i used single thread write and read, everything was ok whether i restart the program after write or not.
1.0
multithread write seems to lost data - Bug reports must include: 1. sled version 0.16.2 Hi,I wrote a simple multi-threaded write / read test for sled in [https://github.com/WenjunHuang/SledMultiThreadTest](https://github.com/WenjunHuang/SledMultiThreadTest) I found that when i first used multi-threaded to write some data to sled and then __restart__ the program to start as multi-thread read, some data will lost. But if i first write then read(without restart the program),everything was ok. And If i used single thread write and read, everything was ok whether i restart the program after write or not.
priority
multithread write seems to lost data bug reports must include sled version hi i wrote a simple multi threaded write read test for sled in i found that when i first used multi threaded to write some data to sled and then restart the program to start as multi thread read some data will lost but if i first write then read without restart the program everything was ok and if i used single thread write and read everything was ok whether i restart the program after write or not
1
509,846
14,750,213,939
IssuesEvent
2021-01-08 01:32:08
alibaba/nacos
https://api.github.com/repos/alibaba/nacos
closed
JRaft RPC :Error in object information registration returned
area/Nacos Core kind/bug priority/high
<!-- Here is for bug reports and feature requests ONLY! If you're looking for help, please check our mail list、WeChat group and the Gitter room. Please try to use English to describe your issue, or at least provide a snippet of English translation. 我们鼓励使用英文,如果不能直接使用,可以使用翻译软件,您仍旧可以保留中文原文。 --> **Describe the bug** ![F439D49A-ED1B-4F2F-98D9-D3BB30CC4F5C](https://user-images.githubusercontent.com/23445172/103853287-b12bbb00-50e8-11eb-8d2b-ab1614b278c4.png) **Expected behavior** A clear and concise description of what you expected to happen. **Acutally behavior** A clear and concise description of what you actually to happen. **How to Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Desktop (please complete the following information):** - OS: [e.g. Centos] - Version [e.g. nacos-server 1.3.1, nacos-client 1.3.1] - Module [e.g. naming/config] - SDK [e.g. original, spring-cloud-alibaba-nacos, dubbo] **Additional context** Add any other context about the problem here.
1.0
JRaft RPC :Error in object information registration returned - <!-- Here is for bug reports and feature requests ONLY! If you're looking for help, please check our mail list、WeChat group and the Gitter room. Please try to use English to describe your issue, or at least provide a snippet of English translation. 我们鼓励使用英文,如果不能直接使用,可以使用翻译软件,您仍旧可以保留中文原文。 --> **Describe the bug** ![F439D49A-ED1B-4F2F-98D9-D3BB30CC4F5C](https://user-images.githubusercontent.com/23445172/103853287-b12bbb00-50e8-11eb-8d2b-ab1614b278c4.png) **Expected behavior** A clear and concise description of what you expected to happen. **Acutally behavior** A clear and concise description of what you actually to happen. **How to Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Desktop (please complete the following information):** - OS: [e.g. Centos] - Version [e.g. nacos-server 1.3.1, nacos-client 1.3.1] - Module [e.g. naming/config] - SDK [e.g. original, spring-cloud-alibaba-nacos, dubbo] **Additional context** Add any other context about the problem here.
priority
jraft rpc :error in object information registration returned here is for bug reports and feature requests only if you re looking for help please check our mail list、wechat group and the gitter room please try to use english to describe your issue or at least provide a snippet of english translation 我们鼓励使用英文,如果不能直接使用,可以使用翻译软件,您仍旧可以保留中文原文。 describe the bug expected behavior a clear and concise description of what you expected to happen acutally behavior a clear and concise description of what you actually to happen how to reproduce steps to reproduce the behavior go to click on scroll down to see error desktop please complete the following information os version module sdk additional context add any other context about the problem here
1
396,291
11,707,061,953
IssuesEvent
2020-03-08 03:06:43
sjsumlclub/Website
https://api.github.com/repos/sjsumlclub/Website
opened
Add Committees Section
high priority
Here is the text for committees below # Committees Our meetings alternate between _Committee Days_ and _Club Days_. Committee Days let members become experts in their topic of interest, while Club Days promote symbiotic learning as a community. - **Committee Days**: members work within their committee to learn, share resources, and complete projects as a team. - **Club Days**: Committees rotate to host workshops, paper readings, presentations, or fun games (e.g. trivia) for symbiotic club learning. Learners are exposed to important concepts from other parts of AI, and presenters refine their knowledge by teaching others. ## Computer Vision Committee The Computer Vision committee is dedicated to retrieve, process, and analyze digital images with deep learning. Our main focus is on object recognition and classification using a variety of convolutional neural networks. We aim to gain expertise in vision models through hands-on experience. ## Reinforcement Learning Committee The Reinforcement Learning committee is focused on teaching its members Reinforcement Learning's fundamentals, its applications, and its possibilities. We find exciting projects and case studies to analyze and learn from. Through leading workshops and tutorials, the RL committee teaches actionable skills for members to utilize in their own projects. ## Graphics Committee The Graphics committee explores the use of deep learning for generative graphics, i.e. with Generative Adversarial Networks. These topics include interactive rendering, visualization, art, as well as virtual and augmented reality applications. We aim to create beneficial applications with AI visual imagination. Members will gain practical implementation and research skills. ## Natural Language Committee COMING SOON ## Traditional ML Committee The Traditional Machine Learning committee encompasses both traditional machine learning models such as Random Forests, K-Nearest Neighbors and recommender systems as well as machine learning techniques like hyperparameter optimization, dimensionality reduction, cross-validation. We seek to gain knowledge about machine learning techniques benefits both traditional machine learning models as well as neural networks.
1.0
Add Committees Section - Here is the text for committees below # Committees Our meetings alternate between _Committee Days_ and _Club Days_. Committee Days let members become experts in their topic of interest, while Club Days promote symbiotic learning as a community. - **Committee Days**: members work within their committee to learn, share resources, and complete projects as a team. - **Club Days**: Committees rotate to host workshops, paper readings, presentations, or fun games (e.g. trivia) for symbiotic club learning. Learners are exposed to important concepts from other parts of AI, and presenters refine their knowledge by teaching others. ## Computer Vision Committee The Computer Vision committee is dedicated to retrieve, process, and analyze digital images with deep learning. Our main focus is on object recognition and classification using a variety of convolutional neural networks. We aim to gain expertise in vision models through hands-on experience. ## Reinforcement Learning Committee The Reinforcement Learning committee is focused on teaching its members Reinforcement Learning's fundamentals, its applications, and its possibilities. We find exciting projects and case studies to analyze and learn from. Through leading workshops and tutorials, the RL committee teaches actionable skills for members to utilize in their own projects. ## Graphics Committee The Graphics committee explores the use of deep learning for generative graphics, i.e. with Generative Adversarial Networks. These topics include interactive rendering, visualization, art, as well as virtual and augmented reality applications. We aim to create beneficial applications with AI visual imagination. Members will gain practical implementation and research skills. ## Natural Language Committee COMING SOON ## Traditional ML Committee The Traditional Machine Learning committee encompasses both traditional machine learning models such as Random Forests, K-Nearest Neighbors and recommender systems as well as machine learning techniques like hyperparameter optimization, dimensionality reduction, cross-validation. We seek to gain knowledge about machine learning techniques benefits both traditional machine learning models as well as neural networks.
priority
add committees section here is the text for committees below committees our meetings alternate between committee days and club days committee days let members become experts in their topic of interest while club days promote symbiotic learning as a community committee days members work within their committee to learn share resources and complete projects as a team club days committees rotate to host workshops paper readings presentations or fun games e g trivia for symbiotic club learning learners are exposed to important concepts from other parts of ai and presenters refine their knowledge by teaching others computer vision committee the computer vision committee is dedicated to retrieve process and analyze digital images with deep learning our main focus is on object recognition and classification using a variety of convolutional neural networks we aim to gain expertise in vision models through hands on experience reinforcement learning committee the reinforcement learning committee is focused on teaching its members reinforcement learning s fundamentals its applications and its possibilities we find exciting projects and case studies to analyze and learn from through leading workshops and tutorials the rl committee teaches actionable skills for members to utilize in their own projects graphics committee the graphics committee explores the use of deep learning for generative graphics i e with generative adversarial networks these topics include interactive rendering visualization art as well as virtual and augmented reality applications we aim to create beneficial applications with ai visual imagination members will gain practical implementation and research skills natural language committee coming soon traditional ml committee the traditional machine learning committee encompasses both traditional machine learning models such as random forests k nearest neighbors and recommender systems as well as machine learning techniques like hyperparameter optimization dimensionality reduction cross validation we seek to gain knowledge about machine learning techniques benefits both traditional machine learning models as well as neural networks
1
422,404
12,277,877,796
IssuesEvent
2020-05-08 08:49:38
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
github.com - see bug description
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 77.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52612 --> <!-- @extra_labels: browser-fenix --> **URL**: https://github.com/gorhill/uBlock/releases **Browser / Version**: Firefox Mobile 77.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: cannot install the beta version **Steps to Reproduce**: Should install instead of download <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
github.com - see bug description - <!-- @browser: Firefox Mobile 77.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:77.0) Gecko/77.0 Firefox/77.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/52612 --> <!-- @extra_labels: browser-fenix --> **URL**: https://github.com/gorhill/uBlock/releases **Browser / Version**: Firefox Mobile 77.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: cannot install the beta version **Steps to Reproduce**: Should install instead of download <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
github com see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description cannot install the beta version steps to reproduce should install instead of download browser configuration none from with ❤️
1
83,235
3,632,268,035
IssuesEvent
2016-02-11 08:57:52
codebuddiesdotorg/cb-v2-scratch
https://api.github.com/repos/codebuddiesdotorg/cb-v2-scratch
closed
"Currently working on" status
feature high-priority ready
A spot for every person to fill in a “current status” to describe what they’re “currently working on”. This way, you’ll be able to see at a glance what everyone is working on.
1.0
"Currently working on" status - A spot for every person to fill in a “current status” to describe what they’re “currently working on”. This way, you’ll be able to see at a glance what everyone is working on.
priority
currently working on status a spot for every person to fill in a “current status” to describe what they’re “currently working on” this way you’ll be able to see at a glance what everyone is working on
1
671,186
22,747,201,915
IssuesEvent
2022-07-07 10:11:18
codersforcauses/elucidate
https://api.github.com/repos/codersforcauses/elucidate
closed
Fix frontend CI pipeline
bug enhancement frontend priority::high
## Basic Information Rewrite pipeline so that it runs the now dockerized project ## Other Information N/A
1.0
Fix frontend CI pipeline - ## Basic Information Rewrite pipeline so that it runs the now dockerized project ## Other Information N/A
priority
fix frontend ci pipeline basic information rewrite pipeline so that it runs the now dockerized project other information n a
1
208,031
7,135,222,013
IssuesEvent
2018-01-22 23:55:57
richrd/suplemon
https://api.github.com/repos/richrd/suplemon
opened
Rewrite: Core config handling
priority-high rewrite
The rewrite needs a config handler and now would be a good time to implement it. The idea is to make it more or less compatible with Sublime Text. The new format wont be backwards compatible with the current version of suplemon (0.1.X) A common implementation should be used for both core config and keymap config. - [ ] Config file format should be similar to sublime text - [ ] Variable names should be the same as in sublime text - [ ] Single line comments should be supported - [ ] User configs need to be merged on top of defaults - [ ] Defaults can be different from sublime text defaults - [ ] Not all sublime text options will be implemented
1.0
Rewrite: Core config handling - The rewrite needs a config handler and now would be a good time to implement it. The idea is to make it more or less compatible with Sublime Text. The new format wont be backwards compatible with the current version of suplemon (0.1.X) A common implementation should be used for both core config and keymap config. - [ ] Config file format should be similar to sublime text - [ ] Variable names should be the same as in sublime text - [ ] Single line comments should be supported - [ ] User configs need to be merged on top of defaults - [ ] Defaults can be different from sublime text defaults - [ ] Not all sublime text options will be implemented
priority
rewrite core config handling the rewrite needs a config handler and now would be a good time to implement it the idea is to make it more or less compatible with sublime text the new format wont be backwards compatible with the current version of suplemon x a common implementation should be used for both core config and keymap config config file format should be similar to sublime text variable names should be the same as in sublime text single line comments should be supported user configs need to be merged on top of defaults defaults can be different from sublime text defaults not all sublime text options will be implemented
1
143,803
5,530,373,928
IssuesEvent
2017-03-21 02:17:38
kamal1978/LTFHC
https://api.github.com/repos/kamal1978/LTFHC
closed
Back arrow causing crashes
app bug high priority
If I am on the following screens, pressing the back arrow will cause a crash: - Prenatal Fansidar - Prenatal Iron Folate - Menbendazole ...and it seems all subsequent screens in the flow.
1.0
Back arrow causing crashes - If I am on the following screens, pressing the back arrow will cause a crash: - Prenatal Fansidar - Prenatal Iron Folate - Menbendazole ...and it seems all subsequent screens in the flow.
priority
back arrow causing crashes if i am on the following screens pressing the back arrow will cause a crash prenatal fansidar prenatal iron folate menbendazole and it seems all subsequent screens in the flow
1
404,712
11,861,920,798
IssuesEvent
2020-03-25 17:05:12
ME-ICA/tedana
https://api.github.com/repos/ME-ICA/tedana
closed
importing tedana in python seems to lock my multiprocessing to one core
bug high-priority
<!-- This is a suggested issue template for tedana. If there is other information that would be helpful to include, please do not hesitate to add it! Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #. If this is a usage question, please check out NeuroStars here: https://neurostars.org/ and tag your topic with "multi-echo" --> <!-- Summarize the issue in 1-2 sentences, linking other issues if they are relevant Note: simply typing # will prompt you for open issues to select from --> ### Summary I am writing a preprocessing script in python which mostly involves using nipype to wrap afni terminal commands. Normally afni will either automatically use all available cpus or take a manually given amount to process certain commands, however when I import tendana, nipype will only use one. Could this have to do with #188 ? <!-- If needed, add additional detail for: 1. Recreating a bug/problem 2. Any additional context necessary to understand the issue --> ### Additional Detail Here is a sample code for a wrapper in my script: ``` import nipype from multiprocessing import cpu_count from nipype.interfaces import afni as afni from tedana.workflows import t2smap_workflow def warp(self,fileobj1,fileobj2,out_file,transformation=None,args=None,saved_mat_file=None): ThreeDWarp = afni.Warp(in_file=fileobj1,out_file=out_file) if args is not None: ThreeDWarp.inputs.args=args if transformation == 'card2oblique': ThreeDWarp.inputs.oblique_parent = fileobj2 elif transformation == 'deoblique': ThreeDWarp.inputs.deoblique = True elif transformation == 'mni2tta': ThreeDWarp.inputs.mni2tta = True elif transformation == 'tta2mni': ThreeDWarp.inputs.tta2mni = True elif transformation == 'matrix': ThreeDWarp.inputs.matparent = fileobj2 elif transformation == None: print("Warning: no transformation input given") else: print("Warning: none of the transformation options given match the possible arguments. Matching arguments are card2oblique,"+ " deoblique, mni2tta, tta2mni, and matrix") ThreeDWarp.inputs.num_threads = cpu_count() if saved_mat_file: #this is for if the pipline requires saving the 1D matrix tranformation information print('saving matrix') ThreeDWarp.inputs.verbose = True ThreeDWarp.inputs.save_warp = True ThreeDWarp.run() ``` <!-- If desired, add suggested next steps. If you foresee them in a particular order or priority, please use numbering --> ### Next Steps printing os.environ gives me ```({ 'OMP_NUM_THREADS': '1', 'LESSOPEN': '| /usr/bin/lesspipe %s', '_': '/usr/bin/python3', 'MKL_NUM_THREADS': '1', 'NUMEXPR_NUM_THREADS': '1', 'VECLIB_MAXIMUM_THREADS': '1', 'KMP_DUPLICATE_LIB_OK': 'True', 'KMP_INIT_AT_FORK': 'FALSE'})``` I changed OMP_NUM_THREADS and fixed my issue, but wanted to post this so that the devs here would know how this can affect nipype. <!-- Thank you for submitting your issue! If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention. Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience! -->
1.0
importing tedana in python seems to lock my multiprocessing to one core - <!-- This is a suggested issue template for tedana. If there is other information that would be helpful to include, please do not hesitate to add it! Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #. If this is a usage question, please check out NeuroStars here: https://neurostars.org/ and tag your topic with "multi-echo" --> <!-- Summarize the issue in 1-2 sentences, linking other issues if they are relevant Note: simply typing # will prompt you for open issues to select from --> ### Summary I am writing a preprocessing script in python which mostly involves using nipype to wrap afni terminal commands. Normally afni will either automatically use all available cpus or take a manually given amount to process certain commands, however when I import tendana, nipype will only use one. Could this have to do with #188 ? <!-- If needed, add additional detail for: 1. Recreating a bug/problem 2. Any additional context necessary to understand the issue --> ### Additional Detail Here is a sample code for a wrapper in my script: ``` import nipype from multiprocessing import cpu_count from nipype.interfaces import afni as afni from tedana.workflows import t2smap_workflow def warp(self,fileobj1,fileobj2,out_file,transformation=None,args=None,saved_mat_file=None): ThreeDWarp = afni.Warp(in_file=fileobj1,out_file=out_file) if args is not None: ThreeDWarp.inputs.args=args if transformation == 'card2oblique': ThreeDWarp.inputs.oblique_parent = fileobj2 elif transformation == 'deoblique': ThreeDWarp.inputs.deoblique = True elif transformation == 'mni2tta': ThreeDWarp.inputs.mni2tta = True elif transformation == 'tta2mni': ThreeDWarp.inputs.tta2mni = True elif transformation == 'matrix': ThreeDWarp.inputs.matparent = fileobj2 elif transformation == None: print("Warning: no transformation input given") else: print("Warning: none of the transformation options given match the possible arguments. Matching arguments are card2oblique,"+ " deoblique, mni2tta, tta2mni, and matrix") ThreeDWarp.inputs.num_threads = cpu_count() if saved_mat_file: #this is for if the pipline requires saving the 1D matrix tranformation information print('saving matrix') ThreeDWarp.inputs.verbose = True ThreeDWarp.inputs.save_warp = True ThreeDWarp.run() ``` <!-- If desired, add suggested next steps. If you foresee them in a particular order or priority, please use numbering --> ### Next Steps printing os.environ gives me ```({ 'OMP_NUM_THREADS': '1', 'LESSOPEN': '| /usr/bin/lesspipe %s', '_': '/usr/bin/python3', 'MKL_NUM_THREADS': '1', 'NUMEXPR_NUM_THREADS': '1', 'VECLIB_MAXIMUM_THREADS': '1', 'KMP_DUPLICATE_LIB_OK': 'True', 'KMP_INIT_AT_FORK': 'FALSE'})``` I changed OMP_NUM_THREADS and fixed my issue, but wanted to post this so that the devs here would know how this can affect nipype. <!-- Thank you for submitting your issue! If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention. Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience! -->
priority
importing tedana in python seems to lock my multiprocessing to one core this is a suggested issue template for tedana if there is other information that would be helpful to include please do not hesitate to add it before submitting please check to make sure that the issue is not already addressed if there is a related issue then please cross reference it by if this is a usage question please check out neurostars here and tag your topic with multi echo summarize the issue in sentences linking other issues if they are relevant note simply typing will prompt you for open issues to select from summary i am writing a preprocessing script in python which mostly involves using nipype to wrap afni terminal commands normally afni will either automatically use all available cpus or take a manually given amount to process certain commands however when i import tendana nipype will only use one could this have to do with if needed add additional detail for recreating a bug problem any additional context necessary to understand the issue additional detail here is a sample code for a wrapper in my script import nipype from multiprocessing import cpu count from nipype interfaces import afni as afni from tedana workflows import workflow def warp self out file transformation none args none saved mat file none threedwarp afni warp in file out file out file if args is not none threedwarp inputs args args if transformation threedwarp inputs oblique parent elif transformation deoblique threedwarp inputs deoblique true elif transformation threedwarp inputs true elif transformation threedwarp inputs true elif transformation matrix threedwarp inputs matparent elif transformation none print warning no transformation input given else print warning none of the transformation options given match the possible arguments matching arguments are deoblique and matrix threedwarp inputs num threads cpu count if saved mat file this is for if the pipline requires saving the matrix tranformation information print saving matrix threedwarp inputs verbose true threedwarp inputs save warp true threedwarp run if desired add suggested next steps if you foresee them in a particular order or priority please use numbering next steps printing os environ gives me omp num threads lessopen usr bin lesspipe s usr bin mkl num threads numexpr num threads veclib maximum threads kmp duplicate lib ok true kmp init at fork false i changed omp num threads and fixed my issue but wanted to post this so that the devs here would know how this can affect nipype thank you for submitting your issue if you do not receive a response within a calendar week please post a comment on this issue to catch our attention some issues may not be resolved right away due to the volunteer nature of the project thank you for your patience
1
563,181
16,677,036,003
IssuesEvent
2021-06-07 17:32:04
ucb-rit/coldfront
https://api.github.com/repos/ucb-rit/coldfront
opened
Track that a request is being processed
enhancement high priority
Currently, there is no way for an admin to know whether a given request (e.g., cluster account request, new project request, etc.) is currently being processed, so multiple admins may try to process the same request at the same time. One solution may be to add an `assignee` field to each request. Admins would see that it is already assigned, and avoid processing it themselves.
1.0
Track that a request is being processed - Currently, there is no way for an admin to know whether a given request (e.g., cluster account request, new project request, etc.) is currently being processed, so multiple admins may try to process the same request at the same time. One solution may be to add an `assignee` field to each request. Admins would see that it is already assigned, and avoid processing it themselves.
priority
track that a request is being processed currently there is no way for an admin to know whether a given request e g cluster account request new project request etc is currently being processed so multiple admins may try to process the same request at the same time one solution may be to add an assignee field to each request admins would see that it is already assigned and avoid processing it themselves
1
680,103
23,258,243,502
IssuesEvent
2022-08-04 11:14:53
chloebrett/mlvet
https://api.github.com/repos/chloebrett/mlvet
closed
Trying to open a project (with ctrl+O) when a project is already open should prompt a save first
high priority
We prompt a save when the user tries to quit without saving, so we should do the same when they try to open a project without saving the one they already have open
1.0
Trying to open a project (with ctrl+O) when a project is already open should prompt a save first - We prompt a save when the user tries to quit without saving, so we should do the same when they try to open a project without saving the one they already have open
priority
trying to open a project with ctrl o when a project is already open should prompt a save first we prompt a save when the user tries to quit without saving so we should do the same when they try to open a project without saving the one they already have open
1
710,799
24,434,959,578
IssuesEvent
2022-10-06 10:48:24
IAmTamal/Milan
https://api.github.com/repos/IAmTamal/Milan
opened
Minor UI fixes
🟧 priority: high 🛠 goal: fix 🛠 status : under development hacktoberfest
### Description Hey, I was going through the project and I found some places where there's a need to work on the UI <img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194293332-88e2bbb5-1ebe-487e-83db-7bb0ea8e586f.png"> in the above screenshot, I can see some weird styling has been given to the scrollbar for no reason and the images and boxes are not perfectly aligned and the images are also of bad quality. <img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194293737-a724832c-5515-499f-9e90-e2121e0eec43.png"> in this screenshot above, whenever I visit this route it autofocuses to the confirm password field which is a bad UX. in addition to all of this the footer design is not suitable we need to work on it too. I can help you in doing all these little things fix, please assign me this task. Thanks. ### Screenshots <img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194294318-419b8451-58e9-4db5-b4a6-06ffa0f6aa2c.png"> <img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194294337-7e7848ce-a5b3-44d8-8cee-44edf50ddb3d.png"> ### Additional information _No response_ ### 🥦 Browser Brave ### 👀 Have you checked if this issue has been raised before? - [X] I checked and didn't find similar issue ### 🏢 Have you read the Contributing Guidelines? - [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md) ### Are you willing to work on this issue ? Yes I am willing to submit a PR!
1.0
Minor UI fixes - ### Description Hey, I was going through the project and I found some places where there's a need to work on the UI <img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194293332-88e2bbb5-1ebe-487e-83db-7bb0ea8e586f.png"> in the above screenshot, I can see some weird styling has been given to the scrollbar for no reason and the images and boxes are not perfectly aligned and the images are also of bad quality. <img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194293737-a724832c-5515-499f-9e90-e2121e0eec43.png"> in this screenshot above, whenever I visit this route it autofocuses to the confirm password field which is a bad UX. in addition to all of this the footer design is not suitable we need to work on it too. I can help you in doing all these little things fix, please assign me this task. Thanks. ### Screenshots <img width="1440" alt="Screenshot 2022-10-06 at 4 11 40 PM" src="https://user-images.githubusercontent.com/51208270/194294318-419b8451-58e9-4db5-b4a6-06ffa0f6aa2c.png"> <img width="1440" alt="Screenshot 2022-10-06 at 4 12 02 PM" src="https://user-images.githubusercontent.com/51208270/194294337-7e7848ce-a5b3-44d8-8cee-44edf50ddb3d.png"> ### Additional information _No response_ ### 🥦 Browser Brave ### 👀 Have you checked if this issue has been raised before? - [X] I checked and didn't find similar issue ### 🏢 Have you read the Contributing Guidelines? - [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md) ### Are you willing to work on this issue ? Yes I am willing to submit a PR!
priority
minor ui fixes description hey i was going through the project and i found some places where there s a need to work on the ui img width alt screenshot at pm src in the above screenshot i can see some weird styling has been given to the scrollbar for no reason and the images and boxes are not perfectly aligned and the images are also of bad quality img width alt screenshot at pm src in this screenshot above whenever i visit this route it autofocuses to the confirm password field which is a bad ux in addition to all of this the footer design is not suitable we need to work on it too i can help you in doing all these little things fix please assign me this task thanks screenshots img width alt screenshot at pm src img width alt screenshot at pm src additional information no response 🥦 browser brave 👀 have you checked if this issue has been raised before i checked and didn t find similar issue 🏢 have you read the contributing guidelines i have read the are you willing to work on this issue yes i am willing to submit a pr
1
435,289
12,533,871,524
IssuesEvent
2020-06-04 18:22:58
brunomargenats/mate
https://api.github.com/repos/brunomargenats/mate
closed
PHP WARNING HTTP_USER_AGENT
Priority: High Status: Completed Tech: PHP Type: Bug
if (strpos($_SERVER['HTTP_USER_AGENT'], 'MSIE') !== FALSE || strpos($_SERVER['HTTP_USER_AGENT'], 'Trident') !== FALSE) { /* LOAD SUPPORT ONLY FOR INTERNET EXPLORER */ if (!empty($mate_ie_support)){ get_template_part('functions/ie-support'); } } The fix: https://stackoverflow.com/questions/14130830/php-undefined-index-http-user-agent
1.0
PHP WARNING HTTP_USER_AGENT - if (strpos($_SERVER['HTTP_USER_AGENT'], 'MSIE') !== FALSE || strpos($_SERVER['HTTP_USER_AGENT'], 'Trident') !== FALSE) { /* LOAD SUPPORT ONLY FOR INTERNET EXPLORER */ if (!empty($mate_ie_support)){ get_template_part('functions/ie-support'); } } The fix: https://stackoverflow.com/questions/14130830/php-undefined-index-http-user-agent
priority
php warning http user agent if strpos server msie false strpos server trident false load support only for internet explorer if empty mate ie support get template part functions ie support the fix
1
190,055
6,808,315,577
IssuesEvent
2017-11-04 01:04:13
NewsDiffs/newsdiffs
https://api.github.com/repos/NewsDiffs/newsdiffs
opened
Email scraper errors to admins
enhancement high priority
Right now when the scraper has errors, it only logs them to file. At the end of scraping, if this file is non-empty, email to the Django project admins to communicate errors.
1.0
Email scraper errors to admins - Right now when the scraper has errors, it only logs them to file. At the end of scraping, if this file is non-empty, email to the Django project admins to communicate errors.
priority
email scraper errors to admins right now when the scraper has errors it only logs them to file at the end of scraping if this file is non empty email to the django project admins to communicate errors
1
276,629
8,603,329,448
IssuesEvent
2018-11-16 16:31:48
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Migrate images to/from KVM / vCenter DS
Category: vCenter Priority: High Status: Accepted Type: Feature
**Description** Automatic conversion to/from .vmdk to .raw/.qcow2/etc upon "oneimage clone" operations can be implemented **Use case** To facilitate smooth transitions between KVM and VMware based infrastructures, a **Interface Changes** None ## Progress Status - [x] Branch created - [x] Code committed to development branch - [x] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [x] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Migrate images to/from KVM / vCenter DS - **Description** Automatic conversion to/from .vmdk to .raw/.qcow2/etc upon "oneimage clone" operations can be implemented **Use case** To facilitate smooth transitions between KVM and VMware based infrastructures, a **Interface Changes** None ## Progress Status - [x] Branch created - [x] Code committed to development branch - [x] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [x] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
priority
migrate images to from kvm vcenter ds description automatic conversion to from vmdk to raw etc upon oneimage clone operations can be implemented use case to facilitate smooth transitions between kvm and vmware based infrastructures a interface changes none progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
536,696
15,712,448,288
IssuesEvent
2021-03-27 12:10:09
sopra-fs21-group-06/remys-best-client
https://api.github.com/repos/sopra-fs21-group-06/remys-best-client
opened
Player throws cards away and sits round out
high priority task
If no card can be played, all remaining cards of the player are put back on the deck and the player sits this round out. Time: 7h This task is part of user story #65
1.0
Player throws cards away and sits round out - If no card can be played, all remaining cards of the player are put back on the deck and the player sits this round out. Time: 7h This task is part of user story #65
priority
player throws cards away and sits round out if no card can be played all remaining cards of the player are put back on the deck and the player sits this round out time this task is part of user story
1
642,408
20,887,247,986
IssuesEvent
2022-03-23 07:12:04
AY2122S2-CS2103-F09-2/tp
https://api.github.com/repos/AY2122S2-CS2103-F09-2/tp
closed
Clear all command
type.Task priority.High
Edit the clear command to clear all friends, events and logs in the addressbook
1.0
Clear all command - Edit the clear command to clear all friends, events and logs in the addressbook
priority
clear all command edit the clear command to clear all friends events and logs in the addressbook
1
575,790
17,049,935,013
IssuesEvent
2021-07-06 07:43:37
openshift/odo
https://api.github.com/repos/openshift/odo
closed
Link via ServiceBindingRequest, part 2
kind/feature kind/user-story points/2 priority/High
## Description As a developer, I would like to be able to easily inject my source component with the config/secrets/envvars needed to invoke a backing application, so I can focus on may application code instead of configuring application connections. ## Acceptance criteria * User can issue link command and then push changes, when pushed it will be determined that the components are compatible for `ServiceRequestBinding` and perform the function. * User can list linked components * User can unlink, or remove link. ## User Stories - [x] odo link should store link information in devfile #4208 - [x] Suport linking devfile component with another component (https://github.com/openshift/odo/issues/3423) ## Good to have - [x] Disallow self linking of a component (https://github.com/openshift/odo/issues/1052) - [x] Add linked component/service information in odo describe for devfile component and operator services (https://github.com/openshift/odo/issues/3734) ## JIRA Epic Tracking [ODO-47]( https://issues.redhat.com/browse/ODO-47)
1.0
Link via ServiceBindingRequest, part 2 - ## Description As a developer, I would like to be able to easily inject my source component with the config/secrets/envvars needed to invoke a backing application, so I can focus on may application code instead of configuring application connections. ## Acceptance criteria * User can issue link command and then push changes, when pushed it will be determined that the components are compatible for `ServiceRequestBinding` and perform the function. * User can list linked components * User can unlink, or remove link. ## User Stories - [x] odo link should store link information in devfile #4208 - [x] Suport linking devfile component with another component (https://github.com/openshift/odo/issues/3423) ## Good to have - [x] Disallow self linking of a component (https://github.com/openshift/odo/issues/1052) - [x] Add linked component/service information in odo describe for devfile component and operator services (https://github.com/openshift/odo/issues/3734) ## JIRA Epic Tracking [ODO-47]( https://issues.redhat.com/browse/ODO-47)
priority
link via servicebindingrequest part description as a developer i would like to be able to easily inject my source component with the config secrets envvars needed to invoke a backing application so i can focus on may application code instead of configuring application connections acceptance criteria user can issue link command and then push changes when pushed it will be determined that the components are compatible for servicerequestbinding and perform the function user can list linked components user can unlink or remove link user stories odo link should store link information in devfile suport linking devfile component with another component good to have disallow self linking of a component add linked component service information in odo describe for devfile component and operator services jira epic tracking
1