Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
550,069
16,104,339,364
IssuesEvent
2021-04-27 13:19:28
TEAM-SUITS/Suits
https://api.github.com/repos/TEAM-SUITS/Suits
closed
[BUG] ์Šคํ† ๋ฆฌ๋ถ ์˜ค๋ฅ˜
:bug: bug :orange_circle: Priority: High
**Describe the bug๐ŸฆŸ** **์ด ๋ฒ„๊ทธ๋ฅผ ๋ฐœ๊ฒฌํ•˜๋Š” ๋ฐฉ๋ฒ•** 1. Go to ์Šคํ† ๋ฆฌ๋ถ ์‹คํ–‰ `npm run storybook` 2. See error **์›๋ž˜๋Š” ์–ด๋–ป๊ฒŒ ๋™์ž‘ํ•˜๋‚˜์š”?** (๋ฒ„๊ทธ๊ฐ€ ์—†์œผ๋ฉด ์›๋ž˜ ์–ด๋–ป๊ฒŒ ์ž‘๋™ํ•ด์•ผ ํ•˜๋Š”์ง€ ์•Œ๋ ค์ฃผ์„ธ์š”.) ์Šคํ† ๋ฆฌ๋ถ์ด ์ •์ƒ์ ์œผ๋กœ ํ‘œ์‹œ๋˜์–ด์•ผ ํ•จ. **Screenshots** <img width="1161" alt="Screen Shot 2021-04-26 at 11 44 42 PM" src="https://user-images.githubusercontent.com/72863748/116102484-b76e9300-a6e9-11eb-8744-e80fa3d7b70f.png"> **Additional context** theme.styled.js ํŒŒ์ผ์—์„œ props๋กœ ๋ฐ›๋Š” theme์„ ์ œ๋Œ€๋กœ ์ฝ์–ด๋“ค์ด์ง€ ๋ชปํ•ด์„œ ๋ฐœ์ƒํ•˜๋Š” ๊ฒƒ ๊ฐ™์Šต๋‹ˆ๋‹ค. ์ด๋Ÿฐ ์ €๋Ÿฐ ๊ฒƒ๋“ค์„ ์ˆ˜์ •ํ•ด๋ดค๋Š”๋ฐ ์•„์ง ํ•ด๊ฒฐํ•˜์ง€ ๋ชปํ•˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค. ์•„๋ž˜๋Š” ์—๋Ÿฌ ๋ฉ”์‹œ์ง€์— ํ‘œ์‹œ๋˜๋Š” theme.styled.js ํŒŒ์ผ์˜ ๋ฌธ์ œ ์ฝ”๋“œ์ž…๋‹ˆ๋‹ค. ํ•ด๊ฒฐํ•ด๋ณด๋ ค๊ณ  ๊ณ„์† ๋…ธ๋ ฅํ•ด๋ณด๊ฒ ์ง€๋งŒ ํ˜น์‹œ ์›์ธ์„ ์•„์‹ค ๊ฒƒ ๊ฐ™์œผ๋ฉด ์กฐ์–ธ ๋ถ€ํƒ๋“œ๋ฆฝ๋‹ˆ๋‹ค. <img width="622" alt="Screen Shot 2021-04-26 at 11 45 15 PM" src="https://user-images.githubusercontent.com/72863748/116102709-e422aa80-a6e9-11eb-85cc-9cda0559b034.png">
1.0
[BUG] ์Šคํ† ๋ฆฌ๋ถ ์˜ค๋ฅ˜ - **Describe the bug๐ŸฆŸ** **์ด ๋ฒ„๊ทธ๋ฅผ ๋ฐœ๊ฒฌํ•˜๋Š” ๋ฐฉ๋ฒ•** 1. Go to ์Šคํ† ๋ฆฌ๋ถ ์‹คํ–‰ `npm run storybook` 2. See error **์›๋ž˜๋Š” ์–ด๋–ป๊ฒŒ ๋™์ž‘ํ•˜๋‚˜์š”?** (๋ฒ„๊ทธ๊ฐ€ ์—†์œผ๋ฉด ์›๋ž˜ ์–ด๋–ป๊ฒŒ ์ž‘๋™ํ•ด์•ผ ํ•˜๋Š”์ง€ ์•Œ๋ ค์ฃผ์„ธ์š”.) ์Šคํ† ๋ฆฌ๋ถ์ด ์ •์ƒ์ ์œผ๋กœ ํ‘œ์‹œ๋˜์–ด์•ผ ํ•จ. **Screenshots** <img width="1161" alt="Screen Shot 2021-04-26 at 11 44 42 PM" src="https://user-images.githubusercontent.com/72863748/116102484-b76e9300-a6e9-11eb-8744-e80fa3d7b70f.png"> **Additional context** theme.styled.js ํŒŒ์ผ์—์„œ props๋กœ ๋ฐ›๋Š” theme์„ ์ œ๋Œ€๋กœ ์ฝ์–ด๋“ค์ด์ง€ ๋ชปํ•ด์„œ ๋ฐœ์ƒํ•˜๋Š” ๊ฒƒ ๊ฐ™์Šต๋‹ˆ๋‹ค. ์ด๋Ÿฐ ์ €๋Ÿฐ ๊ฒƒ๋“ค์„ ์ˆ˜์ •ํ•ด๋ดค๋Š”๋ฐ ์•„์ง ํ•ด๊ฒฐํ•˜์ง€ ๋ชปํ•˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค. ์•„๋ž˜๋Š” ์—๋Ÿฌ ๋ฉ”์‹œ์ง€์— ํ‘œ์‹œ๋˜๋Š” theme.styled.js ํŒŒ์ผ์˜ ๋ฌธ์ œ ์ฝ”๋“œ์ž…๋‹ˆ๋‹ค. ํ•ด๊ฒฐํ•ด๋ณด๋ ค๊ณ  ๊ณ„์† ๋…ธ๋ ฅํ•ด๋ณด๊ฒ ์ง€๋งŒ ํ˜น์‹œ ์›์ธ์„ ์•„์‹ค ๊ฒƒ ๊ฐ™์œผ๋ฉด ์กฐ์–ธ ๋ถ€ํƒ๋“œ๋ฆฝ๋‹ˆ๋‹ค. <img width="622" alt="Screen Shot 2021-04-26 at 11 45 15 PM" src="https://user-images.githubusercontent.com/72863748/116102709-e422aa80-a6e9-11eb-85cc-9cda0559b034.png">
priority
์Šคํ† ๋ฆฌ๋ถ ์˜ค๋ฅ˜ describe the bug๐ŸฆŸ ์ด ๋ฒ„๊ทธ๋ฅผ ๋ฐœ๊ฒฌํ•˜๋Š” ๋ฐฉ๋ฒ• go to ์Šคํ† ๋ฆฌ๋ถ ์‹คํ–‰ npm run storybook see error ์›๋ž˜๋Š” ์–ด๋–ป๊ฒŒ ๋™์ž‘ํ•˜๋‚˜์š” ๋ฒ„๊ทธ๊ฐ€ ์—†์œผ๋ฉด ์›๋ž˜ ์–ด๋–ป๊ฒŒ ์ž‘๋™ํ•ด์•ผ ํ•˜๋Š”์ง€ ์•Œ๋ ค์ฃผ์„ธ์š” ์Šคํ† ๋ฆฌ๋ถ์ด ์ •์ƒ์ ์œผ๋กœ ํ‘œ์‹œ๋˜์–ด์•ผ ํ•จ screenshots img width alt screen shot at pm src additional context theme styled js ํŒŒ์ผ์—์„œ props๋กœ ๋ฐ›๋Š” theme์„ ์ œ๋Œ€๋กœ ์ฝ์–ด๋“ค์ด์ง€ ๋ชปํ•ด์„œ ๋ฐœ์ƒํ•˜๋Š” ๊ฒƒ ๊ฐ™์Šต๋‹ˆ๋‹ค ์ด๋Ÿฐ ์ €๋Ÿฐ ๊ฒƒ๋“ค์„ ์ˆ˜์ •ํ•ด๋ดค๋Š”๋ฐ ์•„์ง ํ•ด๊ฒฐํ•˜์ง€ ๋ชปํ•˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค ์•„๋ž˜๋Š” ์—๋Ÿฌ ๋ฉ”์‹œ์ง€์— ํ‘œ์‹œ๋˜๋Š” theme styled js ํŒŒ์ผ์˜ ๋ฌธ์ œ ์ฝ”๋“œ์ž…๋‹ˆ๋‹ค ํ•ด๊ฒฐํ•ด๋ณด๋ ค๊ณ  ๊ณ„์† ๋…ธ๋ ฅํ•ด๋ณด๊ฒ ์ง€๋งŒ ํ˜น์‹œ ์›์ธ์„ ์•„์‹ค ๊ฒƒ ๊ฐ™์œผ๋ฉด ์กฐ์–ธ ๋ถ€ํƒ๋“œ๋ฆฝ๋‹ˆ๋‹ค img width alt screen shot at pm src
1
73,156
3,408,238,796
IssuesEvent
2015-12-04 09:35:58
gama-platform/gama
https://api.github.com/repos/gama-platform/gama
closed
Map variables can't be declared as parameter anymore
> Bug Affects Usability Concerns GAML Priority High Version Git
The following code does not compile (parameter 'my map' initial value must be constant) : ``` model MyFirstModel global { map<float,int> myMap; } experiment my_experiment type: gui { parameter "my map" var:myMap <- [1.0::5]; } ``` NB : in the 1.61 version, no problem.
1.0
Map variables can't be declared as parameter anymore - The following code does not compile (parameter 'my map' initial value must be constant) : ``` model MyFirstModel global { map<float,int> myMap; } experiment my_experiment type: gui { parameter "my map" var:myMap <- [1.0::5]; } ``` NB : in the 1.61 version, no problem.
priority
map variables can t be declared as parameter anymore the following code does not compile parameter my map initial value must be constant model myfirstmodel global map mymap experiment my experiment type gui parameter my map var mymap nb in the version no problem
1
235,542
7,739,883,312
IssuesEvent
2018-05-28 18:09:19
IfyAniefuna/experiment_metadata
https://api.github.com/repos/IfyAniefuna/experiment_metadata
opened
Don't have any redefined data within code
high priority refactor
There should not be any strings, lists, or other data structures that are defined multiple times in the code. Having multiple definitions of the same information is a problem when needing to update the definitions in the code since you must hunt each down separately. For data that is built according to specific configurations, define the parts that are unique and shared by each configuration separately, and combine them during the configuration execution. This implementation strategy will allow for a single data source for multiple similar data configurations which partly differ according to the configuration.
1.0
Don't have any redefined data within code - There should not be any strings, lists, or other data structures that are defined multiple times in the code. Having multiple definitions of the same information is a problem when needing to update the definitions in the code since you must hunt each down separately. For data that is built according to specific configurations, define the parts that are unique and shared by each configuration separately, and combine them during the configuration execution. This implementation strategy will allow for a single data source for multiple similar data configurations which partly differ according to the configuration.
priority
don t have any redefined data within code there should not be any strings lists or other data structures that are defined multiple times in the code having multiple definitions of the same information is a problem when needing to update the definitions in the code since you must hunt each down separately for data that is built according to specific configurations define the parts that are unique and shared by each configuration separately and combine them during the configuration execution this implementation strategy will allow for a single data source for multiple similar data configurations which partly differ according to the configuration
1
528,828
15,375,383,782
IssuesEvent
2021-03-02 14:52:36
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Evaluate integration with CCTbx
Diffraction High Priority Stale
This issue was originally [TRAC 11270](http://trac.mantidproject.org/mantid/ticket/11270) Integration with Computational Crystallography Toolbox http://cctbx.sourceforge.net/ Contact Kate Page at SNS for details --- Keywords: SSC 2015 sns
1.0
Evaluate integration with CCTbx - This issue was originally [TRAC 11270](http://trac.mantidproject.org/mantid/ticket/11270) Integration with Computational Crystallography Toolbox http://cctbx.sourceforge.net/ Contact Kate Page at SNS for details --- Keywords: SSC 2015 sns
priority
evaluate integration with cctbx this issue was originally integration with computational crystallography toolbox contact kate page at sns for details keywords ssc sns
1
484,872
13,958,373,315
IssuesEvent
2020-10-24 11:37:55
AY2021S1-CS2113T-T09-1/tp
https://api.github.com/repos/AY2021S1-CS2113T-T09-1/tp
opened
Sort Tasks by priority
priority.High type.Story
As a software engineering manager i should be able to see a list of all the tasks in the project, sorted by priority
1.0
Sort Tasks by priority - As a software engineering manager i should be able to see a list of all the tasks in the project, sorted by priority
priority
sort tasks by priority as a software engineering manager i should be able to see a list of all the tasks in the project sorted by priority
1
527,008
15,306,802,762
IssuesEvent
2021-02-24 19:59:25
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Public Schedule: Add clear all filters icon next to sort order and change minus icon to clear filter icon
Priority: High bug enhancement
Our clear filter icon is confusing for people. Users overwhelmingly report the do not understand that the minus icon is a clear filter icon. Secondly we do not have one place to clear all filters everywhere. Therefore please change the public schedule page as follows: 1. Add a clear filter icon that is similar as below next to the sorting option. This clear filter option should clear all filters of the entire page no matter what filter it is. 2. Change the clear filter minus sign to a clear filter icon that is similar as below. ![Screenshot from 2021-02-24 20-55-33](https://user-images.githubusercontent.com/1583873/109058302-21b38900-76e3-11eb-99fa-afa031235abf.png)
1.0
Public Schedule: Add clear all filters icon next to sort order and change minus icon to clear filter icon - Our clear filter icon is confusing for people. Users overwhelmingly report the do not understand that the minus icon is a clear filter icon. Secondly we do not have one place to clear all filters everywhere. Therefore please change the public schedule page as follows: 1. Add a clear filter icon that is similar as below next to the sorting option. This clear filter option should clear all filters of the entire page no matter what filter it is. 2. Change the clear filter minus sign to a clear filter icon that is similar as below. ![Screenshot from 2021-02-24 20-55-33](https://user-images.githubusercontent.com/1583873/109058302-21b38900-76e3-11eb-99fa-afa031235abf.png)
priority
public schedule add clear all filters icon next to sort order and change minus icon to clear filter icon our clear filter icon is confusing for people users overwhelmingly report the do not understand that the minus icon is a clear filter icon secondly we do not have one place to clear all filters everywhere therefore please change the public schedule page as follows add a clear filter icon that is similar as below next to the sorting option this clear filter option should clear all filters of the entire page no matter what filter it is change the clear filter minus sign to a clear filter icon that is similar as below
1
148,718
5,694,693,556
IssuesEvent
2017-04-15 15:40:09
fcollman/render-python
https://api.github.com/repos/fcollman/render-python
closed
client.import_tilespecs_parallel unstable for programs that call it multiple times
bug priority: high
I had a bug where using import_tilespecs_parallel multiple times in the same python program was failing the second time i used it. My program was also using pathos pool for another process, which worked the first 2 times... so structure is pool.map() pool.close() pool.join() import_tilespecs_parallel pool.map() pool.close() pool.join() import_tilespecs_parallel <<< FAIL! but replacing it with non parallel import it worked fine
1.0
client.import_tilespecs_parallel unstable for programs that call it multiple times - I had a bug where using import_tilespecs_parallel multiple times in the same python program was failing the second time i used it. My program was also using pathos pool for another process, which worked the first 2 times... so structure is pool.map() pool.close() pool.join() import_tilespecs_parallel pool.map() pool.close() pool.join() import_tilespecs_parallel <<< FAIL! but replacing it with non parallel import it worked fine
priority
client import tilespecs parallel unstable for programs that call it multiple times i had a bug where using import tilespecs parallel multiple times in the same python program was failing the second time i used it my program was also using pathos pool for another process which worked the first times so structure is pool map pool close pool join import tilespecs parallel pool map pool close pool join import tilespecs parallel fail but replacing it with non parallel import it worked fine
1
653,994
21,632,904,983
IssuesEvent
2022-05-05 11:39:18
opensrp/opensrp-client-anc
https://api.github.com/repos/opensrp/opensrp-client-anc
closed
[Ona Support Request]: User cannot be logged in. ---urgent
high priority Tech Partner (Nepal Team)
### Affected App or Server Version v1.6.15 ### What kind of support do you need? We cannot add any location unit and we cannot even assign the team. with this issue we couldn't login to the app ### What is the acceptance criteria for your support request? we need to add location unit and assign the team and be able to login to the app ### Relevant Information everything is fine with ona's server but after we change to dh mira server the problem persistes
1.0
[Ona Support Request]: User cannot be logged in. ---urgent - ### Affected App or Server Version v1.6.15 ### What kind of support do you need? We cannot add any location unit and we cannot even assign the team. with this issue we couldn't login to the app ### What is the acceptance criteria for your support request? we need to add location unit and assign the team and be able to login to the app ### Relevant Information everything is fine with ona's server but after we change to dh mira server the problem persistes
priority
user cannot be logged in urgent affected app or server version what kind of support do you need we cannot add any location unit and we cannot even assign the team with this issue we couldn t login to the app what is the acceptance criteria for your support request we need to add location unit and assign the team and be able to login to the app relevant information everything is fine with ona s server but after we change to dh mira server the problem persistes
1
643,045
20,922,213,931
IssuesEvent
2022-03-24 18:30:55
oceanprotocol/ocean.py
https://api.github.com/repos/oceanprotocol/ocean.py
closed
Ocean.py support V4 smart contracts
Priority: High Epic
Background: v4 smart contracts are built (alpha form). They include js unit tests inside the contracts repo. Next up: update ocean.js and ocean.py to support the new functionality. Context: - [V4 super-epic](https://github.com/oceanprotocol/multi-repo-issue/issues/93) - Blog posts on ERC721 + ERC20 for IP. First of three [here](https://blog.oceanprotocol.com/nfts-ip-1-practical-connections-of-erc721-with-intellectual-property-dc216aaf005d) TODOs for this issue: - [ ] get ocean.py to support each of - [ ] NFT-ize base IP in contracts & ocean.js - [ ] Updated metadata - [ ] Arch for >1 DT type - [ ] Diff't revenue account than publisher - [ ] Help community monetize - [ ] Whatever else is needed - [ ] update ocean.py README for one or more new flows - [ ] update docs
1.0
Ocean.py support V4 smart contracts - Background: v4 smart contracts are built (alpha form). They include js unit tests inside the contracts repo. Next up: update ocean.js and ocean.py to support the new functionality. Context: - [V4 super-epic](https://github.com/oceanprotocol/multi-repo-issue/issues/93) - Blog posts on ERC721 + ERC20 for IP. First of three [here](https://blog.oceanprotocol.com/nfts-ip-1-practical-connections-of-erc721-with-intellectual-property-dc216aaf005d) TODOs for this issue: - [ ] get ocean.py to support each of - [ ] NFT-ize base IP in contracts & ocean.js - [ ] Updated metadata - [ ] Arch for >1 DT type - [ ] Diff't revenue account than publisher - [ ] Help community monetize - [ ] Whatever else is needed - [ ] update ocean.py README for one or more new flows - [ ] update docs
priority
ocean py support smart contracts background smart contracts are built alpha form they include js unit tests inside the contracts repo next up update ocean js and ocean py to support the new functionality context blog posts on for ip first of three todos for this issue get ocean py to support each of nft ize base ip in contracts ocean js updated metadata arch for dt type diff t revenue account than publisher help community monetize whatever else is needed update ocean py readme for one or more new flows update docs
1
165,391
6,275,837,167
IssuesEvent
2017-07-18 08:05:37
Certaincy/Intranet
https://api.github.com/repos/Certaincy/Intranet
closed
Implement EF Core
Priority: High Status: Available Type: Enhancement
<!-- Please add a description and acceptance criterias --> ### Description Use EF Core to connect to the Postgres DB ### Acceptance Criterias * "Hello World"-implementation of EF * Get data from the database to the client app
1.0
Implement EF Core - <!-- Please add a description and acceptance criterias --> ### Description Use EF Core to connect to the Postgres DB ### Acceptance Criterias * "Hello World"-implementation of EF * Get data from the database to the client app
priority
implement ef core please add a description and acceptance criterias description use ef core to connect to the postgres db acceptance criterias hello world implementation of ef get data from the database to the client app
1
413,159
12,061,005,629
IssuesEvent
2020-04-15 22:31:35
DarshanShet777/Model-Airport
https://api.github.com/repos/DarshanShet777/Model-Airport
closed
Mechanical Engineering: IR Sensors Bump Columns
High Priority
The robot sensors seem to hit the Columns \#A1 and \#A2. The sensors have to be extended away from the robot to ensure that they're able to connect with the checkposts.
1.0
Mechanical Engineering: IR Sensors Bump Columns - The robot sensors seem to hit the Columns \#A1 and \#A2. The sensors have to be extended away from the robot to ensure that they're able to connect with the checkposts.
priority
mechanical engineering ir sensors bump columns the robot sensors seem to hit the columns and the sensors have to be extended away from the robot to ensure that they re able to connect with the checkposts
1
659,820
21,942,648,738
IssuesEvent
2022-05-23 19:52:37
aedanmc/travelo-hey
https://api.github.com/repos/aedanmc/travelo-hey
closed
GCP project has exceeded its credit balance
bug back-end high priority
The status to the connection to the Google API is "REQUEST DENIED", error message says"You must enable Billing on the Google Cloud Project". I am not sure if we ran out of credits or if the issue is being caused by something else.
1.0
GCP project has exceeded its credit balance - The status to the connection to the Google API is "REQUEST DENIED", error message says"You must enable Billing on the Google Cloud Project". I am not sure if we ran out of credits or if the issue is being caused by something else.
priority
gcp project has exceeded its credit balance the status to the connection to the google api is request denied error message says you must enable billing on the google cloud project i am not sure if we ran out of credits or if the issue is being caused by something else
1
222,290
7,431,336,983
IssuesEvent
2018-03-25 13:46:56
JoaquimLey/transport-eta
https://api.github.com/repos/JoaquimLey/transport-eta
closed
Setup initial project
Priority: High Type: Feature
## Why ๐Ÿค” - Since the original project has become quite stale and the idea now is to develop this with the [Twitch Stream](http://twitch.tv/joaquimley) there is a need to do the initial setup again. # Tasks โš™๏ธ - [x] Initial project setup - [x] Use versions.gradle configuration - [x] ย Initial project scaffold - [x] Include initial dependencies (AAC/D2/ETC)
1.0
Setup initial project - ## Why ๐Ÿค” - Since the original project has become quite stale and the idea now is to develop this with the [Twitch Stream](http://twitch.tv/joaquimley) there is a need to do the initial setup again. # Tasks โš™๏ธ - [x] Initial project setup - [x] Use versions.gradle configuration - [x] ย Initial project scaffold - [x] Include initial dependencies (AAC/D2/ETC)
priority
setup initial project why ๐Ÿค” since the original project has become quite stale and the idea now is to develop this with the there is a need to do the initial setup again tasks โš™๏ธ initial project setup use versions gradle configuration ย initial project scaffold include initial dependencies aac etc
1
756,769
26,485,002,403
IssuesEvent
2023-01-17 17:17:26
MystenLabs/sui
https://api.github.com/repos/MystenLabs/sui
closed
TS SDK: Publishing a package returns RPC Error
Priority: High devx Priority:High
## Steps to Reproduce Issue Use the example from README "To publish a package". ```typescript import { Ed25519Keypair, JsonRpcProvider, RawSigner } from '@mysten/sui.js'; const { execSync } = require('child_process'); // Generate a new Keypair const keypair = new Ed25519Keypair(); // or with // const keypair = Ed25519Keypair.deriveKeypair("mnemonic here") const provider = new JsonRpcProvider(); const signer = new RawSigner(keypair, provider); const compiledModules = JSON.parse( execSync( `${cliPath} move build --dump-bytecode-as-base64 --path ${packagePath}`, { encoding: 'utf-8' } ) ); const modulesInBytes = compiledModules.map((m) => Array.from(new Base64DataBuffer(m).getData()) ); const publishTxn = await signer.publish({ compiledModules: modulesInBytes, gasBudget: 10000, }); console.log('publishTxn', publishTxn); ``` ## Expected Result Expected to get the package publish and log the publishTxn ## Actual Result ``` Error: RPC Error: invalid type: sequence, expected a string at line 1 column 1 at RpcTxnDataSerializer.serializeToBytes (/home/project/snippets/node_modules/.pnpm/@mysten+sui.js@0.22.0/node_modules/@mysten/sui.js/src/signers/txn-data-serializers/rpc-txn-data-serializer.ts:171:13) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async RawSigner.signAndExecuteTransaction (/home/project/snippets/node_modules/.pnpm/@mysten+sui.js@0.22.0/node_modules/@mysten/sui.js/src/signers/signer-with-provider.ts:132:7) at async /home/project/snippets/src/index.ts:31:24 โ€‰ELIFECYCLEโ€‰ Command failed with exit code 1. ``` The error is thrown line 31 which is ```typescript const publishTxn = await signer.publish({ compiledModules: modulesInBytes, gasBudget: 10000, }); ``` I logged `modulesInBytes` and `compiledModules` and it looks correct. ## System Information * OS: PopOs (Debian) * sui: 0.21.0 * Move.toml: 0.21.0
2.0
TS SDK: Publishing a package returns RPC Error - ## Steps to Reproduce Issue Use the example from README "To publish a package". ```typescript import { Ed25519Keypair, JsonRpcProvider, RawSigner } from '@mysten/sui.js'; const { execSync } = require('child_process'); // Generate a new Keypair const keypair = new Ed25519Keypair(); // or with // const keypair = Ed25519Keypair.deriveKeypair("mnemonic here") const provider = new JsonRpcProvider(); const signer = new RawSigner(keypair, provider); const compiledModules = JSON.parse( execSync( `${cliPath} move build --dump-bytecode-as-base64 --path ${packagePath}`, { encoding: 'utf-8' } ) ); const modulesInBytes = compiledModules.map((m) => Array.from(new Base64DataBuffer(m).getData()) ); const publishTxn = await signer.publish({ compiledModules: modulesInBytes, gasBudget: 10000, }); console.log('publishTxn', publishTxn); ``` ## Expected Result Expected to get the package publish and log the publishTxn ## Actual Result ``` Error: RPC Error: invalid type: sequence, expected a string at line 1 column 1 at RpcTxnDataSerializer.serializeToBytes (/home/project/snippets/node_modules/.pnpm/@mysten+sui.js@0.22.0/node_modules/@mysten/sui.js/src/signers/txn-data-serializers/rpc-txn-data-serializer.ts:171:13) at processTicksAndRejections (node:internal/process/task_queues:96:5) at async RawSigner.signAndExecuteTransaction (/home/project/snippets/node_modules/.pnpm/@mysten+sui.js@0.22.0/node_modules/@mysten/sui.js/src/signers/signer-with-provider.ts:132:7) at async /home/project/snippets/src/index.ts:31:24 โ€‰ELIFECYCLEโ€‰ Command failed with exit code 1. ``` The error is thrown line 31 which is ```typescript const publishTxn = await signer.publish({ compiledModules: modulesInBytes, gasBudget: 10000, }); ``` I logged `modulesInBytes` and `compiledModules` and it looks correct. ## System Information * OS: PopOs (Debian) * sui: 0.21.0 * Move.toml: 0.21.0
priority
ts sdk publishing a package returns rpc error steps to reproduce issue use the example from readme to publish a package typescript import jsonrpcprovider rawsigner from mysten sui js const execsync require child process generate a new keypair const keypair new or with const keypair derivekeypair mnemonic here const provider new jsonrpcprovider const signer new rawsigner keypair provider const compiledmodules json parse execsync clipath move build dump bytecode as path packagepath encoding utf const modulesinbytes compiledmodules map m array from new m getdata const publishtxn await signer publish compiledmodules modulesinbytes gasbudget console log publishtxn publishtxn expected result expected to get the package publish and log the publishtxn actual result error rpc error invalid type sequence expected a string at line column at rpctxndataserializer serializetobytes home project snippets node modules pnpm mysten sui js node modules mysten sui js src signers txn data serializers rpc txn data serializer ts at processticksandrejections node internal process task queues at async rawsigner signandexecutetransaction home project snippets node modules pnpm mysten sui js node modules mysten sui js src signers signer with provider ts at async home project snippets src index ts โ€‰elifecycleโ€‰ command failed with exit code the error is thrown line which is typescript const publishtxn await signer publish compiledmodules modulesinbytes gasbudget i logged modulesinbytes and compiledmodules and it looks correct system information os popos debian sui move toml
1
327,551
9,977,295,901
IssuesEvent
2019-07-09 16:54:30
lbryio/lbry-sdk
https://api.github.com/repos/lbryio/lbry-sdk
closed
asyncio warning on startup / spam on shutdown.
priority: high
startup: ``` 2019-06-25 09:54:38,096 INFO lbry.extras.daemon.Components:258: Starting torba wallet 2019-06-25 09:54:38,151 INFO lbry.extras.daemon.Components:336: start the dht 2019-06-25 09:54:38,152 WARNING lbry.extras.daemon.Components:342: UPnP component failed to get external ip 2019-06-25 09:54:38,163 INFO torba.client.basedatabase:208: connecting to database: c:\testing\lbrytv\wallet\lbc_mainnet\blockchain.db 2019-06-25 09:54:38,263 INFO lbry.extras.daemon.Components:361: Started the dht 2019-06-25 09:54:38,268 INFO lbry.dht.node:118: DHT node listening on UDP 0.0.0.0:4444 2019-06-25 09:54:38,284 DEBUG lbry.dht.protocol.iterative_find:267: probing 18.213.182.111:4444 2019-06-25 09:54:38,286 DEBUG lbry.dht.protocol.iterative_find:267: probing 35.161.227.101:4444 2019-06-25 09:54:38,289 DEBUG lbry.dht.protocol.iterative_find:267: probing 3.122.90.177:4444 2019-06-25 09:54:38,291 DEBUG lbry.dht.protocol.iterative_find:267: probing 52.78.118.133:4444 Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "C:\Users\thoma\Documents\lbry\lbry-venv\Scripts\lbrynet-script.py", line 11, in <module> load_entry_point('lbry', 'console_scripts', 'lbrynet')() File "c:\users\thoma\documents\lbry\lbry\lbry\extras\cli.py", line 287, in main loop.run_until_complete(daemon.start()) File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py", line 571, in run_until_complete self.run_forever() File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py", line 539, in run_forever self._run_once() File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py", line 1767, in _run_once handle._run() File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\events.py", line 88, in _run self._context.run(self._callback, *self._args) File "c:\users\thoma\documents\lbry\torba\torba\client\basenetwork.py", line 205, in ensure_connections await asyncio.wait([asyncio.sleep(3), self._lost_master.wait()], return_when='FIRST_COMPLETED') File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py", line 361, in wait fs = {ensure_future(f, loop=loop) for f in set(fs)} File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py", line 361, in <setcomp> fs = {ensure_future(f, loop=loop) for f in set(fs)} task: <Task pending coro=<Event.wait() running at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\locks.py:293> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x05D9FCB0>()] created at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py:396> cb=[_wait.<locals>._on_completion() at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py:440] created at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py:361> 2019-06-25 09:54:38,292 DEBUG lbry.dht.protocol.iterative_find:192: running 4 probes 2019-06-25 09:54:38,318 INFO torba.client.basenetwork:89: Successfully connected to SP ``` Shutdown: ```019-06-25 01:57:48,687 INFO lbry.extras.daemon.Daemon:471: finished shutting down Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "lbry\extras\cli.py", line 324, in <module> File "lbry\extras\cli.py", line 287, in main File "asyncio\base_events.py", line 571, in run_until_complete File "asyncio\base_events.py", line 539, in run_forever File "asyncio\base_events.py", line 1767, in _run_once File "asyncio\events.py", line 88, in _run File "torba\client\basenetwork.py", line 205, in ensure_connections File "asyncio\tasks.py", line 361, in wait File "asyncio\tasks.py", line 361, in <setcomp> task: <Task pending coro=<Event.wait() running at asyncio\locks.py:293> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x08159DB0>()] created at asyncio\base_events.py:396> cb=[_wait.<locals>._on_completion() at asyncio\tasks.py:440] created at asyncio\tasks.py:361> Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "lbry\extras\cli.py", line 324, in <module> File "lbry\extras\cli.py", line 289, in main File "asyncio\base_events.py", line 571, in run_until_complete File "asyncio\base_events.py", line 539, in run_forever File "asyncio\base_events.py", line 1767, in _run_once File "asyncio\events.py", line 88, in _run File "torba\client\basenetwork.py", line 205, in ensure_connections File "asyncio\tasks.py", line 361, in wait File "asyncio\tasks.py", line 361, in <setcomp> task: <Task pending coro=<Event.wait() done, defined at asyncio\locks.py:280> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x08186DF0>()] created at asyncio\base_events.py:396> cb=[_wait.<locals>._on_completion() at asyncio\tasks.py:440] created at asyncio\tasks.py:361> Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "lbry\extras\cli.py", line 324, in <module> File "lbry\extras\cli.py", line 287, in main File "asyncio\base_events.py", line 571, in run_until_complete File "asyncio\base_events.py", line 539, in run_forever File "asyncio\base_events.py", line 1767, in _run_once File "asyncio\events.py", line 88, in _run File "torba\client\basenetwork.py", line 205, in ensure_connections File "asyncio\tasks.py", line 361, in wait File "asyncio\tasks.py", line 361, in <setcomp> task: <Task pending coro=<Event.wait() running at asyncio\locks.py:293> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x08166150>()] created at asyncio\base_events.py:396> created at asyncio\tasks.py:361> ```
1.0
asyncio warning on startup / spam on shutdown. - startup: ``` 2019-06-25 09:54:38,096 INFO lbry.extras.daemon.Components:258: Starting torba wallet 2019-06-25 09:54:38,151 INFO lbry.extras.daemon.Components:336: start the dht 2019-06-25 09:54:38,152 WARNING lbry.extras.daemon.Components:342: UPnP component failed to get external ip 2019-06-25 09:54:38,163 INFO torba.client.basedatabase:208: connecting to database: c:\testing\lbrytv\wallet\lbc_mainnet\blockchain.db 2019-06-25 09:54:38,263 INFO lbry.extras.daemon.Components:361: Started the dht 2019-06-25 09:54:38,268 INFO lbry.dht.node:118: DHT node listening on UDP 0.0.0.0:4444 2019-06-25 09:54:38,284 DEBUG lbry.dht.protocol.iterative_find:267: probing 18.213.182.111:4444 2019-06-25 09:54:38,286 DEBUG lbry.dht.protocol.iterative_find:267: probing 35.161.227.101:4444 2019-06-25 09:54:38,289 DEBUG lbry.dht.protocol.iterative_find:267: probing 3.122.90.177:4444 2019-06-25 09:54:38,291 DEBUG lbry.dht.protocol.iterative_find:267: probing 52.78.118.133:4444 Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "C:\Users\thoma\Documents\lbry\lbry-venv\Scripts\lbrynet-script.py", line 11, in <module> load_entry_point('lbry', 'console_scripts', 'lbrynet')() File "c:\users\thoma\documents\lbry\lbry\lbry\extras\cli.py", line 287, in main loop.run_until_complete(daemon.start()) File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py", line 571, in run_until_complete self.run_forever() File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py", line 539, in run_forever self._run_once() File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py", line 1767, in _run_once handle._run() File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\events.py", line 88, in _run self._context.run(self._callback, *self._args) File "c:\users\thoma\documents\lbry\torba\torba\client\basenetwork.py", line 205, in ensure_connections await asyncio.wait([asyncio.sleep(3), self._lost_master.wait()], return_when='FIRST_COMPLETED') File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py", line 361, in wait fs = {ensure_future(f, loop=loop) for f in set(fs)} File "C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py", line 361, in <setcomp> fs = {ensure_future(f, loop=loop) for f in set(fs)} task: <Task pending coro=<Event.wait() running at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\locks.py:293> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x05D9FCB0>()] created at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\base_events.py:396> cb=[_wait.<locals>._on_completion() at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py:440] created at C:\Users\thoma\AppData\Local\Programs\Python\Python37-32\lib\asyncio\tasks.py:361> 2019-06-25 09:54:38,292 DEBUG lbry.dht.protocol.iterative_find:192: running 4 probes 2019-06-25 09:54:38,318 INFO torba.client.basenetwork:89: Successfully connected to SP ``` Shutdown: ```019-06-25 01:57:48,687 INFO lbry.extras.daemon.Daemon:471: finished shutting down Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "lbry\extras\cli.py", line 324, in <module> File "lbry\extras\cli.py", line 287, in main File "asyncio\base_events.py", line 571, in run_until_complete File "asyncio\base_events.py", line 539, in run_forever File "asyncio\base_events.py", line 1767, in _run_once File "asyncio\events.py", line 88, in _run File "torba\client\basenetwork.py", line 205, in ensure_connections File "asyncio\tasks.py", line 361, in wait File "asyncio\tasks.py", line 361, in <setcomp> task: <Task pending coro=<Event.wait() running at asyncio\locks.py:293> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x08159DB0>()] created at asyncio\base_events.py:396> cb=[_wait.<locals>._on_completion() at asyncio\tasks.py:440] created at asyncio\tasks.py:361> Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "lbry\extras\cli.py", line 324, in <module> File "lbry\extras\cli.py", line 289, in main File "asyncio\base_events.py", line 571, in run_until_complete File "asyncio\base_events.py", line 539, in run_forever File "asyncio\base_events.py", line 1767, in _run_once File "asyncio\events.py", line 88, in _run File "torba\client\basenetwork.py", line 205, in ensure_connections File "asyncio\tasks.py", line 361, in wait File "asyncio\tasks.py", line 361, in <setcomp> task: <Task pending coro=<Event.wait() done, defined at asyncio\locks.py:280> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x08186DF0>()] created at asyncio\base_events.py:396> cb=[_wait.<locals>._on_completion() at asyncio\tasks.py:440] created at asyncio\tasks.py:361> Task was destroyed but it is pending! source_traceback: Object created at (most recent call last): File "lbry\extras\cli.py", line 324, in <module> File "lbry\extras\cli.py", line 287, in main File "asyncio\base_events.py", line 571, in run_until_complete File "asyncio\base_events.py", line 539, in run_forever File "asyncio\base_events.py", line 1767, in _run_once File "asyncio\events.py", line 88, in _run File "torba\client\basenetwork.py", line 205, in ensure_connections File "asyncio\tasks.py", line 361, in wait File "asyncio\tasks.py", line 361, in <setcomp> task: <Task pending coro=<Event.wait() running at asyncio\locks.py:293> wait_for=<Future pending cb=[<TaskWakeupMethWrapper object at 0x08166150>()] created at asyncio\base_events.py:396> created at asyncio\tasks.py:361> ```
priority
asyncio warning on startup spam on shutdown startup info lbry extras daemon components starting torba wallet info lbry extras daemon components start the dht warning lbry extras daemon components upnp component failed to get external ip info torba client basedatabase connecting to database c testing lbrytv wallet lbc mainnet blockchain db info lbry extras daemon components started the dht info lbry dht node dht node listening on udp debug lbry dht protocol iterative find probing debug lbry dht protocol iterative find probing debug lbry dht protocol iterative find probing debug lbry dht protocol iterative find probing task was destroyed but it is pending source traceback object created at most recent call last file c users thoma documents lbry lbry venv scripts lbrynet script py line in load entry point lbry console scripts lbrynet file c users thoma documents lbry lbry lbry extras cli py line in main loop run until complete daemon start file c users thoma appdata local programs python lib asyncio base events py line in run until complete self run forever file c users thoma appdata local programs python lib asyncio base events py line in run forever self run once file c users thoma appdata local programs python lib asyncio base events py line in run once handle run file c users thoma appdata local programs python lib asyncio events py line in run self context run self callback self args file c users thoma documents lbry torba torba client basenetwork py line in ensure connections await asyncio wait return when first completed file c users thoma appdata local programs python lib asyncio tasks py line in wait fs ensure future f loop loop for f in set fs file c users thoma appdata local programs python lib asyncio tasks py line in fs ensure future f loop loop for f in set fs task wait for cb created at c users thoma appdata local programs python lib asyncio tasks py debug lbry dht protocol iterative find running probes info torba client basenetwork successfully connected to sp shutdown info lbry extras daemon daemon finished shutting down task was destroyed but it is pending source traceback object created at most recent call last file lbry extras cli py line in file lbry extras cli py line in main file asyncio base events py line in run until complete file asyncio base events py line in run forever file asyncio base events py line in run once file asyncio events py line in run file torba client basenetwork py line in ensure connections file asyncio tasks py line in wait file asyncio tasks py line in task wait for cb created at asyncio tasks py task was destroyed but it is pending source traceback object created at most recent call last file lbry extras cli py line in file lbry extras cli py line in main file asyncio base events py line in run until complete file asyncio base events py line in run forever file asyncio base events py line in run once file asyncio events py line in run file torba client basenetwork py line in ensure connections file asyncio tasks py line in wait file asyncio tasks py line in task wait for cb created at asyncio tasks py task was destroyed but it is pending source traceback object created at most recent call last file lbry extras cli py line in file lbry extras cli py line in main file asyncio base events py line in run until complete file asyncio base events py line in run forever file asyncio base events py line in run once file asyncio events py line in run file torba client basenetwork py line in ensure connections file asyncio tasks py line in wait file asyncio tasks py line in task wait for created at asyncio tasks py
1
718,612
24,725,705,360
IssuesEvent
2022-10-20 13:55:24
mmohare2019/SE_Project
https://api.github.com/repos/mmohare2019/SE_Project
opened
P3 - Log out (1 pt)
User Story High priority
As a parent, I want to log out of the app so that I can assure my data and credentials are safe
1.0
P3 - Log out (1 pt) - As a parent, I want to log out of the app so that I can assure my data and credentials are safe
priority
log out pt as a parent i want to log out of the app so that i can assure my data and credentials are safe
1
169,161
6,395,900,568
IssuesEvent
2017-08-04 14:20:19
InVisionApp/kit-deploymentizer
https://api.github.com/repos/InVisionApp/kit-deploymentizer
closed
Change the mustache templates in kubernetes-coreos to support the new labels.
high-priority
**Manifests: **(generated as: $service-deployment-$sha.yaml) metadata.name: $service-deployment-`{{{deployment.sha}}}` metadata.labels.state: live metadata.labels.sha: `{{{deployment.sha}}}` metadata.labels.service: `{{{name}}}` selector.matchLabels.state: live From: [Quick Rollbacks](https://invisionapp.quip.com/npgPANELpg94) Start of PR: https://github.com/InVisionApp/kit-deploymentizer/issues/43
1.0
Change the mustache templates in kubernetes-coreos to support the new labels. - **Manifests: **(generated as: $service-deployment-$sha.yaml) metadata.name: $service-deployment-`{{{deployment.sha}}}` metadata.labels.state: live metadata.labels.sha: `{{{deployment.sha}}}` metadata.labels.service: `{{{name}}}` selector.matchLabels.state: live From: [Quick Rollbacks](https://invisionapp.quip.com/npgPANELpg94) Start of PR: https://github.com/InVisionApp/kit-deploymentizer/issues/43
priority
change the mustache templates in kubernetes coreos to support the new labels manifests generated as service deployment sha yaml metadata name service deployment deployment sha metadata labels state live metadata labels sha deployment sha metadata labels service name selector matchlabels state live from start of pr
1
705,540
24,238,370,077
IssuesEvent
2022-09-27 03:02:30
paperclip-ui/paperclip
https://api.github.com/repos/paperclip-ui/paperclip
closed
Ability to embed Paperclip templates in JSX code?
priority: low impact: high effort: high area: DSL
Or basically something that allows people to work out of the same file vs having two separate files. Maybe go in the direction of allowing scripts to be embedded in PC files, much like Vue.
1.0
Ability to embed Paperclip templates in JSX code? - Or basically something that allows people to work out of the same file vs having two separate files. Maybe go in the direction of allowing scripts to be embedded in PC files, much like Vue.
priority
ability to embed paperclip templates in jsx code or basically something that allows people to work out of the same file vs having two separate files maybe go in the direction of allowing scripts to be embedded in pc files much like vue
1
328,753
9,999,573,032
IssuesEvent
2019-07-12 11:03:02
turbolabz/transfer-bug-track
https://api.github.com/repos/turbolabz/transfer-bug-track
opened
Game crashes when changing the wifi connection on opponent's turn while playing Quick Match
Priority: High Severity: Critical Type: bug
Pre-Condition: Player is playing a Quick Match Steps: On opponent's turn, change the wifi connection by going in the settings Get back into the game Opponent's clock is ticking After sometime, game crashes.
1.0
Game crashes when changing the wifi connection on opponent's turn while playing Quick Match - Pre-Condition: Player is playing a Quick Match Steps: On opponent's turn, change the wifi connection by going in the settings Get back into the game Opponent's clock is ticking After sometime, game crashes.
priority
game crashes when changing the wifi connection on opponent s turn while playing quick match pre condition player is playing a quick match steps on opponent s turn change the wifi connection by going in the settings get back into the game opponent s clock is ticking after sometime game crashes
1
34,013
2,774,382,696
IssuesEvent
2015-05-04 08:24:31
punongbayan-araullo/tickets
https://api.github.com/repos/punongbayan-araullo/tickets
opened
Change label from "An authorized project has been created" to "New Authorized Project Created"
other priority - high status - accepted system - projects
Change label from "An authorized project has been created" to "New Authorized Project Created"
1.0
Change label from "An authorized project has been created" to "New Authorized Project Created" - Change label from "An authorized project has been created" to "New Authorized Project Created"
priority
change label from an authorized project has been created to new authorized project created change label from an authorized project has been created to new authorized project created
1
693,108
23,763,414,237
IssuesEvent
2022-09-01 10:47:20
CarmenMariaMP/Clap
https://api.github.com/repos/CarmenMariaMP/Clap
closed
H5+H6+H7+H8 - Upload and edit artistic content
Epic high-priority
Allow a content creator to upload artistic content from the different categories that exits, give credit to the content creators involve and assign multiple tags to that content. The categories are as follows: - Cinema: they can upload short films and posters. Its publication will have characteristics such as genre, duration, synopsis... - Painting: they can upload paintings. Its publication will have characteristics such as name of the painting, artist, size... - Photography: they can upload photos. Its publication will have characteristics such as title, location, model/sโ€ฆ - Music: they can upload music videos and images. Its publication will have characteristics such as name of song, singer, lyricist... - Dance: they can upload dance videos and images. Its publication will have characteristics such as name of song, dancer, dance style... - General: they can upload videos and images of general content. Its publication will have characteristics such as title, descriptionโ€ฆ Moreover, allow the content creator to edit the information related to a post.
1.0
H5+H6+H7+H8 - Upload and edit artistic content - Allow a content creator to upload artistic content from the different categories that exits, give credit to the content creators involve and assign multiple tags to that content. The categories are as follows: - Cinema: they can upload short films and posters. Its publication will have characteristics such as genre, duration, synopsis... - Painting: they can upload paintings. Its publication will have characteristics such as name of the painting, artist, size... - Photography: they can upload photos. Its publication will have characteristics such as title, location, model/sโ€ฆ - Music: they can upload music videos and images. Its publication will have characteristics such as name of song, singer, lyricist... - Dance: they can upload dance videos and images. Its publication will have characteristics such as name of song, dancer, dance style... - General: they can upload videos and images of general content. Its publication will have characteristics such as title, descriptionโ€ฆ Moreover, allow the content creator to edit the information related to a post.
priority
upload and edit artistic content allow a content creator to upload artistic content from the different categories that exits give credit to the content creators involve and assign multiple tags to that content the categories are as follows cinema they can upload short films and posters its publication will have characteristics such as genre duration synopsis painting they can upload paintings its publication will have characteristics such as name of the painting artist size photography they can upload photos its publication will have characteristics such as title location model sโ€ฆ music they can upload music videos and images its publication will have characteristics such as name of song singer lyricist dance they can upload dance videos and images its publication will have characteristics such as name of song dancer dance style general they can upload videos and images of general content its publication will have characteristics such as title descriptionโ€ฆ moreover allow the content creator to edit the information related to a post
1
746,240
26,022,200,885
IssuesEvent
2022-12-21 13:34:56
conan-io/conan
https://api.github.com/repos/conan-io/conan
closed
Graph propagation improvements
type: look into stage: queue priority: high complex: high
POC/Investigation required. Looks like every node needs to have the complete computation of the graph, but specific for every node, upstream nodes can "change" (e.g. require/build require) depending on the consumer. But, computing that is very very expensive, so we need to investigate if is there a way to only calculate the `deps_cpp_info` for the nodes to be built. Currently, this issue is the blocker for the POC of x-build #5592 and the origin of other issues like https://github.com/conan-io/conan/issues/5474 and https://github.com/conan-io/conan/issues/5682
1.0
Graph propagation improvements - POC/Investigation required. Looks like every node needs to have the complete computation of the graph, but specific for every node, upstream nodes can "change" (e.g. require/build require) depending on the consumer. But, computing that is very very expensive, so we need to investigate if is there a way to only calculate the `deps_cpp_info` for the nodes to be built. Currently, this issue is the blocker for the POC of x-build #5592 and the origin of other issues like https://github.com/conan-io/conan/issues/5474 and https://github.com/conan-io/conan/issues/5682
priority
graph propagation improvements poc investigation required looks like every node needs to have the complete computation of the graph but specific for every node upstream nodes can change e g require build require depending on the consumer but computing that is very very expensive so we need to investigate if is there a way to only calculate the deps cpp info for the nodes to be built currently this issue is the blocker for the poc of x build and the origin of other issues like and
1
317,253
9,662,506,874
IssuesEvent
2019-05-20 21:01:21
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
Create contract SDK package
Chore Monorepo Priority: High
Should include createMarket, placeTrade, etc. wrappers. Should also add interface for batching.
1.0
Create contract SDK package - Should include createMarket, placeTrade, etc. wrappers. Should also add interface for batching.
priority
create contract sdk package should include createmarket placetrade etc wrappers should also add interface for batching
1
284,864
8,751,423,797
IssuesEvent
2018-12-13 22:17:12
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Client server to cielo doesn't work
bug likelihood low priority reviewed severity high
I was running client server to cielo from inca3 and it wouldn't connect with the following error message: can't open master pty: /dev/ttyTfcan't open sale pty: -1dup2 error to stdindup2 error to stdoutdup2 error to stderrinca3{brugger}23: Version 2.4.2 still works. version 2.5.0 doesn't work either. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1108 Status: Resolved Project: VisIt Tracker: Bug Priority: Urgent Subject: Client server to cielo doesn't work Assigned to: Eric Brugger Category: - Target version: 2.6.2 Author: Eric Brugger Start: 06/26/2012 Due date: % Done: 100% Estimated time: 16.00 hours Created: 06/26/2012 04:07 pm Updated: 03/13/2013 11:34 am Likelihood: 2 - Rare Severity: 5 - Very Serious Found in version: 2.5.0 Impact: Expected Use: OS: All Support Group: Any Description: I was running client server to cielo from inca3 and it wouldn't connect with the following error message: can't open master pty: /dev/ttyTfcan't open sale pty: -1dup2 error to stdindup2 error to stdoutdup2 error to stderrinca3{brugger}23: Version 2.4.2 still works. version 2.5.0 doesn't work either. Comments: See if -nopty helps. It turns out using nopty was the source of the problem. When I fixed the issue with pseudo terminals and no longer had to specify nopty to get visit to connect client/server to cielo, things started to work again. If I add the -nopty it still hangs.
1.0
Client server to cielo doesn't work - I was running client server to cielo from inca3 and it wouldn't connect with the following error message: can't open master pty: /dev/ttyTfcan't open sale pty: -1dup2 error to stdindup2 error to stdoutdup2 error to stderrinca3{brugger}23: Version 2.4.2 still works. version 2.5.0 doesn't work either. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1108 Status: Resolved Project: VisIt Tracker: Bug Priority: Urgent Subject: Client server to cielo doesn't work Assigned to: Eric Brugger Category: - Target version: 2.6.2 Author: Eric Brugger Start: 06/26/2012 Due date: % Done: 100% Estimated time: 16.00 hours Created: 06/26/2012 04:07 pm Updated: 03/13/2013 11:34 am Likelihood: 2 - Rare Severity: 5 - Very Serious Found in version: 2.5.0 Impact: Expected Use: OS: All Support Group: Any Description: I was running client server to cielo from inca3 and it wouldn't connect with the following error message: can't open master pty: /dev/ttyTfcan't open sale pty: -1dup2 error to stdindup2 error to stdoutdup2 error to stderrinca3{brugger}23: Version 2.4.2 still works. version 2.5.0 doesn't work either. Comments: See if -nopty helps. It turns out using nopty was the source of the problem. When I fixed the issue with pseudo terminals and no longer had to specify nopty to get visit to connect client/server to cielo, things started to work again. If I add the -nopty it still hangs.
priority
client server to cielo doesn t work i was running client server to cielo from and it wouldn t connect with the following error message can t open master pty dev ttytfcan t open sale pty error to error to error to brugger version still works version doesn t work either redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority urgent subject client server to cielo doesn t work assigned to eric brugger category target version author eric brugger start due date done estimated time hours created pm updated am likelihood rare severity very serious found in version impact expected use os all support group any description i was running client server to cielo from and it wouldn t connect with the following error message can t open master pty dev ttytfcan t open sale pty error to error to error to brugger version still works version doesn t work either comments see if nopty helps it turns out using nopty was the source of the problem when i fixed the issue with pseudo terminals and no longer had to specify nopty to get visit to connect client server to cielo things started to work again if i add the nopty it still hangs
1
259,677
8,198,844,071
IssuesEvent
2018-08-31 17:53:33
GMOD/jbrowse
https://api.github.com/repos/GMOD/jbrowse
closed
factor out parsers and stores as independent npm modules
big task high priority
It would be great if JBrowse could publish their parsers (GFF,BAM,Wig,VCF,etc.) as separate modules to npm so that they can be easily reused by other JavaScript projects like BioJS or Bionode. I would suggest this tiny roadmap 1. Create a list of interesting packages that can be reused 2. (Distribute the work and), convert the files to npm modules and publish them For 1) we should also see whether there are other already existing JS projects. I know of these two parsers, but I haven't tested one of them. https://github.com/anilthanki/biojs-io-wig http://registry.biojs.net/client/#/detail/biojs-io-sam I just start the list for 1) - GFF3 - GTF - VCF - BAM - BigWig (please modify this list) @bmpvieira @homonecloco @saketkc
1.0
factor out parsers and stores as independent npm modules - It would be great if JBrowse could publish their parsers (GFF,BAM,Wig,VCF,etc.) as separate modules to npm so that they can be easily reused by other JavaScript projects like BioJS or Bionode. I would suggest this tiny roadmap 1. Create a list of interesting packages that can be reused 2. (Distribute the work and), convert the files to npm modules and publish them For 1) we should also see whether there are other already existing JS projects. I know of these two parsers, but I haven't tested one of them. https://github.com/anilthanki/biojs-io-wig http://registry.biojs.net/client/#/detail/biojs-io-sam I just start the list for 1) - GFF3 - GTF - VCF - BAM - BigWig (please modify this list) @bmpvieira @homonecloco @saketkc
priority
factor out parsers and stores as independent npm modules it would be great if jbrowse could publish their parsers gff bam wig vcf etc as separate modules to npm so that they can be easily reused by other javascript projects like biojs or bionode i would suggest this tiny roadmap create a list of interesting packages that can be reused distribute the work and convert the files to npm modules and publish them for we should also see whether there are other already existing js projects i know of these two parsers but i haven t tested one of them i just start the list for gtf vcf bam bigwig please modify this list bmpvieira homonecloco saketkc
1
221,447
7,388,449,985
IssuesEvent
2018-03-16 02:44:31
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
Can't build EnergyPlus on Centos 6.6
Priority2 S1 - High
@Myoldmopar @lefticus I've tried building EnergyPlus 8.6 on Centos 6.6. The default compiler is gcc 4.4 which doesn't seem to support c++11. I get compiler crashes using gcc 4.7 and 4.8. When I tried to install clang 3.4 I get GLIBCXX_3.4.15, GLIBC_2.15 and GLIBC_2.14 not found. What compiler should I use?
1.0
Can't build EnergyPlus on Centos 6.6 - @Myoldmopar @lefticus I've tried building EnergyPlus 8.6 on Centos 6.6. The default compiler is gcc 4.4 which doesn't seem to support c++11. I get compiler crashes using gcc 4.7 and 4.8. When I tried to install clang 3.4 I get GLIBCXX_3.4.15, GLIBC_2.15 and GLIBC_2.14 not found. What compiler should I use?
priority
can t build energyplus on centos myoldmopar lefticus i ve tried building energyplus on centos the default compiler is gcc which doesn t seem to support c i get compiler crashes using gcc and when i tried to install clang i get glibcxx glibc and glibc not found what compiler should i use
1
376,920
11,158,163,683
IssuesEvent
2019-12-25 18:14:05
alexandrudsc/University-timetable-Android
https://api.github.com/repos/alexandrudsc/University-timetable-Android
closed
Delete button for downloaded timetables
enhancement high-priority
Since multiple timetables are supported, there is the need to delete one or multiple timetables. This should be simple, and should not affect any other functions. Two edge cases would be when the last timetable is deleted and when the default timetable is deleted (maybe the default timetable should not be deletable)
1.0
Delete button for downloaded timetables - Since multiple timetables are supported, there is the need to delete one or multiple timetables. This should be simple, and should not affect any other functions. Two edge cases would be when the last timetable is deleted and when the default timetable is deleted (maybe the default timetable should not be deletable)
priority
delete button for downloaded timetables since multiple timetables are supported there is the need to delete one or multiple timetables this should be simple and should not affect any other functions two edge cases would be when the last timetable is deleted and when the default timetable is deleted maybe the default timetable should not be deletable
1
509,201
14,724,053,876
IssuesEvent
2021-01-06 01:44:08
googlefonts/noto-fonts
https://api.github.com/repos/googlefonts/noto-fonts
closed
(U+FDF2) โ€œ๏ทฒโ€ ARABIC LIGATURE ALLAH ligature forming
Kufi Priority-High Script-Arabic
(U+FDF2) โ€œ๏ทฒโ€ ARABIC LIGATURE ALLAH ISOLATED FORM is a very common ligature, used to display the name of God FDF2 = (0627 0644 0644 <b>0651</b> <b>0670</b> 0647). Currently NotoSansArabic, NotoSansArabicUI and NotoKufiArabic will form the ligature FDF2 at least in the following combinations: 1- 0627 0644 0644 0647 โ€”> FDF2 2- 0627 0644 0644 <b>0651</b> 0647 โ€”> FDF2 3- 0627 0644 0644 <b>0651</b> <b>0670</b> 0647 โ€”> FDF2 The issue appears in the second and the third combination; <b>Second combaniation:</b> When 0651 ARABIC SHADDA (ู‘) exist in the original combaniation. The sequence will form the ligature FDF2 and add the original 0651 ARABIC SHADDA to FDF2, this will show duplicate 0651 SHADDA and misplace the original SHADDA <b>Third combaniation:</b> Acts same as the second combination but also will duplicate 0670 (ูฐ)ARABIC LETTER SUPERSCRIPT ALEF exist in the original combaniation and misplace both SUPERSCRIPT ALEF and the SHADDA 90% the Arabic text has the SHADDA in the original text. Adding another SHADDA and SUPERSCRIPT ALEF will make a serious issue. We should generally be conservative in those ligatures. My suggestions are: 1- Form combination only if SHADDA <b>or</b> SHADDA and SUPERSCRIPT ALEF exist in the original text and the very important part is dropping SHADDA and SUPERSCRIPT ALEF to avoid duplication 2- All Google Arabic fonts should act the same to maintain consistency to give users the flexibility to use all fonts without any issues <b>Issue found in:</b> NotoSansArabic 2.002 https://github.com/googlefonts/noto-fonts/tree/master/unhinted/ttf/NotoSansArabic/NotoSansArabic-Regular.ttf NotoSansArabicUI 2.002 https://github.com/googlefonts/noto-fonts/tree/master/unhinted/ttf/NotoSansArabicUI/NotoSansArabicUI-Regular.ttf NotoKufiArabic Version 2.101 https://github.com/googlefonts/noto-fonts/tree/master/unhinted/ttf/NotoKufiArabic/NotoKufiArabic.ttf <b>Characters data used in the test below:</b> 0627 0644 0644 0647 (No SHADDA or SUPERSCRIPT ALEF) 0627 0644 0644 0651 0647 (With SHADDA) 0627 0644 0644 0651 0670 0647 (With SHADDA and SUPERSCRIPT ALEF) <img width="1392" alt="Screen Shot 2020-10-02 at 11 19 28 PM" src="https://user-images.githubusercontent.com/22158882/94984875-baaff080-0505-11eb-8a84-b766e045de0a.png">
1.0
(U+FDF2) โ€œ๏ทฒโ€ ARABIC LIGATURE ALLAH ligature forming - (U+FDF2) โ€œ๏ทฒโ€ ARABIC LIGATURE ALLAH ISOLATED FORM is a very common ligature, used to display the name of God FDF2 = (0627 0644 0644 <b>0651</b> <b>0670</b> 0647). Currently NotoSansArabic, NotoSansArabicUI and NotoKufiArabic will form the ligature FDF2 at least in the following combinations: 1- 0627 0644 0644 0647 โ€”> FDF2 2- 0627 0644 0644 <b>0651</b> 0647 โ€”> FDF2 3- 0627 0644 0644 <b>0651</b> <b>0670</b> 0647 โ€”> FDF2 The issue appears in the second and the third combination; <b>Second combaniation:</b> When 0651 ARABIC SHADDA (ู‘) exist in the original combaniation. The sequence will form the ligature FDF2 and add the original 0651 ARABIC SHADDA to FDF2, this will show duplicate 0651 SHADDA and misplace the original SHADDA <b>Third combaniation:</b> Acts same as the second combination but also will duplicate 0670 (ูฐ)ARABIC LETTER SUPERSCRIPT ALEF exist in the original combaniation and misplace both SUPERSCRIPT ALEF and the SHADDA 90% the Arabic text has the SHADDA in the original text. Adding another SHADDA and SUPERSCRIPT ALEF will make a serious issue. We should generally be conservative in those ligatures. My suggestions are: 1- Form combination only if SHADDA <b>or</b> SHADDA and SUPERSCRIPT ALEF exist in the original text and the very important part is dropping SHADDA and SUPERSCRIPT ALEF to avoid duplication 2- All Google Arabic fonts should act the same to maintain consistency to give users the flexibility to use all fonts without any issues <b>Issue found in:</b> NotoSansArabic 2.002 https://github.com/googlefonts/noto-fonts/tree/master/unhinted/ttf/NotoSansArabic/NotoSansArabic-Regular.ttf NotoSansArabicUI 2.002 https://github.com/googlefonts/noto-fonts/tree/master/unhinted/ttf/NotoSansArabicUI/NotoSansArabicUI-Regular.ttf NotoKufiArabic Version 2.101 https://github.com/googlefonts/noto-fonts/tree/master/unhinted/ttf/NotoKufiArabic/NotoKufiArabic.ttf <b>Characters data used in the test below:</b> 0627 0644 0644 0647 (No SHADDA or SUPERSCRIPT ALEF) 0627 0644 0644 0651 0647 (With SHADDA) 0627 0644 0644 0651 0670 0647 (With SHADDA and SUPERSCRIPT ALEF) <img width="1392" alt="Screen Shot 2020-10-02 at 11 19 28 PM" src="https://user-images.githubusercontent.com/22158882/94984875-baaff080-0505-11eb-8a84-b766e045de0a.png">
priority
u โ€œ๏ทฒโ€ arabic ligature allah ligature forming u โ€œ๏ทฒโ€ arabic ligature allah isolated form is a very common ligature used to display the name of god currently notosansarabic notosansarabicui and notokufiarabic will form the ligature at least in the following combinations โ€” โ€” โ€” the issue appears in the second and the third combination second combaniation when arabic shadda ู‘ exist in the original combaniation the sequence will form the ligature and add the original arabic shadda to this will show duplicate shadda and misplace the original shadda third combaniation acts same as the second combination but also will duplicate ูฐ arabic letter superscript alef exist in the original combaniation and misplace both superscript alef and the shadda the arabic text has the shadda in the original text adding another shadda and superscript alef will make a serious issue we should generally be conservative in those ligatures my suggestions are form combination only if shadda or shadda and superscript alef exist in the original text and the very important part is dropping shadda and superscript alef to avoid duplication all google arabic fonts should act the same to maintain consistency to give users the flexibility to use all fonts without any issues issue found in notosansarabic notosansarabicui notokufiarabic version characters data used in the test below no shadda or superscript alef with shadda with shadda and superscript alef img width alt screen shot at pm src
1
356,989
10,600,402,850
IssuesEvent
2019-10-10 09:58:48
dotkom/onlineweb4
https://api.github.com/repos/dotkom/onlineweb4
closed
Atendee list bug when adding or removing users
Easy Priority: High
<!-- If this is a security issue or information leakage (having access to something you (probably) shouldn't), please send an email rather than opening a public issue. --> ## What kind of an issue is this? - Bug report ## What is the expected behaviour? It should look normal where every box and number is under its corresponding column. It should just look normal. Even if we remove users or add users through the dashboard, nothing should change. ![screenshot from 2018-09-25 21-07-06](https://user-images.githubusercontent.com/41551253/46037365-3b160c80-c108-11e8-9de5-91ef62505aeb.png) ## What is the current behaviour? If you remove or add a user through the dashboard menu at the attendee list it will look like the screenshot above. We have randomly two boxes, the x for removing users is all the way to the right and the text "none" is occupying the remove column. If you refresh the site it will go back to expected behaviour, its only after deleteing/adding a user ![screenshot from 2018-09-25 20-00-23](https://user-images.githubusercontent.com/41551253/46037343-29346980-c108-11e8-99e1-c69b77695d3e.png) <!-- if this is a bug report --> ## How do you reproduce this problem? Remove or add a user to the attendee list manually. <!-- if this is a bug report --> <!-- provide steps to reproduce this problem, preferably in a bullet point list --> 1. go to the attendee list 2. Add a user by writing their name OR remove a user from the list ## Other information This might be a bug which I didn't catch when I added "year of study" to the attendee list. I'm not sure if this was an issue before, but since it hasn't been brought up I will assume this is a bug from that pull request of mine
1.0
Atendee list bug when adding or removing users - <!-- If this is a security issue or information leakage (having access to something you (probably) shouldn't), please send an email rather than opening a public issue. --> ## What kind of an issue is this? - Bug report ## What is the expected behaviour? It should look normal where every box and number is under its corresponding column. It should just look normal. Even if we remove users or add users through the dashboard, nothing should change. ![screenshot from 2018-09-25 21-07-06](https://user-images.githubusercontent.com/41551253/46037365-3b160c80-c108-11e8-9de5-91ef62505aeb.png) ## What is the current behaviour? If you remove or add a user through the dashboard menu at the attendee list it will look like the screenshot above. We have randomly two boxes, the x for removing users is all the way to the right and the text "none" is occupying the remove column. If you refresh the site it will go back to expected behaviour, its only after deleteing/adding a user ![screenshot from 2018-09-25 20-00-23](https://user-images.githubusercontent.com/41551253/46037343-29346980-c108-11e8-99e1-c69b77695d3e.png) <!-- if this is a bug report --> ## How do you reproduce this problem? Remove or add a user to the attendee list manually. <!-- if this is a bug report --> <!-- provide steps to reproduce this problem, preferably in a bullet point list --> 1. go to the attendee list 2. Add a user by writing their name OR remove a user from the list ## Other information This might be a bug which I didn't catch when I added "year of study" to the attendee list. I'm not sure if this was an issue before, but since it hasn't been brought up I will assume this is a bug from that pull request of mine
priority
atendee list bug when adding or removing users what kind of an issue is this bug report what is the expected behaviour it should look normal where every box and number is under its corresponding column it should just look normal even if we remove users or add users through the dashboard nothing should change what is the current behaviour if you remove or add a user through the dashboard menu at the attendee list it will look like the screenshot above we have randomly two boxes the x for removing users is all the way to the right and the text none is occupying the remove column if you refresh the site it will go back to expected behaviour its only after deleteing adding a user how do you reproduce this problem remove or add a user to the attendee list manually go to the attendee list add a user by writing their name or remove a user from the list other information this might be a bug which i didn t catch when i added year of study to the attendee list i m not sure if this was an issue before but since it hasn t been brought up i will assume this is a bug from that pull request of mine
1
597,094
18,154,459,825
IssuesEvent
2021-09-26 20:44:11
nunit/vs-project-loader
https://api.github.com/repos/nunit/vs-project-loader
closed
Support multi-targeting files
High Priority Bug
I have a csproj which is targeting two versions of the .NET Framework. The vs-project-loader crashes with a NullReferenceException while looking for the `TargetFramework` element. (I have `TargetFrameworks` instead.) Sample csproj: ``` <Project Sdk="Microsoft.NET.Sdk"> <PropertyGroup> <TargetFrameworks>net20;net45</TargetFrameworks> </PropertyGroup> </Project> ``` This might be best solved eventually by: https://github.com/nunit/vs-project-loader/issues/25
1.0
Support multi-targeting files - I have a csproj which is targeting two versions of the .NET Framework. The vs-project-loader crashes with a NullReferenceException while looking for the `TargetFramework` element. (I have `TargetFrameworks` instead.) Sample csproj: ``` <Project Sdk="Microsoft.NET.Sdk"> <PropertyGroup> <TargetFrameworks>net20;net45</TargetFrameworks> </PropertyGroup> </Project> ``` This might be best solved eventually by: https://github.com/nunit/vs-project-loader/issues/25
priority
support multi targeting files i have a csproj which is targeting two versions of the net framework the vs project loader crashes with a nullreferenceexception while looking for the targetframework element i have targetframeworks instead sample csproj this might be best solved eventually by
1
592,361
17,876,579,726
IssuesEvent
2021-09-07 05:14:32
AyeCode/userswp
https://api.github.com/repos/AyeCode/userswp
opened
Doc for user role based redirect functionality
Priority: High Type: Needs Docs
We have added user role-based login and logout redirect with hiding the admin bar option in the new setting tab. Also, we have added restrict admin for user role setting in the login settings. We need a doc for the same.
1.0
Doc for user role based redirect functionality - We have added user role-based login and logout redirect with hiding the admin bar option in the new setting tab. Also, we have added restrict admin for user role setting in the login settings. We need a doc for the same.
priority
doc for user role based redirect functionality we have added user role based login and logout redirect with hiding the admin bar option in the new setting tab also we have added restrict admin for user role setting in the login settings we need a doc for the same
1
613,006
19,061,973,286
IssuesEvent
2021-11-26 09:02:52
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Incorrect order in anon function expression completion item sorting due to a bug in `ContextTypeResolver`
Type/Bug Type/Task Priority/High Team/LanguageServer Points/1 SwanLakeDump Area/Completion
**Description:** <!-- Give a brief description of the task --> Consider, <img width="878" alt="Screenshot 2021-10-21 at 11 47 46" src="https://user-images.githubusercontent.com/27485094/138222083-8018b182-c2e4-4cef-a9b9-9bf3387ef7dd.png"> `testFunc2` (which is assignable to cursor position) is assigned `sortText : D`, same as the variables `s` and `i`, whereas it should get a higher priority. **Describe your task(s)** This is due to the `ContextTypeResolver` returning the `signature` in `context.getContextType().get().signature()` incorrectly as `function (int i) returns function (int i, string s) returns int` when it should be `function (int i, string s) returns int`. Source code : ``` int i = 3; string s = "s"; function (int i, string s) returns int testFunc2 = function (int i, string s) returns int { return 3; }; function (int i) returns function (int i, string s) returns int testFunc = intVal => <cursor> ``` **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> Relates to #32879
1.0
Incorrect order in anon function expression completion item sorting due to a bug in `ContextTypeResolver` - **Description:** <!-- Give a brief description of the task --> Consider, <img width="878" alt="Screenshot 2021-10-21 at 11 47 46" src="https://user-images.githubusercontent.com/27485094/138222083-8018b182-c2e4-4cef-a9b9-9bf3387ef7dd.png"> `testFunc2` (which is assignable to cursor position) is assigned `sortText : D`, same as the variables `s` and `i`, whereas it should get a higher priority. **Describe your task(s)** This is due to the `ContextTypeResolver` returning the `signature` in `context.getContextType().get().signature()` incorrectly as `function (int i) returns function (int i, string s) returns int` when it should be `function (int i, string s) returns int`. Source code : ``` int i = 3; string s = "s"; function (int i, string s) returns int testFunc2 = function (int i, string s) returns int { return 3; }; function (int i) returns function (int i, string s) returns int testFunc = intVal => <cursor> ``` **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> Relates to #32879
priority
incorrect order in anon function expression completion item sorting due to a bug in contexttyperesolver description consider img width alt screenshot at src which is assignable to cursor position is assigned sorttext d same as the variables s and i whereas it should get a higher priority describe your task s this is due to the contexttyperesolver returning the signature in context getcontexttype get signature incorrectly as function int i returns function int i string s returns int when it should be function int i string s returns int source code int i string s s function int i string s returns int function int i string s returns int return function int i returns function int i string s returns int testfunc intval related issues optional relates to
1
525,258
15,242,013,966
IssuesEvent
2021-02-19 09:17:23
ubtue/DatenProbleme
https://api.github.com/repos/ubtue/DatenProbleme
opened
ISSN 2237-6461 | Estudos Teolรณgicos | Fehler beim Download
Fehlerquelle: unbekannt Zotero_SEMI-AUTO high priority
http://periodicos.est.edu.br/index.php/estudos_teologicos/issue/view/330 Es gelingt mir nicht, die Daten mir dem Open_Journal_Systems_Standard die Aufsรคtze nach Zotero zu รผbertragen. Am Ende bricht die รœbertragung mit der Fehlermeldung "Es ist ein Fehler aufgetreten ..." ab.
1.0
ISSN 2237-6461 | Estudos Teolรณgicos | Fehler beim Download - http://periodicos.est.edu.br/index.php/estudos_teologicos/issue/view/330 Es gelingt mir nicht, die Daten mir dem Open_Journal_Systems_Standard die Aufsรคtze nach Zotero zu รผbertragen. Am Ende bricht die รœbertragung mit der Fehlermeldung "Es ist ein Fehler aufgetreten ..." ab.
priority
issn estudos teolรณgicos fehler beim download es gelingt mir nicht die daten mir dem open journal systems standard die aufsรคtze nach zotero zu รผbertragen am ende bricht die รผbertragung mit der fehlermeldung es ist ein fehler aufgetreten ab
1
137,168
5,294,884,451
IssuesEvent
2017-02-09 12:13:20
cdnjs/cdnjs
https://api.github.com/repos/cdnjs/cdnjs
closed
[Request] validate.js
High Priority Library - Request to Add/Update
**Library name:** validate.js **Git repository url:** https://github.com/rickharrison/validate.js **npm package url(optional):** **License(s):** MIT License **Official homepage:** http://rickharrison.github.io/validate.js/ **Wanna say something? Leave message here:** Please add it soon :) ===================== Notes from cdnjs maintainer: Please read the README.md and CONTRIBUTING.md document first. You are welcome to add a library via sending pull request, it'll be faster than just opening a request issue, and please don't forget to read the guidelines for contributing, thanks!!
1.0
[Request] validate.js - **Library name:** validate.js **Git repository url:** https://github.com/rickharrison/validate.js **npm package url(optional):** **License(s):** MIT License **Official homepage:** http://rickharrison.github.io/validate.js/ **Wanna say something? Leave message here:** Please add it soon :) ===================== Notes from cdnjs maintainer: Please read the README.md and CONTRIBUTING.md document first. You are welcome to add a library via sending pull request, it'll be faster than just opening a request issue, and please don't forget to read the guidelines for contributing, thanks!!
priority
validate js library name validate js git repository url npm package url optional license s mit license official homepage wanna say something leave message here please add it soon notes from cdnjs maintainer please read the readme md and contributing md document first you are welcome to add a library via sending pull request it ll be faster than just opening a request issue and please don t forget to read the guidelines for contributing thanks
1
792,064
27,945,055,507
IssuesEvent
2023-03-24 01:53:22
ploomber/contributing
https://api.github.com/repos/ploomber/contributing
closed
migrating to jupyter-book 0.14
high priority
jupyter-book 0.14 introduced breaking API changes: https://github.com/executablebooks/jupyter-book/releases/tag/v0.14.0 this caused [jupysql](https://readthedocs.org/projects/jupysql/builds/19639344/) docs to fail, I believe it's because it got more picky and it's displaying warnings that didn't exist before: ``` /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:16: WARNING: 'myst' reference target not found: #list-connections /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:19: WARNING: 'myst' reference target not found: #close-connection /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:28: WARNING: 'myst' reference target not found: #create-table /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:31: WARNING: 'myst' reference target not found: #append-to-table /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:37: WARNING: 'myst' reference target not found: #run-query-from-file /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:43: WARNING: 'myst' reference target not found: #create-table-without-dataframe-index /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:49: WARNING: 'myst' reference target not found: #compose-large-queries /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:52: WARNING: 'myst' reference target not found: #compose-large-queries /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:58: WARNING: 'myst' reference target not found: #connect-to-database /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/plot.md:17: WARNING: 'myst' reference target not found: api/python.html#sql-plot /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/plot.md:130: WARNING: 'myst' reference target not found: #boxplot /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/quick-start.md:66: WARNING: 'myst' reference target not found: howto.html#switch-connections generating indices... genindex done ``` but there might be other issues as well repos to migrate (ordered by priority) * [x] jupysql https://github.com/ploomber/jupysql/pull/211 * [x] sklearn-evaluation * [x] ploomber-engine * [x] ploomber-core * [x] jupyblog * [ ] pkgmt there is a migration guide here: https://jupyterbook.org/en/stable/explain/migration.html
1.0
migrating to jupyter-book 0.14 - jupyter-book 0.14 introduced breaking API changes: https://github.com/executablebooks/jupyter-book/releases/tag/v0.14.0 this caused [jupysql](https://readthedocs.org/projects/jupysql/builds/19639344/) docs to fail, I believe it's because it got more picky and it's displaying warnings that didn't exist before: ``` /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:16: WARNING: 'myst' reference target not found: #list-connections /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:19: WARNING: 'myst' reference target not found: #close-connection /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:28: WARNING: 'myst' reference target not found: #create-table /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:31: WARNING: 'myst' reference target not found: #append-to-table /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:37: WARNING: 'myst' reference target not found: #run-query-from-file /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:43: WARNING: 'myst' reference target not found: #create-table-without-dataframe-index /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:49: WARNING: 'myst' reference target not found: #compose-large-queries /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:52: WARNING: 'myst' reference target not found: #compose-large-queries /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/api/magic-sql.md:58: WARNING: 'myst' reference target not found: #connect-to-database /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/plot.md:17: WARNING: 'myst' reference target not found: api/python.html#sql-plot /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/plot.md:130: WARNING: 'myst' reference target not found: #boxplot /home/docs/checkouts/readthedocs.org/user_builds/jupysql/checkouts/181/doc/quick-start.md:66: WARNING: 'myst' reference target not found: howto.html#switch-connections generating indices... genindex done ``` but there might be other issues as well repos to migrate (ordered by priority) * [x] jupysql https://github.com/ploomber/jupysql/pull/211 * [x] sklearn-evaluation * [x] ploomber-engine * [x] ploomber-core * [x] jupyblog * [ ] pkgmt there is a migration guide here: https://jupyterbook.org/en/stable/explain/migration.html
priority
migrating to jupyter book jupyter book introduced breaking api changes this caused docs to fail i believe it s because it got more picky and it s displaying warnings that didn t exist before home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found list connections home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found close connection home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found create table home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found append to table home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found run query from file home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found create table without dataframe index home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found compose large queries home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found compose large queries home docs checkouts readthedocs org user builds jupysql checkouts doc api magic sql md warning myst reference target not found connect to database home docs checkouts readthedocs org user builds jupysql checkouts doc plot md warning myst reference target not found api python html sql plot home docs checkouts readthedocs org user builds jupysql checkouts doc plot md warning myst reference target not found boxplot home docs checkouts readthedocs org user builds jupysql checkouts doc quick start md warning myst reference target not found howto html switch connections generating indices genindex done but there might be other issues as well repos to migrate ordered by priority jupysql sklearn evaluation ploomber engine ploomber core jupyblog pkgmt there is a migration guide here
1
499,309
14,444,553,471
IssuesEvent
2020-12-07 21:25:50
Eli017/rating-project-backend
https://api.github.com/repos/Eli017/rating-project-backend
opened
Categories Database Update
High Priority enhancement
## Description - Currently, reviews have set categories (frontend/backend). This needs to be flexible to the client. - By separating categories out into their own table and connecting reviews via foreign keys, then categories are more manageable. ## Task Verification - When the user sets up the database, the default `Seed.ts` reviews and the database schema must have a `Categories` table with the proper foreign key connections.
1.0
Categories Database Update - ## Description - Currently, reviews have set categories (frontend/backend). This needs to be flexible to the client. - By separating categories out into their own table and connecting reviews via foreign keys, then categories are more manageable. ## Task Verification - When the user sets up the database, the default `Seed.ts` reviews and the database schema must have a `Categories` table with the proper foreign key connections.
priority
categories database update description currently reviews have set categories frontend backend this needs to be flexible to the client by separating categories out into their own table and connecting reviews via foreign keys then categories are more manageable task verification when the user sets up the database the default seed ts reviews and the database schema must have a categories table with the proper foreign key connections
1
346,141
10,384,646,181
IssuesEvent
2019-09-10 12:29:21
webkom/lego
https://api.github.com/repos/webkom/lego
closed
Compact event list filters out alternative presentations
bug frontend priority:high
The compact event list on the front page does not include alternative presentations :(( Seems like it was simply forgotten when we added the new event type this summer! ![image](https://user-images.githubusercontent.com/8725291/64612638-94885180-d3d4-11e9-8530-a18db366386c.png) ![image](https://user-images.githubusercontent.com/8725291/64612682-a669f480-d3d4-11e9-8660-204b860e2f4c.png) https://github.com/webkom/lego-webapp/blob/0f955ee10b42a85726071a88fc34892d77cf9009/app/routes/overview/components/CompactEvents.js#L61
1.0
Compact event list filters out alternative presentations - The compact event list on the front page does not include alternative presentations :(( Seems like it was simply forgotten when we added the new event type this summer! ![image](https://user-images.githubusercontent.com/8725291/64612638-94885180-d3d4-11e9-8530-a18db366386c.png) ![image](https://user-images.githubusercontent.com/8725291/64612682-a669f480-d3d4-11e9-8660-204b860e2f4c.png) https://github.com/webkom/lego-webapp/blob/0f955ee10b42a85726071a88fc34892d77cf9009/app/routes/overview/components/CompactEvents.js#L61
priority
compact event list filters out alternative presentations the compact event list on the front page does not include alternative presentations seems like it was simply forgotten when we added the new event type this summer
1
524,268
15,209,979,117
IssuesEvent
2021-02-17 06:33:07
ballerina-platform/nballerina
https://api.github.com/repos/ballerina-platform/nballerina
closed
Add Mac OS build support
Points/3 Priority/High Type/Task
**Description:** Add support to build in Mac OS **Suggested Labels:** Type/Improvement,Priority/High **Suggested Assignees:**
1.0
Add Mac OS build support - **Description:** Add support to build in Mac OS **Suggested Labels:** Type/Improvement,Priority/High **Suggested Assignees:**
priority
add mac os build support description add support to build in mac os suggested labels type improvement priority high suggested assignees
1
389,744
11,516,538,215
IssuesEvent
2020-02-14 05:29:16
Novusphere/discussions-app
https://api.github.com/repos/Novusphere/discussions-app
closed
Link previews should use the icon specified by site json
bug high priority visual
i.e. https://discussions.app/tag/voice/3f0d4449n7hm5/voice_isnt_designed_for_you should have the icon `https://atmosdb.novusphere.io/discussions/upload/image/1580752273405.jpeg` as specified by http://atmosdb.novusphere.io/discussions/site
1.0
Link previews should use the icon specified by site json - i.e. https://discussions.app/tag/voice/3f0d4449n7hm5/voice_isnt_designed_for_you should have the icon `https://atmosdb.novusphere.io/discussions/upload/image/1580752273405.jpeg` as specified by http://atmosdb.novusphere.io/discussions/site
priority
link previews should use the icon specified by site json i e should have the icon as specified by
1
617,527
19,377,598,103
IssuesEvent
2021-12-17 00:56:37
geopm/geopm
https://api.github.com/repos/geopm/geopm
closed
test_geopmio.py failures
bug bug-priority-low bug-exposure-low bug-quality-high
**Describe the bug** I tried to run `test_geopmio.py` with the service installed and direct msr-safe access removed, and I expected the test to pass instead some of the tests in that file fail. **GEOPM version** `1.1.0+dev1674g69e2becc` https://github.com/bgeltz/geopm/tree/issue-2015 **Expected behavior** The test runs to completion with a return code of 0. **Actual behavior** ``` [nomsr@mcfly9 test](issue-2015)$ python3 ./test_geopmio.py -v > test_geopmio.log 2>&1 ``` Examining the log file: ``` [nomsr@mcfly test](issue-2015)$ cat test_geopmio.log test_geopmread_all_signal_agg (__main__.TestIntegrationGeopmio) ... ok test_geopmread_command_line (__main__.TestIntegrationGeopmio) ... ok test_geopmread_custom_msr (__main__.TestIntegrationGeopmio) ... FAIL test_geopmread_signal_value (__main__.TestIntegrationGeopmio) ... ok test_geopmwrite_command_line (__main__.TestIntegrationGeopmio) ... ok test_geopmwrite_set_freq (__main__.TestIntegrationGeopmio) ... stress-ng: info: [84867] defaulting to a 86400 second (1 day, 0.00 secs) run per stressor stress-ng: info: [84867] dispatching hogs: 88 cpu stress-ng: info: [84867] successful run completed in 9.22s FAIL ====================================================================== FAIL: test_geopmread_custom_msr (__main__.TestIntegrationGeopmio) ---------------------------------------------------------------------- Traceback (most recent call last): File "./test_geopmio.py", line 195, in test_geopmread_custom_msr self.assertIn(b'MSR::CORE_PERF_LIMIT_REASONS#', all_signals) AssertionError: b'MSR::CORE_PERF_LIMIT_REASONS#' not found in [b'CPUINFO::FREQ_MAX', b'CPUINFO::FREQ_MIN', b'CPUINFO::FREQ_STEP', b'CPUINFO::FREQ_STICKER', b'CPU_FREQUENCY_CONTROL', b'CPU_FREQUENCY_MAX', b'CPU_FREQUENCY_MIN', b'CPU_FREQUENCY_STATUS', b'CPU_FREQUENCY_STEP', b'CPU_FREQUENCY_STICKER', b'CYCLES_REFERENCE', b'CYCLES_THREAD', b'ENERGY_DRAM', b'ENERGY_PACKAGE', b'FREQUENCY', b'FREQUENCY_MAX', b'FREQUENCY_MIN', b'FREQUENCY_STEP', b'FREQUENCY_STICKER', b'INSTRUCTIONS_RETIRED', b'MSR::APERF#', b'MSR::APERF:ACNT', b'MSR::DRAM_ENERGY_STATUS#', b'MSR::DRAM_ENERGY_STATUS:ENERGY', b'MSR::DRAM_PERF_STATUS#', b'MSR::DRAM_PERF_STATUS:THROTTLE_TIME', b'MSR::DRAM_POWER_INFO#', b'MSR::DRAM_POWER_INFO:MAX_POWER', b'MSR::DRAM_POWER_INFO:MAX_TIME_WINDOW', b'MSR::DRAM_POWER_INFO:MIN_POWER', b'MSR::DRAM_POWER_INFO:THERMAL_SPEC_POWER', b'MSR::DRAM_POWER_LIMIT#', b'MSR::DRAM_POWER_LIMIT:ENABLE', b'MSR::DRAM_POWER_LIMIT:LOCK', b'MSR::DRAM_POWER_LIMIT:POWER_LIMIT', b'MSR::DRAM_POWER_LIMIT:TIME_WINDOW', b'MSR::FIXED_CTR0#', b'MSR::FIXED_CTR0:INST_RETIRED_ANY', b'MSR::FIXED_CTR1#', b'MSR::FIXED_CTR1:CPU_CLK_UNHALTED_THREAD', b'MSR::FIXED_CTR2#', b'MSR::FIXED_CTR2:CPU_CLK_UNHALTED_REF_TSC', b'MSR::FIXED_CTR_CTRL#', b'MSR::FIXED_CTR_CTRL:EN0_OS', b'MSR::FIXED_CTR_CTRL:EN0_PMI', b'MSR::FIXED_CTR_CTRL:EN0_USR', b'MSR::FIXED_CTR_CTRL:EN1_OS', b'MSR::FIXED_CTR_CTRL:EN1_PMI', b'MSR::FIXED_CTR_CTRL:EN1_USR', b'MSR::FIXED_CTR_CTRL:EN2_OS', b'MSR::FIXED_CTR_CTRL:EN2_PMI', b'MSR::FIXED_CTR_CTRL:EN2_USR', b'MSR::IA32_PERFEVTSEL0#', b'MSR::IA32_PERFEVTSEL0:ANYTHREAD', b'MSR::IA32_PERFEVTSEL0:CMASK', b'MSR::IA32_PERFEVTSEL0:EDGE', b'MSR::IA32_PERFEVTSEL0:EN', b'MSR::IA32_PERFEVTSEL0:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL0:INT', b'MSR::IA32_PERFEVTSEL0:INV', b'MSR::IA32_PERFEVTSEL0:OS', b'MSR::IA32_PERFEVTSEL0:PC', b'MSR::IA32_PERFEVTSEL0:UMASK', b'MSR::IA32_PERFEVTSEL0:USR', b'MSR::IA32_PERFEVTSEL1#', b'MSR::IA32_PERFEVTSEL1:ANYTHREAD', b'MSR::IA32_PERFEVTSEL1:CMASK', b'MSR::IA32_PERFEVTSEL1:EDGE', b'MSR::IA32_PERFEVTSEL1:EN', b'MSR::IA32_PERFEVTSEL1:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL1:INT', b'MSR::IA32_PERFEVTSEL1:INV', b'MSR::IA32_PERFEVTSEL1:OS', b'MSR::IA32_PERFEVTSEL1:PC', b'MSR::IA32_PERFEVTSEL1:UMASK', b'MSR::IA32_PERFEVTSEL1:USR', b'MSR::IA32_PERFEVTSEL2#', b'MSR::IA32_PERFEVTSEL2:ANYTHREAD', b'MSR::IA32_PERFEVTSEL2:CMASK', b'MSR::IA32_PERFEVTSEL2:EDGE', b'MSR::IA32_PERFEVTSEL2:EN', b'MSR::IA32_PERFEVTSEL2:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL2:INT', b'MSR::IA32_PERFEVTSEL2:INV', b'MSR::IA32_PERFEVTSEL2:OS', b'MSR::IA32_PERFEVTSEL2:PC', b'MSR::IA32_PERFEVTSEL2:UMASK', b'MSR::IA32_PERFEVTSEL2:USR', b'MSR::IA32_PERFEVTSEL3#', b'MSR::IA32_PERFEVTSEL3:ANYTHREAD', b'MSR::IA32_PERFEVTSEL3:CMASK', b'MSR::IA32_PERFEVTSEL3:EDGE', b'MSR::IA32_PERFEVTSEL3:EN', b'MSR::IA32_PERFEVTSEL3:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL3:INT', b'MSR::IA32_PERFEVTSEL3:INV', b'MSR::IA32_PERFEVTSEL3:OS', b'MSR::IA32_PERFEVTSEL3:PC', b'MSR::IA32_PERFEVTSEL3:UMASK', b'MSR::IA32_PERFEVTSEL3:USR', b'MSR::IA32_PMC0#', b'MSR::IA32_PMC0:PERFCTR', b'MSR::IA32_PMC1#', b'MSR::IA32_PMC1:PERFCTR', b'MSR::IA32_PMC2#', b'MSR::IA32_PMC2:PERFCTR', b'MSR::IA32_PMC3#', b'MSR::IA32_PMC3:PERFCTR', b'MSR::MISC_ENABLE#', b'MSR::MISC_ENABLE:ENHANCED_SPEEDSTEP_TECH_ENABLE', b'MSR::MISC_ENABLE:LIMIT_CPUID_MAXVAL', b'MSR::MISC_ENABLE:TURBO_MODE_DISABLE', b'MSR::MPERF#', b'MSR::MPERF:MCNT', b'MSR::PACKAGE_THERM_STATUS#', b'MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_LOG', b'MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_STATUS', b'MSR::PACKAGE_THERM_STATUS:DIGITAL_READOUT', b'MSR::PACKAGE_THERM_STATUS:POWER_LIMIT_STATUS', b'MSR::PACKAGE_THERM_STATUS:POWER_NOTIFICATION_LOG', b'MSR::PACKAGE_THERM_STATUS:PROCHOT_EVENT', b'MSR::PACKAGE_THERM_STATUS:PROCHOT_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_FLAG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_STATUS', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_STATUS', b'MSR::PERF_CTL#', b'MSR::PERF_CTL:FREQ', b'MSR::PERF_GLOBAL_CTRL#', b'MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR0', b'MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR1', b'MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR2', b'MSR::PERF_GLOBAL_CTRL:EN_PMC0', b'MSR::PERF_GLOBAL_CTRL:EN_PMC1', b'MSR::PERF_GLOBAL_CTRL:EN_PMC2', b'MSR::PERF_GLOBAL_CTRL:EN_PMC3', b'MSR::PERF_GLOBAL_OVF_CTRL#', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR0', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR1', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR2', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC0', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC1', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC2', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC3', b'MSR::PERF_STATUS#', b'MSR::PERF_STATUS:FREQ', b'MSR::PKG_ENERGY_STATUS#', b'MSR::PKG_ENERGY_STATUS:ENERGY', b'MSR::PKG_POWER_INFO#', b'MSR::PKG_POWER_INFO:MAX_POWER', b'MSR::PKG_POWER_INFO:MAX_TIME_WINDOW', b'MSR::PKG_POWER_INFO:MIN_POWER', b'MSR::PKG_POWER_INFO:THERMAL_SPEC_POWER', b'MSR::PKG_POWER_LIMIT#', b'MSR::PKG_POWER_LIMIT:LOCK', b'MSR::PKG_POWER_LIMIT:PL1_CLAMP_ENABLE', b'MSR::PKG_POWER_LIMIT:PL1_LIMIT_ENABLE', b'MSR::PKG_POWER_LIMIT:PL1_POWER_LIMIT', b'MSR::PKG_POWER_LIMIT:PL1_TIME_WINDOW', b'MSR::PKG_POWER_LIMIT:PL2_CLAMP_ENABLE', b'MSR::PKG_POWER_LIMIT:PL2_LIMIT_ENABLE', b'MSR::PKG_POWER_LIMIT:PL2_POWER_LIMIT', b'MSR::PKG_POWER_LIMIT:PL2_TIME_WINDOW', b'MSR::PLATFORM_INFO#', b'MSR::PLATFORM_INFO:MAX_EFFICIENCY_RATIO', b'MSR::PLATFORM_INFO:MAX_NON_TURBO_RATIO', b'MSR::PLATFORM_INFO:PROGRAMMABLE_RATIO_LIMITS_TURBO_MODE', b'MSR::PLATFORM_INFO:PROGRAMMABLE_TCC_ACTIVATION_OFFSET', b'MSR::PLATFORM_INFO:PROGRAMMABLE_TDP_LIMITS_TURBO_MODE', b'MSR::PPERF#', b'MSR::PPERF:PCNT', b'MSR::PQR_ASSOC#', b'MSR::PQR_ASSOC:RMID', b'MSR::QM_CTR#', b'MSR::QM_CTR:ERROR', b'MSR::QM_CTR:RM_DATA', b'MSR::QM_CTR:UNAVAILABLE', b'MSR::QM_EVTSEL#', b'MSR::QM_EVTSEL:EVENT_ID', b'MSR::QM_EVTSEL:RMID', b'MSR::RAPL_POWER_UNIT#', b'MSR::RAPL_POWER_UNIT:ENERGY', b'MSR::RAPL_POWER_UNIT:POWER', b'MSR::RAPL_POWER_UNIT:TIME', b'MSR::TEMPERATURE_TARGET#', b'MSR::TEMPERATURE_TARGET:PROCHOT_MIN', b'MSR::TEMPERATURE_TARGET:TCC_ACTIVE_OFFSET', b'MSR::THERM_STATUS#', b'MSR::THERM_STATUS:CRITICAL_TEMP_LOG', b'MSR::THERM_STATUS:CRITICAL_TEMP_STATUS', b'MSR::THERM_STATUS:DIGITAL_READOUT', b'MSR::THERM_STATUS:POWER_LIMIT_STATUS', b'MSR::THERM_STATUS:POWER_NOTIFICATION_LOG', b'MSR::THERM_STATUS:PROCHOT_EVENT', b'MSR::THERM_STATUS:PROCHOT_LOG', b'MSR::THERM_STATUS:READING_VALID', b'MSR::THERM_STATUS:RESOLUTION', b'MSR::THERM_STATUS:THERMAL_STATUS_FLAG', b'MSR::THERM_STATUS:THERMAL_STATUS_LOG', b'MSR::THERM_STATUS:THERMAL_THRESH_1_LOG', b'MSR::THERM_STATUS:THERMAL_THRESH_1_STATUS', b'MSR::THERM_STATUS:THERMAL_THRESH_2_LOG', b'MSR::THERM_STATUS:THERMAL_THRESH_2_STATUS', b'MSR::TIME', b'MSR::TIME_STAMP_COUNTER#', b'MSR::TIME_STAMP_COUNTER:TIMESTAMP_COUNT', b'MSR::TURBO_RATIO_LIMIT#', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_0', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_1', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_2', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_3', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_4', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_5', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_6', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_7', b'MSR::TURBO_RATIO_LIMIT_CORES#', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_0', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_1', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_2', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_3', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_4', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_5', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_6', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_7', b'MSR::UNCORE_PERF_STATUS#', b'MSR::UNCORE_PERF_STATUS:FREQ', b'MSR::UNCORE_RATIO_LIMIT#', b'MSR::UNCORE_RATIO_LIMIT:MAX_RATIO', b'MSR::UNCORE_RATIO_LIMIT:MIN_RATIO', b'POWER_DRAM', b'POWER_PACKAGE', b'POWER_PACKAGE_MAX', b'POWER_PACKAGE_MIN', b'POWER_PACKAGE_TDP', b'QM_CTR_SCALED', b'QM_CTR_SCALED_RATE', b'SERVICE::CPUINFO::FREQ_MAX', b'SERVICE::CPUINFO::FREQ_MIN', b'SERVICE::CPUINFO::FREQ_STEP', b'SERVICE::CPUINFO::FREQ_STICKER', b'SERVICE::CPU_FREQUENCY_CONTROL', b'SERVICE::CPU_FREQUENCY_MAX', b'SERVICE::CPU_FREQUENCY_MIN', b'SERVICE::CPU_FREQUENCY_STATUS', b'SERVICE::CPU_FREQUENCY_STEP', b'SERVICE::CPU_FREQUENCY_STICKER', b'SERVICE::CYCLES_REFERENCE', b'SERVICE::CYCLES_THREAD', b'SERVICE::ENERGY_DRAM', b'SERVICE::ENERGY_PACKAGE', b'SERVICE::FREQUENCY', b'SERVICE::FREQUENCY_MAX', b'SERVICE::FREQUENCY_MIN', b'SERVICE::FREQUENCY_STEP', b'SERVICE::FREQUENCY_STICKER', b'SERVICE::INSTRUCTIONS_RETIRED', b'SERVICE::MSR::APERF#', b'SERVICE::MSR::APERF:ACNT', b'SERVICE::MSR::DRAM_ENERGY_STATUS#', b'SERVICE::MSR::DRAM_ENERGY_STATUS:ENERGY', b'SERVICE::MSR::DRAM_PERF_STATUS#', b'SERVICE::MSR::DRAM_PERF_STATUS:THROTTLE_TIME', b'SERVICE::MSR::DRAM_POWER_INFO#', b'SERVICE::MSR::DRAM_POWER_INFO:MAX_POWER', b'SERVICE::MSR::DRAM_POWER_INFO:MAX_TIME_WINDOW', b'SERVICE::MSR::DRAM_POWER_INFO:MIN_POWER', b'SERVICE::MSR::DRAM_POWER_INFO:THERMAL_SPEC_POWER', b'SERVICE::MSR::DRAM_POWER_LIMIT#', b'SERVICE::MSR::DRAM_POWER_LIMIT:ENABLE', b'SERVICE::MSR::DRAM_POWER_LIMIT:LOCK', b'SERVICE::MSR::DRAM_POWER_LIMIT:POWER_LIMIT', b'SERVICE::MSR::DRAM_POWER_LIMIT:TIME_WINDOW', b'SERVICE::MSR::FIXED_CTR0#', b'SERVICE::MSR::FIXED_CTR0:INST_RETIRED_ANY', b'SERVICE::MSR::FIXED_CTR1#', b'SERVICE::MSR::FIXED_CTR1:CPU_CLK_UNHALTED_THREAD', b'SERVICE::MSR::FIXED_CTR2#', b'SERVICE::MSR::FIXED_CTR2:CPU_CLK_UNHALTED_REF_TSC', b'SERVICE::MSR::FIXED_CTR_CTRL#', b'SERVICE::MSR::FIXED_CTR_CTRL:EN0_OS', b'SERVICE::MSR::FIXED_CTR_CTRL:EN0_PMI', b'SERVICE::MSR::FIXED_CTR_CTRL:EN0_USR', b'SERVICE::MSR::FIXED_CTR_CTRL:EN1_OS', b'SERVICE::MSR::FIXED_CTR_CTRL:EN1_PMI', b'SERVICE::MSR::FIXED_CTR_CTRL:EN1_USR', b'SERVICE::MSR::FIXED_CTR_CTRL:EN2_OS', b'SERVICE::MSR::FIXED_CTR_CTRL:EN2_PMI', b'SERVICE::MSR::FIXED_CTR_CTRL:EN2_USR', b'SERVICE::MSR::IA32_PERFEVTSEL0#', b'SERVICE::MSR::IA32_PERFEVTSEL0:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL0:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL0:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL0:EN', b'SERVICE::MSR::IA32_PERFEVTSEL0:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL0:INT', b'SERVICE::MSR::IA32_PERFEVTSEL0:INV', b'SERVICE::MSR::IA32_PERFEVTSEL0:OS', b'SERVICE::MSR::IA32_PERFEVTSEL0:PC', b'SERVICE::MSR::IA32_PERFEVTSEL0:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL0:USR', b'SERVICE::MSR::IA32_PERFEVTSEL1#', b'SERVICE::MSR::IA32_PERFEVTSEL1:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL1:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL1:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL1:EN', b'SERVICE::MSR::IA32_PERFEVTSEL1:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL1:INT', b'SERVICE::MSR::IA32_PERFEVTSEL1:INV', b'SERVICE::MSR::IA32_PERFEVTSEL1:OS', b'SERVICE::MSR::IA32_PERFEVTSEL1:PC', b'SERVICE::MSR::IA32_PERFEVTSEL1:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL1:USR', b'SERVICE::MSR::IA32_PERFEVTSEL2#', b'SERVICE::MSR::IA32_PERFEVTSEL2:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL2:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL2:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL2:EN', b'SERVICE::MSR::IA32_PERFEVTSEL2:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL2:INT', b'SERVICE::MSR::IA32_PERFEVTSEL2:INV', b'SERVICE::MSR::IA32_PERFEVTSEL2:OS', b'SERVICE::MSR::IA32_PERFEVTSEL2:PC', b'SERVICE::MSR::IA32_PERFEVTSEL2:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL2:USR', b'SERVICE::MSR::IA32_PERFEVTSEL3#', b'SERVICE::MSR::IA32_PERFEVTSEL3:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL3:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL3:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL3:EN', b'SERVICE::MSR::IA32_PERFEVTSEL3:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL3:INT', b'SERVICE::MSR::IA32_PERFEVTSEL3:INV', b'SERVICE::MSR::IA32_PERFEVTSEL3:OS', b'SERVICE::MSR::IA32_PERFEVTSEL3:PC', b'SERVICE::MSR::IA32_PERFEVTSEL3:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL3:USR', b'SERVICE::MSR::IA32_PMC0#', b'SERVICE::MSR::IA32_PMC0:PERFCTR', b'SERVICE::MSR::IA32_PMC1#', b'SERVICE::MSR::IA32_PMC1:PERFCTR', b'SERVICE::MSR::IA32_PMC2#', b'SERVICE::MSR::IA32_PMC2:PERFCTR', b'SERVICE::MSR::IA32_PMC3#', b'SERVICE::MSR::IA32_PMC3:PERFCTR', b'SERVICE::MSR::MISC_ENABLE#', b'SERVICE::MSR::MISC_ENABLE:ENHANCED_SPEEDSTEP_TECH_ENABLE', b'SERVICE::MSR::MISC_ENABLE:LIMIT_CPUID_MAXVAL', b'SERVICE::MSR::MISC_ENABLE:TURBO_MODE_DISABLE', b'SERVICE::MSR::MPERF#', b'SERVICE::MSR::MPERF:MCNT', b'SERVICE::MSR::PACKAGE_THERM_STATUS#', b'SERVICE::MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_STATUS', b'SERVICE::MSR::PACKAGE_THERM_STATUS:DIGITAL_READOUT', b'SERVICE::MSR::PACKAGE_THERM_STATUS:POWER_LIMIT_STATUS', b'SERVICE::MSR::PACKAGE_THERM_STATUS:POWER_NOTIFICATION_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:PROCHOT_EVENT', b'SERVICE::MSR::PACKAGE_THERM_STATUS:PROCHOT_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_FLAG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_STATUS', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_STATUS', b'SERVICE::MSR::PERF_CTL#', b'SERVICE::MSR::PERF_CTL:FREQ', b'SERVICE::MSR::PERF_GLOBAL_CTRL#', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR0', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR1', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR2', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC0', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC1', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC2', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC3', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL#', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR0', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR1', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR2', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC0', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC1', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC2', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC3', b'SERVICE::MSR::PERF_STATUS#', b'SERVICE::MSR::PERF_STATUS:FREQ', b'SERVICE::MSR::PKG_ENERGY_STATUS#', b'SERVICE::MSR::PKG_ENERGY_STATUS:ENERGY', b'SERVICE::MSR::PKG_POWER_INFO#', b'SERVICE::MSR::PKG_POWER_INFO:MAX_POWER', b'SERVICE::MSR::PKG_POWER_INFO:MAX_TIME_WINDOW', b'SERVICE::MSR::PKG_POWER_INFO:MIN_POWER', b'SERVICE::MSR::PKG_POWER_INFO:THERMAL_SPEC_POWER', b'SERVICE::MSR::PKG_POWER_LIMIT#', b'SERVICE::MSR::PKG_POWER_LIMIT:LOCK', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_CLAMP_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_LIMIT_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_POWER_LIMIT', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_TIME_WINDOW', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_CLAMP_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_LIMIT_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_POWER_LIMIT', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_TIME_WINDOW', b'SERVICE::MSR::PLATFORM_INFO#', b'SERVICE::MSR::PLATFORM_INFO:MAX_EFFICIENCY_RATIO', b'SERVICE::MSR::PLATFORM_INFO:MAX_NON_TURBO_RATIO', b'SERVICE::MSR::PLATFORM_INFO:PROGRAMMABLE_RATIO_LIMITS_TURBO_MODE', b'SERVICE::MSR::PLATFORM_INFO:PROGRAMMABLE_TCC_ACTIVATION_OFFSET', b'SERVICE::MSR::PLATFORM_INFO:PROGRAMMABLE_TDP_LIMITS_TURBO_MODE', b'SERVICE::MSR::PPERF#', b'SERVICE::MSR::PPERF:PCNT', b'SERVICE::MSR::PQR_ASSOC#', b'SERVICE::MSR::PQR_ASSOC:RMID', b'SERVICE::MSR::QM_CTR#', b'SERVICE::MSR::QM_CTR:ERROR', b'SERVICE::MSR::QM_CTR:RM_DATA', b'SERVICE::MSR::QM_CTR:UNAVAILABLE', b'SERVICE::MSR::QM_EVTSEL#', b'SERVICE::MSR::QM_EVTSEL:EVENT_ID', b'SERVICE::MSR::QM_EVTSEL:RMID', b'SERVICE::MSR::RAPL_POWER_UNIT#', b'SERVICE::MSR::RAPL_POWER_UNIT:ENERGY', b'SERVICE::MSR::RAPL_POWER_UNIT:POWER', b'SERVICE::MSR::RAPL_POWER_UNIT:TIME', b'SERVICE::MSR::TEMPERATURE_TARGET#', b'SERVICE::MSR::TEMPERATURE_TARGET:PROCHOT_MIN', b'SERVICE::MSR::TEMPERATURE_TARGET:TCC_ACTIVE_OFFSET', b'SERVICE::MSR::THERM_STATUS#', b'SERVICE::MSR::THERM_STATUS:CRITICAL_TEMP_LOG', b'SERVICE::MSR::THERM_STATUS:CRITICAL_TEMP_STATUS', b'SERVICE::MSR::THERM_STATUS:DIGITAL_READOUT', b'SERVICE::MSR::THERM_STATUS:POWER_LIMIT_STATUS', b'SERVICE::MSR::THERM_STATUS:POWER_NOTIFICATION_LOG', b'SERVICE::MSR::THERM_STATUS:PROCHOT_EVENT', b'SERVICE::MSR::THERM_STATUS:PROCHOT_LOG', b'SERVICE::MSR::THERM_STATUS:READING_VALID', b'SERVICE::MSR::THERM_STATUS:RESOLUTION', b'SERVICE::MSR::THERM_STATUS:THERMAL_STATUS_FLAG', b'SERVICE::MSR::THERM_STATUS:THERMAL_STATUS_LOG', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_1_LOG', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_1_STATUS', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_2_LOG', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_2_STATUS', b'SERVICE::MSR::TIME', b'SERVICE::MSR::TIME_STAMP_COUNTER#', b'SERVICE::MSR::TIME_STAMP_COUNTER:TIMESTAMP_COUNT', b'SERVICE::MSR::TURBO_RATIO_LIMIT#', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_0', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_1', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_2', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_3', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_4', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_5', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_6', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_7', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES#', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_0', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_1', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_2', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_3', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_4', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_5', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_6', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_7', b'SERVICE::MSR::UNCORE_PERF_STATUS#', b'SERVICE::MSR::UNCORE_PERF_STATUS:FREQ', b'SERVICE::MSR::UNCORE_RATIO_LIMIT#', b'SERVICE::MSR::UNCORE_RATIO_LIMIT:MAX_RATIO', b'SERVICE::MSR::UNCORE_RATIO_LIMIT:MIN_RATIO', b'SERVICE::POWER_DRAM', b'SERVICE::POWER_PACKAGE', b'SERVICE::POWER_PACKAGE_MAX', b'SERVICE::POWER_PACKAGE_MIN', b'SERVICE::POWER_PACKAGE_TDP', b'SERVICE::QM_CTR_SCALED', b'SERVICE::QM_CTR_SCALED_RATE', b'SERVICE::TEMPERATURE_CORE', b'SERVICE::TEMPERATURE_PACKAGE', b'SERVICE::TIME', b'SERVICE::TIME::ELAPSED', b'SERVICE::TIMESTAMP_COUNTER', b'TEMPERATURE_CORE', b'TEMPERATURE_PACKAGE', b'TIME', b'TIME::ELAPSED', b'TIMESTAMP_COUNTER'] ====================================================================== FAIL: test_geopmwrite_set_freq (__main__.TestIntegrationGeopmio) ---------------------------------------------------------------------- Traceback (most recent call last): File "./test_geopmio.py", line 268, in test_geopmwrite_set_freq self.assertEqual(min_freq, result) AssertionError: 1000000000.0 != 2800000000.0 ---------------------------------------------------------------------- Ran 6 tests in 209.203s FAILED (failures=2) ``` **Additional context** There are 2 failures above: 1. The custom MSR plugin is not working. 2. `geopmwrite` is not able to set the frequency
1.0
test_geopmio.py failures - **Describe the bug** I tried to run `test_geopmio.py` with the service installed and direct msr-safe access removed, and I expected the test to pass instead some of the tests in that file fail. **GEOPM version** `1.1.0+dev1674g69e2becc` https://github.com/bgeltz/geopm/tree/issue-2015 **Expected behavior** The test runs to completion with a return code of 0. **Actual behavior** ``` [nomsr@mcfly9 test](issue-2015)$ python3 ./test_geopmio.py -v > test_geopmio.log 2>&1 ``` Examining the log file: ``` [nomsr@mcfly test](issue-2015)$ cat test_geopmio.log test_geopmread_all_signal_agg (__main__.TestIntegrationGeopmio) ... ok test_geopmread_command_line (__main__.TestIntegrationGeopmio) ... ok test_geopmread_custom_msr (__main__.TestIntegrationGeopmio) ... FAIL test_geopmread_signal_value (__main__.TestIntegrationGeopmio) ... ok test_geopmwrite_command_line (__main__.TestIntegrationGeopmio) ... ok test_geopmwrite_set_freq (__main__.TestIntegrationGeopmio) ... stress-ng: info: [84867] defaulting to a 86400 second (1 day, 0.00 secs) run per stressor stress-ng: info: [84867] dispatching hogs: 88 cpu stress-ng: info: [84867] successful run completed in 9.22s FAIL ====================================================================== FAIL: test_geopmread_custom_msr (__main__.TestIntegrationGeopmio) ---------------------------------------------------------------------- Traceback (most recent call last): File "./test_geopmio.py", line 195, in test_geopmread_custom_msr self.assertIn(b'MSR::CORE_PERF_LIMIT_REASONS#', all_signals) AssertionError: b'MSR::CORE_PERF_LIMIT_REASONS#' not found in [b'CPUINFO::FREQ_MAX', b'CPUINFO::FREQ_MIN', b'CPUINFO::FREQ_STEP', b'CPUINFO::FREQ_STICKER', b'CPU_FREQUENCY_CONTROL', b'CPU_FREQUENCY_MAX', b'CPU_FREQUENCY_MIN', b'CPU_FREQUENCY_STATUS', b'CPU_FREQUENCY_STEP', b'CPU_FREQUENCY_STICKER', b'CYCLES_REFERENCE', b'CYCLES_THREAD', b'ENERGY_DRAM', b'ENERGY_PACKAGE', b'FREQUENCY', b'FREQUENCY_MAX', b'FREQUENCY_MIN', b'FREQUENCY_STEP', b'FREQUENCY_STICKER', b'INSTRUCTIONS_RETIRED', b'MSR::APERF#', b'MSR::APERF:ACNT', b'MSR::DRAM_ENERGY_STATUS#', b'MSR::DRAM_ENERGY_STATUS:ENERGY', b'MSR::DRAM_PERF_STATUS#', b'MSR::DRAM_PERF_STATUS:THROTTLE_TIME', b'MSR::DRAM_POWER_INFO#', b'MSR::DRAM_POWER_INFO:MAX_POWER', b'MSR::DRAM_POWER_INFO:MAX_TIME_WINDOW', b'MSR::DRAM_POWER_INFO:MIN_POWER', b'MSR::DRAM_POWER_INFO:THERMAL_SPEC_POWER', b'MSR::DRAM_POWER_LIMIT#', b'MSR::DRAM_POWER_LIMIT:ENABLE', b'MSR::DRAM_POWER_LIMIT:LOCK', b'MSR::DRAM_POWER_LIMIT:POWER_LIMIT', b'MSR::DRAM_POWER_LIMIT:TIME_WINDOW', b'MSR::FIXED_CTR0#', b'MSR::FIXED_CTR0:INST_RETIRED_ANY', b'MSR::FIXED_CTR1#', b'MSR::FIXED_CTR1:CPU_CLK_UNHALTED_THREAD', b'MSR::FIXED_CTR2#', b'MSR::FIXED_CTR2:CPU_CLK_UNHALTED_REF_TSC', b'MSR::FIXED_CTR_CTRL#', b'MSR::FIXED_CTR_CTRL:EN0_OS', b'MSR::FIXED_CTR_CTRL:EN0_PMI', b'MSR::FIXED_CTR_CTRL:EN0_USR', b'MSR::FIXED_CTR_CTRL:EN1_OS', b'MSR::FIXED_CTR_CTRL:EN1_PMI', b'MSR::FIXED_CTR_CTRL:EN1_USR', b'MSR::FIXED_CTR_CTRL:EN2_OS', b'MSR::FIXED_CTR_CTRL:EN2_PMI', b'MSR::FIXED_CTR_CTRL:EN2_USR', b'MSR::IA32_PERFEVTSEL0#', b'MSR::IA32_PERFEVTSEL0:ANYTHREAD', b'MSR::IA32_PERFEVTSEL0:CMASK', b'MSR::IA32_PERFEVTSEL0:EDGE', b'MSR::IA32_PERFEVTSEL0:EN', b'MSR::IA32_PERFEVTSEL0:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL0:INT', b'MSR::IA32_PERFEVTSEL0:INV', b'MSR::IA32_PERFEVTSEL0:OS', b'MSR::IA32_PERFEVTSEL0:PC', b'MSR::IA32_PERFEVTSEL0:UMASK', b'MSR::IA32_PERFEVTSEL0:USR', b'MSR::IA32_PERFEVTSEL1#', b'MSR::IA32_PERFEVTSEL1:ANYTHREAD', b'MSR::IA32_PERFEVTSEL1:CMASK', b'MSR::IA32_PERFEVTSEL1:EDGE', b'MSR::IA32_PERFEVTSEL1:EN', b'MSR::IA32_PERFEVTSEL1:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL1:INT', b'MSR::IA32_PERFEVTSEL1:INV', b'MSR::IA32_PERFEVTSEL1:OS', b'MSR::IA32_PERFEVTSEL1:PC', b'MSR::IA32_PERFEVTSEL1:UMASK', b'MSR::IA32_PERFEVTSEL1:USR', b'MSR::IA32_PERFEVTSEL2#', b'MSR::IA32_PERFEVTSEL2:ANYTHREAD', b'MSR::IA32_PERFEVTSEL2:CMASK', b'MSR::IA32_PERFEVTSEL2:EDGE', b'MSR::IA32_PERFEVTSEL2:EN', b'MSR::IA32_PERFEVTSEL2:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL2:INT', b'MSR::IA32_PERFEVTSEL2:INV', b'MSR::IA32_PERFEVTSEL2:OS', b'MSR::IA32_PERFEVTSEL2:PC', b'MSR::IA32_PERFEVTSEL2:UMASK', b'MSR::IA32_PERFEVTSEL2:USR', b'MSR::IA32_PERFEVTSEL3#', b'MSR::IA32_PERFEVTSEL3:ANYTHREAD', b'MSR::IA32_PERFEVTSEL3:CMASK', b'MSR::IA32_PERFEVTSEL3:EDGE', b'MSR::IA32_PERFEVTSEL3:EN', b'MSR::IA32_PERFEVTSEL3:EVENT_SELECT', b'MSR::IA32_PERFEVTSEL3:INT', b'MSR::IA32_PERFEVTSEL3:INV', b'MSR::IA32_PERFEVTSEL3:OS', b'MSR::IA32_PERFEVTSEL3:PC', b'MSR::IA32_PERFEVTSEL3:UMASK', b'MSR::IA32_PERFEVTSEL3:USR', b'MSR::IA32_PMC0#', b'MSR::IA32_PMC0:PERFCTR', b'MSR::IA32_PMC1#', b'MSR::IA32_PMC1:PERFCTR', b'MSR::IA32_PMC2#', b'MSR::IA32_PMC2:PERFCTR', b'MSR::IA32_PMC3#', b'MSR::IA32_PMC3:PERFCTR', b'MSR::MISC_ENABLE#', b'MSR::MISC_ENABLE:ENHANCED_SPEEDSTEP_TECH_ENABLE', b'MSR::MISC_ENABLE:LIMIT_CPUID_MAXVAL', b'MSR::MISC_ENABLE:TURBO_MODE_DISABLE', b'MSR::MPERF#', b'MSR::MPERF:MCNT', b'MSR::PACKAGE_THERM_STATUS#', b'MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_LOG', b'MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_STATUS', b'MSR::PACKAGE_THERM_STATUS:DIGITAL_READOUT', b'MSR::PACKAGE_THERM_STATUS:POWER_LIMIT_STATUS', b'MSR::PACKAGE_THERM_STATUS:POWER_NOTIFICATION_LOG', b'MSR::PACKAGE_THERM_STATUS:PROCHOT_EVENT', b'MSR::PACKAGE_THERM_STATUS:PROCHOT_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_FLAG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_STATUS', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_LOG', b'MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_STATUS', b'MSR::PERF_CTL#', b'MSR::PERF_CTL:FREQ', b'MSR::PERF_GLOBAL_CTRL#', b'MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR0', b'MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR1', b'MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR2', b'MSR::PERF_GLOBAL_CTRL:EN_PMC0', b'MSR::PERF_GLOBAL_CTRL:EN_PMC1', b'MSR::PERF_GLOBAL_CTRL:EN_PMC2', b'MSR::PERF_GLOBAL_CTRL:EN_PMC3', b'MSR::PERF_GLOBAL_OVF_CTRL#', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR0', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR1', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR2', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC0', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC1', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC2', b'MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC3', b'MSR::PERF_STATUS#', b'MSR::PERF_STATUS:FREQ', b'MSR::PKG_ENERGY_STATUS#', b'MSR::PKG_ENERGY_STATUS:ENERGY', b'MSR::PKG_POWER_INFO#', b'MSR::PKG_POWER_INFO:MAX_POWER', b'MSR::PKG_POWER_INFO:MAX_TIME_WINDOW', b'MSR::PKG_POWER_INFO:MIN_POWER', b'MSR::PKG_POWER_INFO:THERMAL_SPEC_POWER', b'MSR::PKG_POWER_LIMIT#', b'MSR::PKG_POWER_LIMIT:LOCK', b'MSR::PKG_POWER_LIMIT:PL1_CLAMP_ENABLE', b'MSR::PKG_POWER_LIMIT:PL1_LIMIT_ENABLE', b'MSR::PKG_POWER_LIMIT:PL1_POWER_LIMIT', b'MSR::PKG_POWER_LIMIT:PL1_TIME_WINDOW', b'MSR::PKG_POWER_LIMIT:PL2_CLAMP_ENABLE', b'MSR::PKG_POWER_LIMIT:PL2_LIMIT_ENABLE', b'MSR::PKG_POWER_LIMIT:PL2_POWER_LIMIT', b'MSR::PKG_POWER_LIMIT:PL2_TIME_WINDOW', b'MSR::PLATFORM_INFO#', b'MSR::PLATFORM_INFO:MAX_EFFICIENCY_RATIO', b'MSR::PLATFORM_INFO:MAX_NON_TURBO_RATIO', b'MSR::PLATFORM_INFO:PROGRAMMABLE_RATIO_LIMITS_TURBO_MODE', b'MSR::PLATFORM_INFO:PROGRAMMABLE_TCC_ACTIVATION_OFFSET', b'MSR::PLATFORM_INFO:PROGRAMMABLE_TDP_LIMITS_TURBO_MODE', b'MSR::PPERF#', b'MSR::PPERF:PCNT', b'MSR::PQR_ASSOC#', b'MSR::PQR_ASSOC:RMID', b'MSR::QM_CTR#', b'MSR::QM_CTR:ERROR', b'MSR::QM_CTR:RM_DATA', b'MSR::QM_CTR:UNAVAILABLE', b'MSR::QM_EVTSEL#', b'MSR::QM_EVTSEL:EVENT_ID', b'MSR::QM_EVTSEL:RMID', b'MSR::RAPL_POWER_UNIT#', b'MSR::RAPL_POWER_UNIT:ENERGY', b'MSR::RAPL_POWER_UNIT:POWER', b'MSR::RAPL_POWER_UNIT:TIME', b'MSR::TEMPERATURE_TARGET#', b'MSR::TEMPERATURE_TARGET:PROCHOT_MIN', b'MSR::TEMPERATURE_TARGET:TCC_ACTIVE_OFFSET', b'MSR::THERM_STATUS#', b'MSR::THERM_STATUS:CRITICAL_TEMP_LOG', b'MSR::THERM_STATUS:CRITICAL_TEMP_STATUS', b'MSR::THERM_STATUS:DIGITAL_READOUT', b'MSR::THERM_STATUS:POWER_LIMIT_STATUS', b'MSR::THERM_STATUS:POWER_NOTIFICATION_LOG', b'MSR::THERM_STATUS:PROCHOT_EVENT', b'MSR::THERM_STATUS:PROCHOT_LOG', b'MSR::THERM_STATUS:READING_VALID', b'MSR::THERM_STATUS:RESOLUTION', b'MSR::THERM_STATUS:THERMAL_STATUS_FLAG', b'MSR::THERM_STATUS:THERMAL_STATUS_LOG', b'MSR::THERM_STATUS:THERMAL_THRESH_1_LOG', b'MSR::THERM_STATUS:THERMAL_THRESH_1_STATUS', b'MSR::THERM_STATUS:THERMAL_THRESH_2_LOG', b'MSR::THERM_STATUS:THERMAL_THRESH_2_STATUS', b'MSR::TIME', b'MSR::TIME_STAMP_COUNTER#', b'MSR::TIME_STAMP_COUNTER:TIMESTAMP_COUNT', b'MSR::TURBO_RATIO_LIMIT#', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_0', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_1', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_2', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_3', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_4', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_5', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_6', b'MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_7', b'MSR::TURBO_RATIO_LIMIT_CORES#', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_0', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_1', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_2', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_3', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_4', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_5', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_6', b'MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_7', b'MSR::UNCORE_PERF_STATUS#', b'MSR::UNCORE_PERF_STATUS:FREQ', b'MSR::UNCORE_RATIO_LIMIT#', b'MSR::UNCORE_RATIO_LIMIT:MAX_RATIO', b'MSR::UNCORE_RATIO_LIMIT:MIN_RATIO', b'POWER_DRAM', b'POWER_PACKAGE', b'POWER_PACKAGE_MAX', b'POWER_PACKAGE_MIN', b'POWER_PACKAGE_TDP', b'QM_CTR_SCALED', b'QM_CTR_SCALED_RATE', b'SERVICE::CPUINFO::FREQ_MAX', b'SERVICE::CPUINFO::FREQ_MIN', b'SERVICE::CPUINFO::FREQ_STEP', b'SERVICE::CPUINFO::FREQ_STICKER', b'SERVICE::CPU_FREQUENCY_CONTROL', b'SERVICE::CPU_FREQUENCY_MAX', b'SERVICE::CPU_FREQUENCY_MIN', b'SERVICE::CPU_FREQUENCY_STATUS', b'SERVICE::CPU_FREQUENCY_STEP', b'SERVICE::CPU_FREQUENCY_STICKER', b'SERVICE::CYCLES_REFERENCE', b'SERVICE::CYCLES_THREAD', b'SERVICE::ENERGY_DRAM', b'SERVICE::ENERGY_PACKAGE', b'SERVICE::FREQUENCY', b'SERVICE::FREQUENCY_MAX', b'SERVICE::FREQUENCY_MIN', b'SERVICE::FREQUENCY_STEP', b'SERVICE::FREQUENCY_STICKER', b'SERVICE::INSTRUCTIONS_RETIRED', b'SERVICE::MSR::APERF#', b'SERVICE::MSR::APERF:ACNT', b'SERVICE::MSR::DRAM_ENERGY_STATUS#', b'SERVICE::MSR::DRAM_ENERGY_STATUS:ENERGY', b'SERVICE::MSR::DRAM_PERF_STATUS#', b'SERVICE::MSR::DRAM_PERF_STATUS:THROTTLE_TIME', b'SERVICE::MSR::DRAM_POWER_INFO#', b'SERVICE::MSR::DRAM_POWER_INFO:MAX_POWER', b'SERVICE::MSR::DRAM_POWER_INFO:MAX_TIME_WINDOW', b'SERVICE::MSR::DRAM_POWER_INFO:MIN_POWER', b'SERVICE::MSR::DRAM_POWER_INFO:THERMAL_SPEC_POWER', b'SERVICE::MSR::DRAM_POWER_LIMIT#', b'SERVICE::MSR::DRAM_POWER_LIMIT:ENABLE', b'SERVICE::MSR::DRAM_POWER_LIMIT:LOCK', b'SERVICE::MSR::DRAM_POWER_LIMIT:POWER_LIMIT', b'SERVICE::MSR::DRAM_POWER_LIMIT:TIME_WINDOW', b'SERVICE::MSR::FIXED_CTR0#', b'SERVICE::MSR::FIXED_CTR0:INST_RETIRED_ANY', b'SERVICE::MSR::FIXED_CTR1#', b'SERVICE::MSR::FIXED_CTR1:CPU_CLK_UNHALTED_THREAD', b'SERVICE::MSR::FIXED_CTR2#', b'SERVICE::MSR::FIXED_CTR2:CPU_CLK_UNHALTED_REF_TSC', b'SERVICE::MSR::FIXED_CTR_CTRL#', b'SERVICE::MSR::FIXED_CTR_CTRL:EN0_OS', b'SERVICE::MSR::FIXED_CTR_CTRL:EN0_PMI', b'SERVICE::MSR::FIXED_CTR_CTRL:EN0_USR', b'SERVICE::MSR::FIXED_CTR_CTRL:EN1_OS', b'SERVICE::MSR::FIXED_CTR_CTRL:EN1_PMI', b'SERVICE::MSR::FIXED_CTR_CTRL:EN1_USR', b'SERVICE::MSR::FIXED_CTR_CTRL:EN2_OS', b'SERVICE::MSR::FIXED_CTR_CTRL:EN2_PMI', b'SERVICE::MSR::FIXED_CTR_CTRL:EN2_USR', b'SERVICE::MSR::IA32_PERFEVTSEL0#', b'SERVICE::MSR::IA32_PERFEVTSEL0:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL0:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL0:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL0:EN', b'SERVICE::MSR::IA32_PERFEVTSEL0:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL0:INT', b'SERVICE::MSR::IA32_PERFEVTSEL0:INV', b'SERVICE::MSR::IA32_PERFEVTSEL0:OS', b'SERVICE::MSR::IA32_PERFEVTSEL0:PC', b'SERVICE::MSR::IA32_PERFEVTSEL0:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL0:USR', b'SERVICE::MSR::IA32_PERFEVTSEL1#', b'SERVICE::MSR::IA32_PERFEVTSEL1:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL1:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL1:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL1:EN', b'SERVICE::MSR::IA32_PERFEVTSEL1:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL1:INT', b'SERVICE::MSR::IA32_PERFEVTSEL1:INV', b'SERVICE::MSR::IA32_PERFEVTSEL1:OS', b'SERVICE::MSR::IA32_PERFEVTSEL1:PC', b'SERVICE::MSR::IA32_PERFEVTSEL1:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL1:USR', b'SERVICE::MSR::IA32_PERFEVTSEL2#', b'SERVICE::MSR::IA32_PERFEVTSEL2:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL2:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL2:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL2:EN', b'SERVICE::MSR::IA32_PERFEVTSEL2:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL2:INT', b'SERVICE::MSR::IA32_PERFEVTSEL2:INV', b'SERVICE::MSR::IA32_PERFEVTSEL2:OS', b'SERVICE::MSR::IA32_PERFEVTSEL2:PC', b'SERVICE::MSR::IA32_PERFEVTSEL2:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL2:USR', b'SERVICE::MSR::IA32_PERFEVTSEL3#', b'SERVICE::MSR::IA32_PERFEVTSEL3:ANYTHREAD', b'SERVICE::MSR::IA32_PERFEVTSEL3:CMASK', b'SERVICE::MSR::IA32_PERFEVTSEL3:EDGE', b'SERVICE::MSR::IA32_PERFEVTSEL3:EN', b'SERVICE::MSR::IA32_PERFEVTSEL3:EVENT_SELECT', b'SERVICE::MSR::IA32_PERFEVTSEL3:INT', b'SERVICE::MSR::IA32_PERFEVTSEL3:INV', b'SERVICE::MSR::IA32_PERFEVTSEL3:OS', b'SERVICE::MSR::IA32_PERFEVTSEL3:PC', b'SERVICE::MSR::IA32_PERFEVTSEL3:UMASK', b'SERVICE::MSR::IA32_PERFEVTSEL3:USR', b'SERVICE::MSR::IA32_PMC0#', b'SERVICE::MSR::IA32_PMC0:PERFCTR', b'SERVICE::MSR::IA32_PMC1#', b'SERVICE::MSR::IA32_PMC1:PERFCTR', b'SERVICE::MSR::IA32_PMC2#', b'SERVICE::MSR::IA32_PMC2:PERFCTR', b'SERVICE::MSR::IA32_PMC3#', b'SERVICE::MSR::IA32_PMC3:PERFCTR', b'SERVICE::MSR::MISC_ENABLE#', b'SERVICE::MSR::MISC_ENABLE:ENHANCED_SPEEDSTEP_TECH_ENABLE', b'SERVICE::MSR::MISC_ENABLE:LIMIT_CPUID_MAXVAL', b'SERVICE::MSR::MISC_ENABLE:TURBO_MODE_DISABLE', b'SERVICE::MSR::MPERF#', b'SERVICE::MSR::MPERF:MCNT', b'SERVICE::MSR::PACKAGE_THERM_STATUS#', b'SERVICE::MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:CRITICAL_TEMP_STATUS', b'SERVICE::MSR::PACKAGE_THERM_STATUS:DIGITAL_READOUT', b'SERVICE::MSR::PACKAGE_THERM_STATUS:POWER_LIMIT_STATUS', b'SERVICE::MSR::PACKAGE_THERM_STATUS:POWER_NOTIFICATION_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:PROCHOT_EVENT', b'SERVICE::MSR::PACKAGE_THERM_STATUS:PROCHOT_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_FLAG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_STATUS_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_1_STATUS', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_LOG', b'SERVICE::MSR::PACKAGE_THERM_STATUS:THERMAL_THRESH_2_STATUS', b'SERVICE::MSR::PERF_CTL#', b'SERVICE::MSR::PERF_CTL:FREQ', b'SERVICE::MSR::PERF_GLOBAL_CTRL#', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR0', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR1', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_FIXED_CTR2', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC0', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC1', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC2', b'SERVICE::MSR::PERF_GLOBAL_CTRL:EN_PMC3', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL#', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR0', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR1', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_FIXED_CTR2', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC0', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC1', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC2', b'SERVICE::MSR::PERF_GLOBAL_OVF_CTRL:CLEAR_OVF_PMC3', b'SERVICE::MSR::PERF_STATUS#', b'SERVICE::MSR::PERF_STATUS:FREQ', b'SERVICE::MSR::PKG_ENERGY_STATUS#', b'SERVICE::MSR::PKG_ENERGY_STATUS:ENERGY', b'SERVICE::MSR::PKG_POWER_INFO#', b'SERVICE::MSR::PKG_POWER_INFO:MAX_POWER', b'SERVICE::MSR::PKG_POWER_INFO:MAX_TIME_WINDOW', b'SERVICE::MSR::PKG_POWER_INFO:MIN_POWER', b'SERVICE::MSR::PKG_POWER_INFO:THERMAL_SPEC_POWER', b'SERVICE::MSR::PKG_POWER_LIMIT#', b'SERVICE::MSR::PKG_POWER_LIMIT:LOCK', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_CLAMP_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_LIMIT_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_POWER_LIMIT', b'SERVICE::MSR::PKG_POWER_LIMIT:PL1_TIME_WINDOW', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_CLAMP_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_LIMIT_ENABLE', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_POWER_LIMIT', b'SERVICE::MSR::PKG_POWER_LIMIT:PL2_TIME_WINDOW', b'SERVICE::MSR::PLATFORM_INFO#', b'SERVICE::MSR::PLATFORM_INFO:MAX_EFFICIENCY_RATIO', b'SERVICE::MSR::PLATFORM_INFO:MAX_NON_TURBO_RATIO', b'SERVICE::MSR::PLATFORM_INFO:PROGRAMMABLE_RATIO_LIMITS_TURBO_MODE', b'SERVICE::MSR::PLATFORM_INFO:PROGRAMMABLE_TCC_ACTIVATION_OFFSET', b'SERVICE::MSR::PLATFORM_INFO:PROGRAMMABLE_TDP_LIMITS_TURBO_MODE', b'SERVICE::MSR::PPERF#', b'SERVICE::MSR::PPERF:PCNT', b'SERVICE::MSR::PQR_ASSOC#', b'SERVICE::MSR::PQR_ASSOC:RMID', b'SERVICE::MSR::QM_CTR#', b'SERVICE::MSR::QM_CTR:ERROR', b'SERVICE::MSR::QM_CTR:RM_DATA', b'SERVICE::MSR::QM_CTR:UNAVAILABLE', b'SERVICE::MSR::QM_EVTSEL#', b'SERVICE::MSR::QM_EVTSEL:EVENT_ID', b'SERVICE::MSR::QM_EVTSEL:RMID', b'SERVICE::MSR::RAPL_POWER_UNIT#', b'SERVICE::MSR::RAPL_POWER_UNIT:ENERGY', b'SERVICE::MSR::RAPL_POWER_UNIT:POWER', b'SERVICE::MSR::RAPL_POWER_UNIT:TIME', b'SERVICE::MSR::TEMPERATURE_TARGET#', b'SERVICE::MSR::TEMPERATURE_TARGET:PROCHOT_MIN', b'SERVICE::MSR::TEMPERATURE_TARGET:TCC_ACTIVE_OFFSET', b'SERVICE::MSR::THERM_STATUS#', b'SERVICE::MSR::THERM_STATUS:CRITICAL_TEMP_LOG', b'SERVICE::MSR::THERM_STATUS:CRITICAL_TEMP_STATUS', b'SERVICE::MSR::THERM_STATUS:DIGITAL_READOUT', b'SERVICE::MSR::THERM_STATUS:POWER_LIMIT_STATUS', b'SERVICE::MSR::THERM_STATUS:POWER_NOTIFICATION_LOG', b'SERVICE::MSR::THERM_STATUS:PROCHOT_EVENT', b'SERVICE::MSR::THERM_STATUS:PROCHOT_LOG', b'SERVICE::MSR::THERM_STATUS:READING_VALID', b'SERVICE::MSR::THERM_STATUS:RESOLUTION', b'SERVICE::MSR::THERM_STATUS:THERMAL_STATUS_FLAG', b'SERVICE::MSR::THERM_STATUS:THERMAL_STATUS_LOG', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_1_LOG', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_1_STATUS', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_2_LOG', b'SERVICE::MSR::THERM_STATUS:THERMAL_THRESH_2_STATUS', b'SERVICE::MSR::TIME', b'SERVICE::MSR::TIME_STAMP_COUNTER#', b'SERVICE::MSR::TIME_STAMP_COUNTER:TIMESTAMP_COUNT', b'SERVICE::MSR::TURBO_RATIO_LIMIT#', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_0', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_1', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_2', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_3', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_4', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_5', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_6', b'SERVICE::MSR::TURBO_RATIO_LIMIT:MAX_RATIO_LIMIT_7', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES#', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_0', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_1', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_2', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_3', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_4', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_5', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_6', b'SERVICE::MSR::TURBO_RATIO_LIMIT_CORES:NUMCORE_7', b'SERVICE::MSR::UNCORE_PERF_STATUS#', b'SERVICE::MSR::UNCORE_PERF_STATUS:FREQ', b'SERVICE::MSR::UNCORE_RATIO_LIMIT#', b'SERVICE::MSR::UNCORE_RATIO_LIMIT:MAX_RATIO', b'SERVICE::MSR::UNCORE_RATIO_LIMIT:MIN_RATIO', b'SERVICE::POWER_DRAM', b'SERVICE::POWER_PACKAGE', b'SERVICE::POWER_PACKAGE_MAX', b'SERVICE::POWER_PACKAGE_MIN', b'SERVICE::POWER_PACKAGE_TDP', b'SERVICE::QM_CTR_SCALED', b'SERVICE::QM_CTR_SCALED_RATE', b'SERVICE::TEMPERATURE_CORE', b'SERVICE::TEMPERATURE_PACKAGE', b'SERVICE::TIME', b'SERVICE::TIME::ELAPSED', b'SERVICE::TIMESTAMP_COUNTER', b'TEMPERATURE_CORE', b'TEMPERATURE_PACKAGE', b'TIME', b'TIME::ELAPSED', b'TIMESTAMP_COUNTER'] ====================================================================== FAIL: test_geopmwrite_set_freq (__main__.TestIntegrationGeopmio) ---------------------------------------------------------------------- Traceback (most recent call last): File "./test_geopmio.py", line 268, in test_geopmwrite_set_freq self.assertEqual(min_freq, result) AssertionError: 1000000000.0 != 2800000000.0 ---------------------------------------------------------------------- Ran 6 tests in 209.203s FAILED (failures=2) ``` **Additional context** There are 2 failures above: 1. The custom MSR plugin is not working. 2. `geopmwrite` is not able to set the frequency
priority
test geopmio py failures describe the bug i tried to run test geopmio py with the service installed and direct msr safe access removed and i expected the test to pass instead some of the tests in that file fail geopm version expected behavior the test runs to completion with a return code of actual behavior issue test geopmio py v test geopmio log examining the log file issue cat test geopmio log test geopmread all signal agg main testintegrationgeopmio ok test geopmread command line main testintegrationgeopmio ok test geopmread custom msr main testintegrationgeopmio fail test geopmread signal value main testintegrationgeopmio ok test geopmwrite command line main testintegrationgeopmio ok test geopmwrite set freq main testintegrationgeopmio stress ng info defaulting to a second day secs run per stressor stress ng info dispatching hogs cpu stress ng info successful run completed in fail fail test geopmread custom msr main testintegrationgeopmio traceback most recent call last file test geopmio py line in test geopmread custom msr self assertin b msr core perf limit reasons all signals assertionerror b msr core perf limit reasons not found in fail test geopmwrite set freq main testintegrationgeopmio traceback most recent call last file test geopmio py line in test geopmwrite set freq self assertequal min freq result assertionerror ran tests in failed failures additional context there are failures above the custom msr plugin is not working geopmwrite is not able to set the frequency
1
240,845
7,806,468,534
IssuesEvent
2018-06-11 14:09:18
larray-project/larray
https://api.github.com/repos/larray-project/larray
opened
string labels dumped to Excel should be formatted as strings
bug component: excel priority: high
```python >>> arr = ndtest('a=10E01,10E03') >>> arr.to_excel() ``` produces something like: ``` a | 1.00E+02 | 1.00E+04 ย  | 0 | 1 ``` This is not a visual transformation, it actually converts the values to numbers (whatever the actual dtype of labels were -- .dump() converts to raw Python strings anyway) ```python >>> arr.a.labels.dtype dtype('<U5') >>> arr.dump() [['a', '10E01', '10E03'], ['', 0, 1]] ``` The solution is to set NumberFormat on the target cells ***before*** pasting the values (https://github.com/ZoomerAnalytics/xlwings/issues/436). ```python >>> wb = open_excel() >>> wb[0]['A1'] = '10E03' >>> wb[0]['A2'].api.NumberFormat = "@" >>> wb[0]['A2'] = '10E03' ``` Note that set_labels (currently) returns object arrays. This should be irrelevant when we go via .dump(), which is always the case, I think (when using `wb[0] = array`, .dump() is called automatically). ```python >>> ndtest(3).set_labels('a', {'a1': '10E03'}).a.labels.dtype dtype('O') ``` Ideally, this should be fixed in xlwings. Honestly, this issue seems easy to fix for small arrays (by scanning through each value and setting the format accordingly) but hard to do efficiently (set the format in chunks). Maybe .dump should not return a simple list of list but a custom object which retains the notion of axes etc. so that we can set the format for each axis in a single operation?
1.0
string labels dumped to Excel should be formatted as strings - ```python >>> arr = ndtest('a=10E01,10E03') >>> arr.to_excel() ``` produces something like: ``` a | 1.00E+02 | 1.00E+04 ย  | 0 | 1 ``` This is not a visual transformation, it actually converts the values to numbers (whatever the actual dtype of labels were -- .dump() converts to raw Python strings anyway) ```python >>> arr.a.labels.dtype dtype('<U5') >>> arr.dump() [['a', '10E01', '10E03'], ['', 0, 1]] ``` The solution is to set NumberFormat on the target cells ***before*** pasting the values (https://github.com/ZoomerAnalytics/xlwings/issues/436). ```python >>> wb = open_excel() >>> wb[0]['A1'] = '10E03' >>> wb[0]['A2'].api.NumberFormat = "@" >>> wb[0]['A2'] = '10E03' ``` Note that set_labels (currently) returns object arrays. This should be irrelevant when we go via .dump(), which is always the case, I think (when using `wb[0] = array`, .dump() is called automatically). ```python >>> ndtest(3).set_labels('a', {'a1': '10E03'}).a.labels.dtype dtype('O') ``` Ideally, this should be fixed in xlwings. Honestly, this issue seems easy to fix for small arrays (by scanning through each value and setting the format accordingly) but hard to do efficiently (set the format in chunks). Maybe .dump should not return a simple list of list but a custom object which retains the notion of axes etc. so that we can set the format for each axis in a single operation?
priority
string labels dumped to excel should be formatted as strings python arr ndtest a arr to excel produces something like a ย  this is not a visual transformation it actually converts the values to numbers whatever the actual dtype of labels were dump converts to raw python strings anyway python arr a labels dtype dtype arr dump the solution is to set numberformat on the target cells before pasting the values python wb open excel wb wb api numberformat wb note that set labels currently returns object arrays this should be irrelevant when we go via dump which is always the case i think when using wb array dump is called automatically python ndtest set labels a a labels dtype dtype o ideally this should be fixed in xlwings honestly this issue seems easy to fix for small arrays by scanning through each value and setting the format accordingly but hard to do efficiently set the format in chunks maybe dump should not return a simple list of list but a custom object which retains the notion of axes etc so that we can set the format for each axis in a single operation
1
405,402
11,872,670,889
IssuesEvent
2020-03-26 16:09:05
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
forking modals are missing processing tags
Bug Needed for V2 launch Priority: High
Migrating and releasing Rep buttons should get processing buttons so user knows when the tx is going through or failed
1.0
forking modals are missing processing tags - Migrating and releasing Rep buttons should get processing buttons so user knows when the tx is going through or failed
priority
forking modals are missing processing tags migrating and releasing rep buttons should get processing buttons so user knows when the tx is going through or failed
1
771,868
27,096,881,190
IssuesEvent
2023-02-15 04:10:44
SuddenDevelopment/StopMotion
https://api.github.com/repos/SuddenDevelopment/StopMotion
closed
Spacing keys is confusing
Priority High
We might need to chat about this and then get feedback from the user. Because we are animating at 3 different levels when should the keyframes stop spacing? ### A couple of directions we can move forward 1. Stick to the current method. 2. Current method and a keyframe selection method. 3. Just a selection method **Examples** shows off keys closing in on the gap at all levels. How can the use have more control instead of just being aware of the current behavior? https://user-images.githubusercontent.com/16330340/216887876-1132105e-85d5-44ca-b8b7-67e06a9555d5.mp4
1.0
Spacing keys is confusing - We might need to chat about this and then get feedback from the user. Because we are animating at 3 different levels when should the keyframes stop spacing? ### A couple of directions we can move forward 1. Stick to the current method. 2. Current method and a keyframe selection method. 3. Just a selection method **Examples** shows off keys closing in on the gap at all levels. How can the use have more control instead of just being aware of the current behavior? https://user-images.githubusercontent.com/16330340/216887876-1132105e-85d5-44ca-b8b7-67e06a9555d5.mp4
priority
spacing keys is confusing we might need to chat about this and then get feedback from the user because we are animating at different levels when should the keyframes stop spacing a couple of directions we can move forward stick to the current method current method and a keyframe selection method just a selection method examples shows off keys closing in on the gap at all levels how can the use have more control instead of just being aware of the current behavior
1
575,314
17,027,037,453
IssuesEvent
2021-07-03 18:51:12
MasterCruelty/eMerger
https://api.github.com/repos/MasterCruelty/eMerger
closed
It's not working on Manjaro
bug help wanted high priority
If you launch the setup on Manjaro you would see this error: ```/src/utils/global.sh is missing``` Also it doesnt find ```.cache``` returning another error: ```cat: /home/user/Updater/src/utils/.cache: No such file or directory```
1.0
It's not working on Manjaro - If you launch the setup on Manjaro you would see this error: ```/src/utils/global.sh is missing``` Also it doesnt find ```.cache``` returning another error: ```cat: /home/user/Updater/src/utils/.cache: No such file or directory```
priority
it s not working on manjaro if you launch the setup on manjaro you would see this error src utils global sh is missing also it doesnt find cache returning another error cat home user updater src utils cache no such file or directory
1
195,341
6,910,903,234
IssuesEvent
2017-11-28 05:21:14
vmware/harbor
https://api.github.com/repos/vmware/harbor
closed
Harbor tile: Support multiple auth mode.
area/bosh-release priority/high target/pks-0.8
For PKS 0.8, we should allow user to set db, ldap auth mode, the UAA can be optional.
1.0
Harbor tile: Support multiple auth mode. - For PKS 0.8, we should allow user to set db, ldap auth mode, the UAA can be optional.
priority
harbor tile support multiple auth mode for pks we should allow user to set db ldap auth mode the uaa can be optional
1
274,880
8,569,051,457
IssuesEvent
2018-11-11 05:38:21
red-eclipse/base
https://api.github.com/repos/red-eclipse/base
closed
Should low health alert sound play in instagib mode?
branch: master difficulty: easy priority: high status: enhancement
Currently the low health alert sound is more like a die sound in instagib. IMO it would be best to remove the alert sound entirely from instagib.
1.0
Should low health alert sound play in instagib mode? - Currently the low health alert sound is more like a die sound in instagib. IMO it would be best to remove the alert sound entirely from instagib.
priority
should low health alert sound play in instagib mode currently the low health alert sound is more like a die sound in instagib imo it would be best to remove the alert sound entirely from instagib
1
219,770
7,345,742,527
IssuesEvent
2018-03-07 18:23:28
hackoregon/civic-devops
https://api.github.com/repos/hackoregon/civic-devops
closed
S3 Bucket request for Transportation raw "congestion" dataset
Priority: high
This is our first big dataset and it will be getting bigger. Currently the raw data file is a `.rar` archive that takes up about 4.3 GB compressed and unpacks to 24 GB, most of which is CSV files we need to input into a PostgreSQL database. We expect a new dump every quarter of about the same size. I'm currently building the ingestion process - see <https://github.com/hackoregon/transportation-congestion-analysis/tree/master/src/data> for the code. My workstation has the capacity to do this, but the end goal is a fully automated AWS process, either containerized. or in a full VM if it's too big for containers. I will be attempting the containerized solution first as I build the processing for this first dataset. Ideally we'd want to store both the compressed and uncompressed data now, but once it's containerized we can live with just the compressed data file.
1.0
S3 Bucket request for Transportation raw "congestion" dataset - This is our first big dataset and it will be getting bigger. Currently the raw data file is a `.rar` archive that takes up about 4.3 GB compressed and unpacks to 24 GB, most of which is CSV files we need to input into a PostgreSQL database. We expect a new dump every quarter of about the same size. I'm currently building the ingestion process - see <https://github.com/hackoregon/transportation-congestion-analysis/tree/master/src/data> for the code. My workstation has the capacity to do this, but the end goal is a fully automated AWS process, either containerized. or in a full VM if it's too big for containers. I will be attempting the containerized solution first as I build the processing for this first dataset. Ideally we'd want to store both the compressed and uncompressed data now, but once it's containerized we can live with just the compressed data file.
priority
bucket request for transportation raw congestion dataset this is our first big dataset and it will be getting bigger currently the raw data file is a rar archive that takes up about gb compressed and unpacks to gb most of which is csv files we need to input into a postgresql database we expect a new dump every quarter of about the same size i m currently building the ingestion process see for the code my workstation has the capacity to do this but the end goal is a fully automated aws process either containerized or in a full vm if it s too big for containers i will be attempting the containerized solution first as i build the processing for this first dataset ideally we d want to store both the compressed and uncompressed data now but once it s containerized we can live with just the compressed data file
1
680,240
23,263,569,045
IssuesEvent
2022-08-04 15:19:52
CredentialEngine/CredentialRegistry
https://api.github.com/repos/CredentialEngine/CredentialRegistry
closed
Empty description sets for Competency Frameworks
bug High Priority Blocker
The following Competency Frameworks have no data in/for their description sets: https://credentialengineregistry.org/resources/ce-fd3e874f-f1d3-4895-8c46-dfed04da9336 https://credentialengineregistry.org/resources/ce-f204c85c-1a3e-4ac7-8a3c-705e5833ac9d https://credentialengineregistry.org/resources/ce-0ac2068b-55df-4b3c-9d21-1f77ac34c327 https://credentialengineregistry.org/resources/ce-0c780d22-2195-46e2-85ee-9acdffd67a2c https://credentialengineregistry.org/resources/ce-cf0efb79-427d-424d-983f-a17fc7714db6 https://credentialengineregistry.org/resources/ce-d9b45ce9-6906-4b6e-a40a-5bee394b5d53 ![image](https://user-images.githubusercontent.com/21346519/181101641-312ebe3c-5462-40e4-a364-d5e2f6ba01f4.png) I believe @jeannekitchens plans to present some of these in a meeting in the near future. Any idea what caused the description sets to not be generated?
1.0
Empty description sets for Competency Frameworks - The following Competency Frameworks have no data in/for their description sets: https://credentialengineregistry.org/resources/ce-fd3e874f-f1d3-4895-8c46-dfed04da9336 https://credentialengineregistry.org/resources/ce-f204c85c-1a3e-4ac7-8a3c-705e5833ac9d https://credentialengineregistry.org/resources/ce-0ac2068b-55df-4b3c-9d21-1f77ac34c327 https://credentialengineregistry.org/resources/ce-0c780d22-2195-46e2-85ee-9acdffd67a2c https://credentialengineregistry.org/resources/ce-cf0efb79-427d-424d-983f-a17fc7714db6 https://credentialengineregistry.org/resources/ce-d9b45ce9-6906-4b6e-a40a-5bee394b5d53 ![image](https://user-images.githubusercontent.com/21346519/181101641-312ebe3c-5462-40e4-a364-d5e2f6ba01f4.png) I believe @jeannekitchens plans to present some of these in a meeting in the near future. Any idea what caused the description sets to not be generated?
priority
empty description sets for competency frameworks the following competency frameworks have no data in for their description sets i believe jeannekitchens plans to present some of these in a meeting in the near future any idea what caused the description sets to not be generated
1
748,701
26,133,678,142
IssuesEvent
2022-12-29 09:17:30
LiskHQ/lisk-desktop
https://api.github.com/repos/LiskHQ/lisk-desktop
closed
Update applications domain texts
type: improvement priority: high
### Description Update all the texts under applications domain ### Acceptance Criteria - Review all the texts under [applications domain](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=41%3A52388&t=prTQCIyJ1X49773R-4) and update the texts according to proposal - [Remove applications](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=129%3A79866&t=prTQCIyJ1X49773R-4) - [Disconnect application](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=195%3A77813&t=prTQCIyJ1X49773R-4) - [Bridge application](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=195%3A77815&t=prTQCIyJ1X49773R-4) - [Application connect flow](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=41%3A52384&t=prTQCIyJ1X49773R-4) - [Sign transaction flow](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=41%3A52386&t=prTQCIyJ1X49773R-4)
1.0
Update applications domain texts - ### Description Update all the texts under applications domain ### Acceptance Criteria - Review all the texts under [applications domain](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=41%3A52388&t=prTQCIyJ1X49773R-4) and update the texts according to proposal - [Remove applications](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=129%3A79866&t=prTQCIyJ1X49773R-4) - [Disconnect application](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=195%3A77813&t=prTQCIyJ1X49773R-4) - [Bridge application](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=195%3A77815&t=prTQCIyJ1X49773R-4) - [Application connect flow](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=41%3A52384&t=prTQCIyJ1X49773R-4) - [Sign transaction flow](https://www.figma.com/file/KcrDpvWEKQhdGwNd4CZ5NY/Desktop-Prototype?node-id=41%3A52386&t=prTQCIyJ1X49773R-4)
priority
update applications domain texts description update all the texts under applications domain acceptance criteria review all the texts under and update the texts according to proposal
1
488,926
14,099,587,262
IssuesEvent
2020-11-06 01:47:52
aws/aws-sdk-js-v3
https://api.github.com/repos/aws/aws-sdk-js-v3
closed
Unit test "NodeHttpHandler โ€บ constructor โ€บ can set httpAgent and httpsAgent" intermittently fails
High Priority
**Describe the bug** Unit test "NodeHttpHandler โ€บ constructor โ€บ can set httpAgent and httpsAgent" intermittently fails with the following error: ``` FAIL packages/node-http-handler/dist/cjs/node-http-handler.spec.js -- 947 | โ— NodeHttpHandler โ€บ constructor โ€บ can set httpAgent and httpsAgent 948 | ย  949 | expect(received).toEqual(expected) // deep equality 950 | ย  951 | Expected: 0 952 | Received: Infinity 953 | ย  954 | 22 \| httpsAgent: new https.Agent({ maxSockets }), 955 | 23 \| }); 956 | > 24 \| expect((nodeHttpHandler as any).httpsAgent.maxSockets).toEqual(maxSockets); 957 | \| ^ 958 | 25 \| }); 959 | 26 \| }); 960 | 27 \| describe("http", () => { 961 | ย  962 | at Object.<anonymous> (src/node-http-handler.spec.ts:24:62) ``` [Example run](https://us-west-2.console.aws.amazon.com/codesuite/codebuild/119327258862/projects/sdk-staging-test/build/sdk-staging-test%3A41343a87-64c7-48fa-a42e-1001e20775ec?region=us-west-2) **SDK version number** master **To Reproduce (observed behavior)** Intermittent failure **Expected behavior** The test succeeds every time it's run
1.0
Unit test "NodeHttpHandler โ€บ constructor โ€บ can set httpAgent and httpsAgent" intermittently fails - **Describe the bug** Unit test "NodeHttpHandler โ€บ constructor โ€บ can set httpAgent and httpsAgent" intermittently fails with the following error: ``` FAIL packages/node-http-handler/dist/cjs/node-http-handler.spec.js -- 947 | โ— NodeHttpHandler โ€บ constructor โ€บ can set httpAgent and httpsAgent 948 | ย  949 | expect(received).toEqual(expected) // deep equality 950 | ย  951 | Expected: 0 952 | Received: Infinity 953 | ย  954 | 22 \| httpsAgent: new https.Agent({ maxSockets }), 955 | 23 \| }); 956 | > 24 \| expect((nodeHttpHandler as any).httpsAgent.maxSockets).toEqual(maxSockets); 957 | \| ^ 958 | 25 \| }); 959 | 26 \| }); 960 | 27 \| describe("http", () => { 961 | ย  962 | at Object.<anonymous> (src/node-http-handler.spec.ts:24:62) ``` [Example run](https://us-west-2.console.aws.amazon.com/codesuite/codebuild/119327258862/projects/sdk-staging-test/build/sdk-staging-test%3A41343a87-64c7-48fa-a42e-1001e20775ec?region=us-west-2) **SDK version number** master **To Reproduce (observed behavior)** Intermittent failure **Expected behavior** The test succeeds every time it's run
priority
unit test nodehttphandler โ€บ constructor โ€บ can set httpagent and httpsagent intermittently fails describe the bug unit test nodehttphandler โ€บ constructor โ€บ can set httpagent and httpsagent intermittently fails with the following error fail packages node http handler dist cjs node http handler spec js โ— nodehttphandler โ€บ constructor โ€บ can set httpagent and httpsagent ย  expect received toequal expected deep equality ย  expected received infinity ย  httpsagent new https agent maxsockets expect nodehttphandler as any httpsagent maxsockets toequal maxsockets describe http ย  at object src node http handler spec ts sdk version number master to reproduce observed behavior intermittent failure expected behavior the test succeeds every time it s run
1
31,509
2,733,240,766
IssuesEvent
2015-04-17 12:47:35
UnifiedViews/Plugins
https://api.github.com/repos/UnifiedViews/Plugins
opened
E-FilesDownload fails to download file using HTTPS with invalid certificate
priority: High severity: bug
Try to download https://www.isvz.cz/ReportingSuite/Explorer/Download/Data/XML/VVZ/2014 using E-FilesDownload. The web server does not have a valid certificate and the download fails with ``` Execution failed because: Exception occured while downloading files.eu.unifiedviews.dpu.DPUException: Exception occured while downloading files. at eu.unifiedviews.helpers.dpu.context.ContextUtils.dpuException(ContextUtils.java:193) at eu.unifiedviews.plugins.extractor.filesdownload.FilesDownload.innerExecute(FilesDownload.java:128) at eu.unifiedviews.helpers.dpu.exec.AbstractDpu.execute(AbstractDpu.java:117) at cz.cuni.mff.xrg.odcs.backend.execution.dpu.DPUExecutor.executeInstance(DPUExecutor.java:232) at cz.cuni.mff.xrg.odcs.backend.execution.dpu.DPUExecutor.execute(DPUExecutor.java:370) at cz.cuni.mff.xrg.odcs.backend.execution.dpu.DPUExecutor.run(DPUExecutor.java:452) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.commons.vfs2.FileSystemException: Could not connect to HTTP server on "www.isvz.cz". at org.apache.commons.vfs2.provider.http.HttpClientFactory.createConnection(HttpClientFactory.java:152) at org.apache.commons.vfs2.provider.http.HttpClientFactory.createConnection(HttpClientFactory.java:50) at org.apache.commons.vfs2.provider.http.HttpFileProvider.doCreateFileSystem(HttpFileProvider.java:83) at org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider.getFileSystem(AbstractOriginatingFileProvider.java:103) at org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider.findFile(AbstractOriginatingFileProvider.java:81) at org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider.findFile(AbstractOriginatingFileProvider.java:65) at org.apache.commons.vfs2.impl.DefaultFileSystemManager.resolveFile(DefaultFileSystemManager.java:693) at org.apache.commons.vfs2.impl.DefaultFileSystemManager.resolveFile(DefaultFileSystemManager.java:621) at eu.unifiedviews.plugins.extractor.filesdownload.FilesDownload.innerExecute(FilesDownload.java:125) ... 5 more Caused by: javax.net.ssl.SSLHandshakeException: sun.security.validator.ValidatorException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target at sun.security.ssl.Alerts.getSSLException(Alerts.java:192) at sun.security.ssl.SSLSocketImpl.fatal(SSLSocketImpl.java:1937) at sun.security.ssl.Handshaker.fatalSE(Handshaker.java:302) at sun.security.ssl.Handshaker.fatalSE(Handshaker.java:296) at sun.security.ssl.ClientHandshaker.serverCertificate(ClientHandshaker.java:1478) at sun.security.ssl.ClientHandshaker.processMessage(ClientHandshaker.java:212) at sun.security.ssl.Handshaker.processLoop(Handshaker.java:969) at sun.security.ssl.Handshaker.process_record(Handshaker.java:904) at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1050) at sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1363) at sun.security.ssl.SSLSocketImpl.writeRecord(SSLSocketImpl.java:735) at sun.security.ssl.AppOutputStream.write(AppOutputStream.java:123) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) at org.apache.commons.httpclient.HttpConnection.flushRequestOutputStream(HttpConnection.java:828) at org.apache.commons.httpclient.MultiThreadedHttpConnectionManager$HttpConnectionAdapter.flushRequestOutputStream(MultiThreadedHttpConnectionManager.java:1565) at org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.java:2116) at org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1096) at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398) at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323) at org.apache.commons.vfs2.provider.http.HttpClientFactory.createConnection(HttpClientFactory.java:148) ... 13 more Caused by: sun.security.validator.ValidatorException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target at sun.security.validator.PKIXValidator.doBuild(PKIXValidator.java:387) at sun.security.validator.PKIXValidator.engineValidate(PKIXValidator.java:292) at sun.security.validator.Validator.validate(Validator.java:260) at sun.security.ssl.X509TrustManagerImpl.validate(X509TrustManagerImpl.java:324) at sun.security.ssl.X509TrustManagerImpl.checkTrusted(X509TrustManagerImpl.java:229) at sun.security.ssl.X509TrustManagerImpl.checkServerTrusted(X509TrustManagerImpl.java:124) at sun.security.ssl.ClientHandshaker.serverCertificate(ClientHandshaker.java:1460) ... 31 more Caused by: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target at sun.security.provider.certpath.SunCertPathBuilder.build(SunCertPathBuilder.java:145) at sun.security.provider.certpath.SunCertPathBuilder.engineBuild(SunCertPathBuilder.java:131) at java.security.cert.CertPathBuilder.build(CertPathBuilder.java:280) at sun.security.validator.PKIXValidator.doBuild(PKIXValidator.java:382) ... 37 more ``` The expected behavior is to ignore invalid certificate and download anyway (maybe report an INFO level message). Before this DPU I used <a href="http://stackoverflow.com/questions/12060250/ignore-ssl-certificate-errors-with-java">this workaround</a>.
1.0
E-FilesDownload fails to download file using HTTPS with invalid certificate - Try to download https://www.isvz.cz/ReportingSuite/Explorer/Download/Data/XML/VVZ/2014 using E-FilesDownload. The web server does not have a valid certificate and the download fails with ``` Execution failed because: Exception occured while downloading files.eu.unifiedviews.dpu.DPUException: Exception occured while downloading files. at eu.unifiedviews.helpers.dpu.context.ContextUtils.dpuException(ContextUtils.java:193) at eu.unifiedviews.plugins.extractor.filesdownload.FilesDownload.innerExecute(FilesDownload.java:128) at eu.unifiedviews.helpers.dpu.exec.AbstractDpu.execute(AbstractDpu.java:117) at cz.cuni.mff.xrg.odcs.backend.execution.dpu.DPUExecutor.executeInstance(DPUExecutor.java:232) at cz.cuni.mff.xrg.odcs.backend.execution.dpu.DPUExecutor.execute(DPUExecutor.java:370) at cz.cuni.mff.xrg.odcs.backend.execution.dpu.DPUExecutor.run(DPUExecutor.java:452) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.commons.vfs2.FileSystemException: Could not connect to HTTP server on "www.isvz.cz". at org.apache.commons.vfs2.provider.http.HttpClientFactory.createConnection(HttpClientFactory.java:152) at org.apache.commons.vfs2.provider.http.HttpClientFactory.createConnection(HttpClientFactory.java:50) at org.apache.commons.vfs2.provider.http.HttpFileProvider.doCreateFileSystem(HttpFileProvider.java:83) at org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider.getFileSystem(AbstractOriginatingFileProvider.java:103) at org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider.findFile(AbstractOriginatingFileProvider.java:81) at org.apache.commons.vfs2.provider.AbstractOriginatingFileProvider.findFile(AbstractOriginatingFileProvider.java:65) at org.apache.commons.vfs2.impl.DefaultFileSystemManager.resolveFile(DefaultFileSystemManager.java:693) at org.apache.commons.vfs2.impl.DefaultFileSystemManager.resolveFile(DefaultFileSystemManager.java:621) at eu.unifiedviews.plugins.extractor.filesdownload.FilesDownload.innerExecute(FilesDownload.java:125) ... 5 more Caused by: javax.net.ssl.SSLHandshakeException: sun.security.validator.ValidatorException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target at sun.security.ssl.Alerts.getSSLException(Alerts.java:192) at sun.security.ssl.SSLSocketImpl.fatal(SSLSocketImpl.java:1937) at sun.security.ssl.Handshaker.fatalSE(Handshaker.java:302) at sun.security.ssl.Handshaker.fatalSE(Handshaker.java:296) at sun.security.ssl.ClientHandshaker.serverCertificate(ClientHandshaker.java:1478) at sun.security.ssl.ClientHandshaker.processMessage(ClientHandshaker.java:212) at sun.security.ssl.Handshaker.processLoop(Handshaker.java:969) at sun.security.ssl.Handshaker.process_record(Handshaker.java:904) at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1050) at sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1363) at sun.security.ssl.SSLSocketImpl.writeRecord(SSLSocketImpl.java:735) at sun.security.ssl.AppOutputStream.write(AppOutputStream.java:123) at java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) at java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) at org.apache.commons.httpclient.HttpConnection.flushRequestOutputStream(HttpConnection.java:828) at org.apache.commons.httpclient.MultiThreadedHttpConnectionManager$HttpConnectionAdapter.flushRequestOutputStream(MultiThreadedHttpConnectionManager.java:1565) at org.apache.commons.httpclient.HttpMethodBase.writeRequest(HttpMethodBase.java:2116) at org.apache.commons.httpclient.HttpMethodBase.execute(HttpMethodBase.java:1096) at org.apache.commons.httpclient.HttpMethodDirector.executeWithRetry(HttpMethodDirector.java:398) at org.apache.commons.httpclient.HttpMethodDirector.executeMethod(HttpMethodDirector.java:171) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:397) at org.apache.commons.httpclient.HttpClient.executeMethod(HttpClient.java:323) at org.apache.commons.vfs2.provider.http.HttpClientFactory.createConnection(HttpClientFactory.java:148) ... 13 more Caused by: sun.security.validator.ValidatorException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target at sun.security.validator.PKIXValidator.doBuild(PKIXValidator.java:387) at sun.security.validator.PKIXValidator.engineValidate(PKIXValidator.java:292) at sun.security.validator.Validator.validate(Validator.java:260) at sun.security.ssl.X509TrustManagerImpl.validate(X509TrustManagerImpl.java:324) at sun.security.ssl.X509TrustManagerImpl.checkTrusted(X509TrustManagerImpl.java:229) at sun.security.ssl.X509TrustManagerImpl.checkServerTrusted(X509TrustManagerImpl.java:124) at sun.security.ssl.ClientHandshaker.serverCertificate(ClientHandshaker.java:1460) ... 31 more Caused by: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target at sun.security.provider.certpath.SunCertPathBuilder.build(SunCertPathBuilder.java:145) at sun.security.provider.certpath.SunCertPathBuilder.engineBuild(SunCertPathBuilder.java:131) at java.security.cert.CertPathBuilder.build(CertPathBuilder.java:280) at sun.security.validator.PKIXValidator.doBuild(PKIXValidator.java:382) ... 37 more ``` The expected behavior is to ignore invalid certificate and download anyway (maybe report an INFO level message). Before this DPU I used <a href="http://stackoverflow.com/questions/12060250/ignore-ssl-certificate-errors-with-java">this workaround</a>.
priority
e filesdownload fails to download file using https with invalid certificate try to download using e filesdownload the web server does not have a valid certificate and the download fails with execution failed because exception occured while downloading files eu unifiedviews dpu dpuexception exception occured while downloading files at eu unifiedviews helpers dpu context contextutils dpuexception contextutils java at eu unifiedviews plugins extractor filesdownload filesdownload innerexecute filesdownload java at eu unifiedviews helpers dpu exec abstractdpu execute abstractdpu java at cz cuni mff xrg odcs backend execution dpu dpuexecutor executeinstance dpuexecutor java at cz cuni mff xrg odcs backend execution dpu dpuexecutor execute dpuexecutor java at cz cuni mff xrg odcs backend execution dpu dpuexecutor run dpuexecutor java at java lang thread run thread java caused by org apache commons filesystemexception could not connect to http server on at org apache commons provider http httpclientfactory createconnection httpclientfactory java at org apache commons provider http httpclientfactory createconnection httpclientfactory java at org apache commons provider http httpfileprovider docreatefilesystem httpfileprovider java at org apache commons provider abstractoriginatingfileprovider getfilesystem abstractoriginatingfileprovider java at org apache commons provider abstractoriginatingfileprovider findfile abstractoriginatingfileprovider java at org apache commons provider abstractoriginatingfileprovider findfile abstractoriginatingfileprovider java at org apache commons impl defaultfilesystemmanager resolvefile defaultfilesystemmanager java at org apache commons impl defaultfilesystemmanager resolvefile defaultfilesystemmanager java at eu unifiedviews plugins extractor filesdownload filesdownload innerexecute filesdownload java more caused by javax net ssl sslhandshakeexception sun security validator validatorexception pkix path building failed sun security provider certpath suncertpathbuilderexception unable to find valid certification path to requested target at sun security ssl alerts getsslexception alerts java at sun security ssl sslsocketimpl fatal sslsocketimpl java at sun security ssl handshaker fatalse handshaker java at sun security ssl handshaker fatalse handshaker java at sun security ssl clienthandshaker servercertificate clienthandshaker java at sun security ssl clienthandshaker processmessage clienthandshaker java at sun security ssl handshaker processloop handshaker java at sun security ssl handshaker process record handshaker java at sun security ssl sslsocketimpl readrecord sslsocketimpl java at sun security ssl sslsocketimpl performinitialhandshake sslsocketimpl java at sun security ssl sslsocketimpl writerecord sslsocketimpl java at sun security ssl appoutputstream write appoutputstream java at java io bufferedoutputstream flushbuffer bufferedoutputstream java at java io bufferedoutputstream flush bufferedoutputstream java at org apache commons httpclient httpconnection flushrequestoutputstream httpconnection java at org apache commons httpclient multithreadedhttpconnectionmanager httpconnectionadapter flushrequestoutputstream multithreadedhttpconnectionmanager java at org apache commons httpclient httpmethodbase writerequest httpmethodbase java at org apache commons httpclient httpmethodbase execute httpmethodbase java at org apache commons httpclient httpmethoddirector executewithretry httpmethoddirector java at org apache commons httpclient httpmethoddirector executemethod httpmethoddirector java at org apache commons httpclient httpclient executemethod httpclient java at org apache commons httpclient httpclient executemethod httpclient java at org apache commons provider http httpclientfactory createconnection httpclientfactory java more caused by sun security validator validatorexception pkix path building failed sun security provider certpath suncertpathbuilderexception unable to find valid certification path to requested target at sun security validator pkixvalidator dobuild pkixvalidator java at sun security validator pkixvalidator enginevalidate pkixvalidator java at sun security validator validator validate validator java at sun security ssl validate java at sun security ssl checktrusted java at sun security ssl checkservertrusted java at sun security ssl clienthandshaker servercertificate clienthandshaker java more caused by sun security provider certpath suncertpathbuilderexception unable to find valid certification path to requested target at sun security provider certpath suncertpathbuilder build suncertpathbuilder java at sun security provider certpath suncertpathbuilder enginebuild suncertpathbuilder java at java security cert certpathbuilder build certpathbuilder java at sun security validator pkixvalidator dobuild pkixvalidator java more the expected behavior is to ignore invalid certificate and download anyway maybe report an info level message before this dpu i used
1
795,594
28,078,650,804
IssuesEvent
2023-03-30 03:20:59
openmsupply/open-msupply
https://api.github.com/repos/openmsupply/open-msupply
closed
Not all Master list items seems to be included or visible/reachable on internal order or in the catalogue of that master list
bug front-end Priority: High
## What went wrong? ๐Ÿ˜ฒ <!-- Provide a clear and concise description of what the bug is. Screenshots are helpful! --> When adding items using the "Add from master list" button on a internal order. Not all the items seems to be included or visible/reachable. The master list I have been using is shared by both the Requesting and Supplier stores. I was looking for 06_4956 Acetylcysteine powder 200mg sachet. I can see this item in the Item list in omsupply When I do the filtering in the internal order ![image](https://user-images.githubusercontent.com/11624516/225487537-8a942724-a2f0-4906-a690-6db1193de2a6.png) I can see it when adding a new item ![image](https://user-images.githubusercontent.com/11624516/225487336-34429461-2a23-4865-9dc1-d4ea29c8faf6.png) Master list on omSupply The are in alphabetical order so should be the first one on top ![image](https://user-images.githubusercontent.com/11624516/225491487-112c1472-6e6a-474f-9782-465565938a8f.png) Same Master list in mSupply ![image](https://user-images.githubusercontent.com/11624516/225492798-8ca358e0-fff9-4aa9-90b3-652b471c5e1c.png) ## Expected behaviour ๐Ÿค” I think all items of the master list should be displayed or accessible in some way ## How to Reproduce ๐Ÿ”จ Steps to reproduce the behaviour: Install omSupply 1.1.5 and linked it to the Alain Abode site of the Timer demo cloud server (available in teamviewer) . 1. After initalisation 2. Click on 'Replenishment > Internal orders' 3. Create a new internal order and press the Add from master list button. Select the SAMES MASTER LIST 4. Try to find 06_4956 Acetylcysteine powder 200mg sachet. See error, it is not there. ## Your environment ๐ŸŒฑ - Version [e.g. 1.1.5] - Platform [e.g. tablet]
1.0
Not all Master list items seems to be included or visible/reachable on internal order or in the catalogue of that master list - ## What went wrong? ๐Ÿ˜ฒ <!-- Provide a clear and concise description of what the bug is. Screenshots are helpful! --> When adding items using the "Add from master list" button on a internal order. Not all the items seems to be included or visible/reachable. The master list I have been using is shared by both the Requesting and Supplier stores. I was looking for 06_4956 Acetylcysteine powder 200mg sachet. I can see this item in the Item list in omsupply When I do the filtering in the internal order ![image](https://user-images.githubusercontent.com/11624516/225487537-8a942724-a2f0-4906-a690-6db1193de2a6.png) I can see it when adding a new item ![image](https://user-images.githubusercontent.com/11624516/225487336-34429461-2a23-4865-9dc1-d4ea29c8faf6.png) Master list on omSupply The are in alphabetical order so should be the first one on top ![image](https://user-images.githubusercontent.com/11624516/225491487-112c1472-6e6a-474f-9782-465565938a8f.png) Same Master list in mSupply ![image](https://user-images.githubusercontent.com/11624516/225492798-8ca358e0-fff9-4aa9-90b3-652b471c5e1c.png) ## Expected behaviour ๐Ÿค” I think all items of the master list should be displayed or accessible in some way ## How to Reproduce ๐Ÿ”จ Steps to reproduce the behaviour: Install omSupply 1.1.5 and linked it to the Alain Abode site of the Timer demo cloud server (available in teamviewer) . 1. After initalisation 2. Click on 'Replenishment > Internal orders' 3. Create a new internal order and press the Add from master list button. Select the SAMES MASTER LIST 4. Try to find 06_4956 Acetylcysteine powder 200mg sachet. See error, it is not there. ## Your environment ๐ŸŒฑ - Version [e.g. 1.1.5] - Platform [e.g. tablet]
priority
not all master list items seems to be included or visible reachable on internal order or in the catalogue of that master list what went wrong ๐Ÿ˜ฒ when adding items using the add from master list button on a internal order not all the items seems to be included or visible reachable the master list i have been using is shared by both the requesting and supplier stores i was looking for acetylcysteine powder sachet i can see this item in the item list in omsupply when i do the filtering in the internal order i can see it when adding a new item master list on omsupply the are in alphabetical order so should be the first one on top same master list in msupply expected behaviour ๐Ÿค” i think all items of the master list should be displayed or accessible in some way how to reproduce ๐Ÿ”จ steps to reproduce the behaviour install omsupply and linked it to the alain abode site of the timer demo cloud server available in teamviewer after initalisation click on replenishment internal orders create a new internal order and press the add from master list button select the sames master list try to find acetylcysteine powder sachet see error it is not there your environment ๐ŸŒฑ version platform
1
283,330
8,718,810,981
IssuesEvent
2018-12-07 21:45:16
openstax/bit
https://api.github.com/repos/openstax/bit
closed
SEO-Optimize H1 tags
Change Request RELEASED Story priority1-high
### Description The following criteria should be met to optimize H1 tags: (a) There should be exactly one H1 per page. Many pages are missing the H1 (e.g. on book pages, should be book title) (b) H1s should be unique (c) The most competitive keywords should be represented in the H1 (d) 3-5 words ideal ### Acceptance H1 tags have been added to all webpages.
1.0
SEO-Optimize H1 tags - ### Description The following criteria should be met to optimize H1 tags: (a) There should be exactly one H1 per page. Many pages are missing the H1 (e.g. on book pages, should be book title) (b) H1s should be unique (c) The most competitive keywords should be represented in the H1 (d) 3-5 words ideal ### Acceptance H1 tags have been added to all webpages.
priority
seo optimize tags description the following criteria should be met to optimize tags a there should be exactly one per page many pages are missing the e g on book pages should be book title b should be unique c the most competitive keywords should be represented in the d words ideal acceptance tags have been added to all webpages
1
642,869
20,916,162,376
IssuesEvent
2022-03-24 13:37:16
AY2122S2-CS2103T-T11-3/tp
https://api.github.com/repos/AY2122S2-CS2103T-T11-3/tp
closed
Docs - Update UG for edit command
type.Task priority.High
Update the UG to reflect the changes made to the `edit` command to allow the editing of priority tag and application status tag of any specified application using the prefix pt/ and ast/ respectively.
1.0
Docs - Update UG for edit command - Update the UG to reflect the changes made to the `edit` command to allow the editing of priority tag and application status tag of any specified application using the prefix pt/ and ast/ respectively.
priority
docs update ug for edit command update the ug to reflect the changes made to the edit command to allow the editing of priority tag and application status tag of any specified application using the prefix pt and ast respectively
1
194,694
6,897,821,446
IssuesEvent
2017-11-24 06:02:46
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
closed
FTP service is shown as file as opposed to sample in Ballerina by example
Priority/Highest Severity/Blocker Type/Bug
Version 0.93 Browser: Chrome Version 61.0.3163.100 As shown in the attached image, for ftp service the configuration of service is shown as <file>. But in [1] it's shown as <ftp>. ![composer_ftpconnector](https://user-images.githubusercontent.com/15624590/31370700-6f9547ae-adaa-11e7-805f-026b007c9ba7.png) [1] https://ballerinalang.org/docs/by-example/ftp-server-connector
1.0
FTP service is shown as file as opposed to sample in Ballerina by example - Version 0.93 Browser: Chrome Version 61.0.3163.100 As shown in the attached image, for ftp service the configuration of service is shown as <file>. But in [1] it's shown as <ftp>. ![composer_ftpconnector](https://user-images.githubusercontent.com/15624590/31370700-6f9547ae-adaa-11e7-805f-026b007c9ba7.png) [1] https://ballerinalang.org/docs/by-example/ftp-server-connector
priority
ftp service is shown as file as opposed to sample in ballerina by example version browser chrome version as shown in the attached image for ftp service the configuration of service is shown as but in it s shown as
1
213,321
7,248,236,474
IssuesEvent
2018-02-15 08:32:25
wso2/puppet-apim
https://api.github.com/repos/wso2/puppet-apim
opened
WSO2 performance tunings are not addressed in puppet scripts
Priority/High Severity/Major Type/Improvement
**Description:** WSO2 performance tuning[1] should be addressed by the puppet scripts. These configuration are production recommended tunings so puppet modules should incorporate those changes to make them production ready. [1] https://docs.wso2.com/display/AM2xx/Tuning+Performance **Affected Product Version:** apim2.1.0
1.0
WSO2 performance tunings are not addressed in puppet scripts - **Description:** WSO2 performance tuning[1] should be addressed by the puppet scripts. These configuration are production recommended tunings so puppet modules should incorporate those changes to make them production ready. [1] https://docs.wso2.com/display/AM2xx/Tuning+Performance **Affected Product Version:** apim2.1.0
priority
performance tunings are not addressed in puppet scripts description performance tuning should be addressed by the puppet scripts these configuration are production recommended tunings so puppet modules should incorporate those changes to make them production ready affected product version
1
546,508
16,013,777,854
IssuesEvent
2021-04-20 13:50:15
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
product assigned to no category
component: product priority: high type: bug
**Prerequisites (mark completed items with an [x]):** - [x] I have have carried out troubleshooting steps and I believe I have found a bug. - [x] I have searched for similar bugs in both open and closed issues and cannot find a duplicate. **Describe the bug** I have a master store and a child store, I exported the products from the master store using the woocommerce core import/export feature. I imported them into my new store (imported all the categories and products). I then decided I wanted to change the default category on my child store which i did. I then removed some categories I did not want on my child store expecting my products within the categories i am removing to be moved to the to the default category. However I noticed that the products were not in there. Due to naivety i had just assumed maybe the products were removed when i deleted the categories, which i thought was strange as normally if a category is deleted then the products within that category are moved to the default category instead. Some time has passed and while i was working on a script today i noticed that the products are actually still on the site somewhere with no assigned category? **Expected behavior** I expected the products within the category i deleted to be moved to the default category. **Actual behavior** products are now on the site with no assigned category **Screenshots** https://imgur.com/a/AQUugvB **Isolating the problem (mark completed items with an [x]):** - [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [] I can reproduce this bug consistently using the steps above. **WordPress Environment** We use the [WooCommerce System Status Report](https://docs.woocommerce.com/document/understanding-the-woocommerce-system-status-report/) to help us evaluate the issue. Without this report we won't be able to fully evaluate this issue. <details> ``` ` ### WordPress Environment ### WordPress address (URL): https://tonsoffunparties.ie Site address (URL): https://tonsoffunparties.ie WC Version: 5.1.0 REST API Version: โœ” 5.1.0 WC Blocks Version: โœ” 4.4.3 Action Scheduler Version: โœ” 3.1.6 WC Admin Version: โœ” 2.0.2 Log Directory Writable: โœ” WP Version: 5.7 WP Multisite: โ€“ WP Memory Limit: 256 MB WP Debug Mode: โœ” WP Cron: โœ” Language: en_US External object cache: โ€“ ### Server Environment ### Server Info: Apache PHP Version: 7.4.16 PHP Post Max Size: 16 MB PHP Time Limit: 300 PHP Max Input Vars: 1000 cURL Version: 7.29.0 NSS/3.53.1 SUHOSIN Installed: โ€“ MySQL Version: 5.5.5-10.3.28-MariaDB Max Upload Size: 16 MB Default Timezone is UTC: โœ” fsockopen/cURL: โœ” SoapClient: โœ” DOMDocument: โœ” GZip: โœ” Multibyte String: โœ” Remote Post: โœ” Remote Get: โœ” ### Database ### WC Database Version: 5.1.0 WC Database Prefix: m7O0KqA_ Total Database Size: 232.57MB Database Data Size: 153.59MB Database Index Size: 78.98MB m7O0KqA_woocommerce_sessions: Data: 0.03MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_api_keys: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_attribute_taxonomies: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_downloadable_product_permissions: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_woocommerce_order_items: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_order_itemmeta: Data: 0.05MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_tax_rates: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_woocommerce_tax_rate_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_shipping_zones: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_woocommerce_shipping_zone_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_shipping_zone_methods: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_woocommerce_payment_tokens: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_payment_tokenmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_log: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_actionscheduler_actions: Data: 1.02MB + Index: 0.81MB + Engine InnoDB m7O0KqA_actionscheduler_claims: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_actionscheduler_groups: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_actionscheduler_logs: Data: 1.09MB + Index: 0.59MB + Engine InnoDB m7O0KqA_aelia_dismissed_messages: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_aelia_exchange_rates_history: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_aioseo_notifications: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_aioseo_posts: Data: 2.52MB + Index: 0.16MB + Engine InnoDB m7O0KqA_aws_cache: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_aws_index: Data: 27.38MB + Index: 13.59MB + Engine InnoDB m7O0KqA_chaty_contact_form_leads: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_cli_cookie_scan: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_cli_cookie_scan_categories: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_cli_cookie_scan_cookies: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_cli_cookie_scan_url: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_cli_scripts: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_commentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_comments: Data: 0.02MB + Index: 0.09MB + Engine InnoDB m7O0KqA_gdbc_attempts: Data: 0.34MB + Index: 0.09MB + Engine InnoDB m7O0KqA_links: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_ms_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_options: Data: 8.02MB + Index: 1.28MB + Engine InnoDB m7O0KqA_pmxi_files: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_hash: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_history: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_images: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_imports: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_posts: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_templates: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_postmeta: Data: 84.47MB + Index: 47.09MB + Engine InnoDB m7O0KqA_posts: Data: 15.30MB + Index: 6.05MB + Engine InnoDB m7O0KqA_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_stock_log: Data: 2.52MB + Index: 0.00MB + Engine InnoDB m7O0KqA_termmeta: Data: 1.48MB + Index: 1.95MB + Engine InnoDB m7O0KqA_terms: Data: 0.48MB + Index: 0.52MB + Engine InnoDB m7O0KqA_term_relationships: Data: 5.02MB + Index: 4.17MB + Engine InnoDB m7O0KqA_term_taxonomy: Data: 1.52MB + Index: 0.63MB + Engine InnoDB m7O0KqA_usermeta: Data: 0.05MB + Index: 0.03MB + Engine InnoDB m7O0KqA_users: Data: 0.02MB + Index: 0.05MB + Engine InnoDB m7O0KqA_wcpdf_invoice_number: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_admin_notes: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_admin_note_actions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_wc_category_lookup: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_customer_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_download_log: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_order_coupon_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_order_product_lookup: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_wc_order_stats: Data: 0.02MB + Index: 0.05MB + Engine InnoDB m7O0KqA_wc_order_tax_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_product_meta_lookup: Data: 1.03MB + Index: 1.02MB + Engine InnoDB m7O0KqA_wc_reserved_stock: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_tax_rate_classes: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_wc_webhooks: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_wss_log: Data: 0.27MB + Index: 0.00MB + Engine InnoDB ### Post Type Counts ### attachment: 13041 cmm4e_menu_theme: 1 cookielawinfo: 6 custom_css: 1 customize_changeset: 47 elementor_library: 9 iksm: 2 nav_menu_item: 63 page: 10 post: 2 product: 5751 revision: 344 seedprod: 2 shop_order: 7 shop_order_refund: 4 sp_wp_carousel: 1 wpmm_theme: 3 ### Security ### Secure connection (HTTPS): โœ” Hide errors from visitors: โŒError messages should not be shown to visitors. ### Active Plugins (21) ### WPBruiser: by Mihai Chelaru โ€“ 3.1.43 Advanced Woo Search: by ILLID โ€“ 2.24 All in One SEO: by All in One SEO Team โ€“ 4.0.17 Chaty: by Premio โ€“ 2.7.6 Code Snippets: by Code Snippets Pro โ€“ 2.14.1 GDPR Cookie Consent: by WebToffee โ€“ 2.0.1 Disable Real MIME Check: by Sergey Biryukov โ€“ 1.0 Elementor Pro: by Elementor.com โ€“ 3.2.0-dev1 Elementor: by Elementor.com โ€“ 3.2.0-dev4 Iks Menu: by IksStudio โ€“ 1.9.0 ID Back To Top: by Idenovasi โ€“ 1.1.1 UberMenu 3 - The Ultimate WordPress Mega Menu: by Chris Mavricos SevenSpark โ€“ 3.7.3 UpdraftPlus - Backup/Restore: by UpdraftPlus.Com DavidAnderson โ€“ 1.16.50 WooCommerce Weight Based Shipping: by weightbasedshipping.com โ€“ 5.3.10 Widget Options: by Widget Options Team โ€“ 3.7.7 Stock Sync for WooCommerce Pro: by Lauri Karisola / WooElements.com โ€“ 2.0.3 WooCommerce EU VAT Assistant: by Aelia โ€“ 2.0.14.210317 WooCommerce Stripe Gateway: by WooCommerce โ€“ 5.0.0 WooCommerce PDF Invoices & Packing Slips: by Ewout Fernhout โ€“ 2.8.2 WooCommerce Stock Manager: by StoreApps โ€“ 2.5.6 WooCommerce: by Automattic โ€“ 5.1.0 ### Inactive Plugins (0) ### ### Dropin Plugins (1) ### maintenance.php: maintenance.php ### Settings ### API Enabled: โœ” Force SSL: โ€“ Currency: EUR (โ‚ฌ) Currency Position: left Thousand Separator: . Decimal Separator: . Number of Decimals: 2 Taxonomies: Product Types: external (external) grouped (grouped) simple (simple) variable (variable) Taxonomies: Product Visibility: exclude-from-catalog (exclude-from-catalog) exclude-from-search (exclude-from-search) featured (featured) outofstock (outofstock) rated-1 (rated-1) rated-2 (rated-2) rated-3 (rated-3) rated-4 (rated-4) rated-5 (rated-5) Connected to WooCommerce.com: โ€“ ### WC Pages ### Shop base: #6 - /shop/ Cart: #7 - /cart/ Checkout: #8 - /checkout/ My account: #9 - /my-account/ Terms and conditions: #58575 - /elementor-58575/ ### Theme ### Name: Supermarket Ecommerce Version: 0.3.4 Author URL: https://www.luzuk.com/ Child Theme: โŒ โ€“ If you are modifying WooCommerce on a parent theme that you did not build personally we recommend using a child theme. See: How to create a child theme WooCommerce Support: โœ” ### Templates ### Overrides: supermarket-ecommerce/woocommerce/archive-product.php supermarket-ecommerce/woocommerce/checkout/form-checkout.php supermarket-ecommerce/woocommerce/global/wrapper-end.php supermarket-ecommerce/woocommerce/global/wrapper-start.php supermarket-ecommerce/woocommerce/single-product.php ### Action Scheduler ### Complete: 3,427 Oldest: 2021-02-27 16:52:31 +0000 Newest: 2021-03-30 16:00:17 +0000 Pending: 3 Oldest: 2021-03-30 16:00:47 +0000 Newest: 2021-03-31 15:31:05 +0000 ### Status report information ### Generated at: 2021-03-30 16:01:02 +00:00 ` ``` </details>
1.0
product assigned to no category - **Prerequisites (mark completed items with an [x]):** - [x] I have have carried out troubleshooting steps and I believe I have found a bug. - [x] I have searched for similar bugs in both open and closed issues and cannot find a duplicate. **Describe the bug** I have a master store and a child store, I exported the products from the master store using the woocommerce core import/export feature. I imported them into my new store (imported all the categories and products). I then decided I wanted to change the default category on my child store which i did. I then removed some categories I did not want on my child store expecting my products within the categories i am removing to be moved to the to the default category. However I noticed that the products were not in there. Due to naivety i had just assumed maybe the products were removed when i deleted the categories, which i thought was strange as normally if a category is deleted then the products within that category are moved to the default category instead. Some time has passed and while i was working on a script today i noticed that the products are actually still on the site somewhere with no assigned category? **Expected behavior** I expected the products within the category i deleted to be moved to the default category. **Actual behavior** products are now on the site with no assigned category **Screenshots** https://imgur.com/a/AQUugvB **Isolating the problem (mark completed items with an [x]):** - [x] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [] I can reproduce this bug consistently using the steps above. **WordPress Environment** We use the [WooCommerce System Status Report](https://docs.woocommerce.com/document/understanding-the-woocommerce-system-status-report/) to help us evaluate the issue. Without this report we won't be able to fully evaluate this issue. <details> ``` ` ### WordPress Environment ### WordPress address (URL): https://tonsoffunparties.ie Site address (URL): https://tonsoffunparties.ie WC Version: 5.1.0 REST API Version: โœ” 5.1.0 WC Blocks Version: โœ” 4.4.3 Action Scheduler Version: โœ” 3.1.6 WC Admin Version: โœ” 2.0.2 Log Directory Writable: โœ” WP Version: 5.7 WP Multisite: โ€“ WP Memory Limit: 256 MB WP Debug Mode: โœ” WP Cron: โœ” Language: en_US External object cache: โ€“ ### Server Environment ### Server Info: Apache PHP Version: 7.4.16 PHP Post Max Size: 16 MB PHP Time Limit: 300 PHP Max Input Vars: 1000 cURL Version: 7.29.0 NSS/3.53.1 SUHOSIN Installed: โ€“ MySQL Version: 5.5.5-10.3.28-MariaDB Max Upload Size: 16 MB Default Timezone is UTC: โœ” fsockopen/cURL: โœ” SoapClient: โœ” DOMDocument: โœ” GZip: โœ” Multibyte String: โœ” Remote Post: โœ” Remote Get: โœ” ### Database ### WC Database Version: 5.1.0 WC Database Prefix: m7O0KqA_ Total Database Size: 232.57MB Database Data Size: 153.59MB Database Index Size: 78.98MB m7O0KqA_woocommerce_sessions: Data: 0.03MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_api_keys: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_attribute_taxonomies: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_downloadable_product_permissions: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_woocommerce_order_items: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_order_itemmeta: Data: 0.05MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_tax_rates: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_woocommerce_tax_rate_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_shipping_zones: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_woocommerce_shipping_zone_locations: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_shipping_zone_methods: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_woocommerce_payment_tokens: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_woocommerce_payment_tokenmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_woocommerce_log: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_actionscheduler_actions: Data: 1.02MB + Index: 0.81MB + Engine InnoDB m7O0KqA_actionscheduler_claims: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_actionscheduler_groups: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_actionscheduler_logs: Data: 1.09MB + Index: 0.59MB + Engine InnoDB m7O0KqA_aelia_dismissed_messages: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_aelia_exchange_rates_history: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_aioseo_notifications: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_aioseo_posts: Data: 2.52MB + Index: 0.16MB + Engine InnoDB m7O0KqA_aws_cache: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_aws_index: Data: 27.38MB + Index: 13.59MB + Engine InnoDB m7O0KqA_chaty_contact_form_leads: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_cli_cookie_scan: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_cli_cookie_scan_categories: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_cli_cookie_scan_cookies: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_cli_cookie_scan_url: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_cli_scripts: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_commentmeta: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_comments: Data: 0.02MB + Index: 0.09MB + Engine InnoDB m7O0KqA_gdbc_attempts: Data: 0.34MB + Index: 0.09MB + Engine InnoDB m7O0KqA_links: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_ms_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_options: Data: 8.02MB + Index: 1.28MB + Engine InnoDB m7O0KqA_pmxi_files: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_hash: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_history: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_images: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_imports: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_posts: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_pmxi_templates: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_postmeta: Data: 84.47MB + Index: 47.09MB + Engine InnoDB m7O0KqA_posts: Data: 15.30MB + Index: 6.05MB + Engine InnoDB m7O0KqA_snippets: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_stock_log: Data: 2.52MB + Index: 0.00MB + Engine InnoDB m7O0KqA_termmeta: Data: 1.48MB + Index: 1.95MB + Engine InnoDB m7O0KqA_terms: Data: 0.48MB + Index: 0.52MB + Engine InnoDB m7O0KqA_term_relationships: Data: 5.02MB + Index: 4.17MB + Engine InnoDB m7O0KqA_term_taxonomy: Data: 1.52MB + Index: 0.63MB + Engine InnoDB m7O0KqA_usermeta: Data: 0.05MB + Index: 0.03MB + Engine InnoDB m7O0KqA_users: Data: 0.02MB + Index: 0.05MB + Engine InnoDB m7O0KqA_wcpdf_invoice_number: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_admin_notes: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_admin_note_actions: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_wc_category_lookup: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_customer_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_download_log: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_order_coupon_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_order_product_lookup: Data: 0.02MB + Index: 0.06MB + Engine InnoDB m7O0KqA_wc_order_stats: Data: 0.02MB + Index: 0.05MB + Engine InnoDB m7O0KqA_wc_order_tax_lookup: Data: 0.02MB + Index: 0.03MB + Engine InnoDB m7O0KqA_wc_product_meta_lookup: Data: 1.03MB + Index: 1.02MB + Engine InnoDB m7O0KqA_wc_reserved_stock: Data: 0.02MB + Index: 0.00MB + Engine InnoDB m7O0KqA_wc_tax_rate_classes: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_wc_webhooks: Data: 0.02MB + Index: 0.02MB + Engine InnoDB m7O0KqA_wss_log: Data: 0.27MB + Index: 0.00MB + Engine InnoDB ### Post Type Counts ### attachment: 13041 cmm4e_menu_theme: 1 cookielawinfo: 6 custom_css: 1 customize_changeset: 47 elementor_library: 9 iksm: 2 nav_menu_item: 63 page: 10 post: 2 product: 5751 revision: 344 seedprod: 2 shop_order: 7 shop_order_refund: 4 sp_wp_carousel: 1 wpmm_theme: 3 ### Security ### Secure connection (HTTPS): โœ” Hide errors from visitors: โŒError messages should not be shown to visitors. ### Active Plugins (21) ### WPBruiser: by Mihai Chelaru โ€“ 3.1.43 Advanced Woo Search: by ILLID โ€“ 2.24 All in One SEO: by All in One SEO Team โ€“ 4.0.17 Chaty: by Premio โ€“ 2.7.6 Code Snippets: by Code Snippets Pro โ€“ 2.14.1 GDPR Cookie Consent: by WebToffee โ€“ 2.0.1 Disable Real MIME Check: by Sergey Biryukov โ€“ 1.0 Elementor Pro: by Elementor.com โ€“ 3.2.0-dev1 Elementor: by Elementor.com โ€“ 3.2.0-dev4 Iks Menu: by IksStudio โ€“ 1.9.0 ID Back To Top: by Idenovasi โ€“ 1.1.1 UberMenu 3 - The Ultimate WordPress Mega Menu: by Chris Mavricos SevenSpark โ€“ 3.7.3 UpdraftPlus - Backup/Restore: by UpdraftPlus.Com DavidAnderson โ€“ 1.16.50 WooCommerce Weight Based Shipping: by weightbasedshipping.com โ€“ 5.3.10 Widget Options: by Widget Options Team โ€“ 3.7.7 Stock Sync for WooCommerce Pro: by Lauri Karisola / WooElements.com โ€“ 2.0.3 WooCommerce EU VAT Assistant: by Aelia โ€“ 2.0.14.210317 WooCommerce Stripe Gateway: by WooCommerce โ€“ 5.0.0 WooCommerce PDF Invoices & Packing Slips: by Ewout Fernhout โ€“ 2.8.2 WooCommerce Stock Manager: by StoreApps โ€“ 2.5.6 WooCommerce: by Automattic โ€“ 5.1.0 ### Inactive Plugins (0) ### ### Dropin Plugins (1) ### maintenance.php: maintenance.php ### Settings ### API Enabled: โœ” Force SSL: โ€“ Currency: EUR (โ‚ฌ) Currency Position: left Thousand Separator: . Decimal Separator: . Number of Decimals: 2 Taxonomies: Product Types: external (external) grouped (grouped) simple (simple) variable (variable) Taxonomies: Product Visibility: exclude-from-catalog (exclude-from-catalog) exclude-from-search (exclude-from-search) featured (featured) outofstock (outofstock) rated-1 (rated-1) rated-2 (rated-2) rated-3 (rated-3) rated-4 (rated-4) rated-5 (rated-5) Connected to WooCommerce.com: โ€“ ### WC Pages ### Shop base: #6 - /shop/ Cart: #7 - /cart/ Checkout: #8 - /checkout/ My account: #9 - /my-account/ Terms and conditions: #58575 - /elementor-58575/ ### Theme ### Name: Supermarket Ecommerce Version: 0.3.4 Author URL: https://www.luzuk.com/ Child Theme: โŒ โ€“ If you are modifying WooCommerce on a parent theme that you did not build personally we recommend using a child theme. See: How to create a child theme WooCommerce Support: โœ” ### Templates ### Overrides: supermarket-ecommerce/woocommerce/archive-product.php supermarket-ecommerce/woocommerce/checkout/form-checkout.php supermarket-ecommerce/woocommerce/global/wrapper-end.php supermarket-ecommerce/woocommerce/global/wrapper-start.php supermarket-ecommerce/woocommerce/single-product.php ### Action Scheduler ### Complete: 3,427 Oldest: 2021-02-27 16:52:31 +0000 Newest: 2021-03-30 16:00:17 +0000 Pending: 3 Oldest: 2021-03-30 16:00:47 +0000 Newest: 2021-03-31 15:31:05 +0000 ### Status report information ### Generated at: 2021-03-30 16:01:02 +00:00 ` ``` </details>
priority
product assigned to no category prerequisites mark completed items with an i have have carried out troubleshooting steps and i believe i have found a bug i have searched for similar bugs in both open and closed issues and cannot find a duplicate describe the bug i have a master store and a child store i exported the products from the master store using the woocommerce core import export feature i imported them into my new store imported all the categories and products i then decided i wanted to change the default category on my child store which i did i then removed some categories i did not want on my child store expecting my products within the categories i am removing to be moved to the to the default category however i noticed that the products were not in there due to naivety i had just assumed maybe the products were removed when i deleted the categories which i thought was strange as normally if a category is deleted then the products within that category are moved to the default category instead some time has passed and while i was working on a script today i noticed that the products are actually still on the site somewhere with no assigned category expected behavior i expected the products within the category i deleted to be moved to the default category actual behavior products are now on the site with no assigned category screenshots isolating the problem mark completed items with an i have deactivated other plugins and confirmed this bug occurs when only woocommerce plugin is active this bug happens with a default wordpress theme active or i can reproduce this bug consistently using the steps above wordpress environment we use the to help us evaluate the issue without this report we won t be able to fully evaluate this issue wordpress environment wordpress address url site address url wc version rest api version โœ” wc blocks version โœ” action scheduler version โœ” wc admin version โœ” log directory writable โœ” wp version wp multisite โ€“ wp memory limit mb wp debug mode โœ” wp cron โœ” language en us external object cache โ€“ server environment server info apache php version php post max size mb php time limit php max input vars curl version nss suhosin installed โ€“ mysql version mariadb max upload size mb default timezone is utc โœ” fsockopen curl โœ” soapclient โœ” domdocument โœ” gzip โœ” multibyte string โœ” remote post โœ” remote get โœ” database wc database version wc database prefix total database size database data size database index size woocommerce sessions data index engine innodb woocommerce api keys data index engine innodb woocommerce attribute taxonomies data index engine innodb woocommerce downloadable product permissions data index engine innodb woocommerce order items data index engine innodb woocommerce order itemmeta data index engine innodb woocommerce tax rates data index engine innodb woocommerce tax rate locations data index engine innodb woocommerce shipping zones data index engine innodb woocommerce shipping zone locations data index engine innodb woocommerce shipping zone methods data index engine innodb woocommerce payment tokens data index engine innodb woocommerce payment tokenmeta data index engine innodb woocommerce log data index engine innodb actionscheduler actions data index engine innodb actionscheduler claims data index engine innodb actionscheduler groups data index engine innodb actionscheduler logs data index engine innodb aelia dismissed messages data index engine innodb aelia exchange rates history data index engine innodb aioseo notifications data index engine innodb aioseo posts data index engine innodb aws cache data index engine innodb aws index data index engine innodb chaty contact form leads data index engine innodb cli cookie scan data index engine innodb cli cookie scan categories data index engine innodb cli cookie scan cookies data index engine innodb cli cookie scan url data index engine innodb cli scripts data index engine innodb commentmeta data index engine innodb comments data index engine innodb gdbc attempts data index engine innodb links data index engine innodb ms snippets data index engine innodb options data index engine innodb pmxi files data index engine innodb pmxi hash data index engine innodb pmxi history data index engine innodb pmxi images data index engine innodb pmxi imports data index engine innodb pmxi posts data index engine innodb pmxi templates data index engine innodb postmeta data index engine innodb posts data index engine innodb snippets data index engine innodb stock log data index engine innodb termmeta data index engine innodb terms data index engine innodb term relationships data index engine innodb term taxonomy data index engine innodb usermeta data index engine innodb users data index engine innodb wcpdf invoice number data index engine innodb wc admin notes data index engine innodb wc admin note actions data index engine innodb wc category lookup data index engine innodb wc customer lookup data index engine innodb wc download log data index engine innodb wc order coupon lookup data index engine innodb wc order product lookup data index engine innodb wc order stats data index engine innodb wc order tax lookup data index engine innodb wc product meta lookup data index engine innodb wc reserved stock data index engine innodb wc tax rate classes data index engine innodb wc webhooks data index engine innodb wss log data index engine innodb post type counts attachment menu theme cookielawinfo custom css customize changeset elementor library iksm nav menu item page post product revision seedprod shop order shop order refund sp wp carousel wpmm theme security secure connection https โœ” hide errors from visitors โŒerror messages should not be shown to visitors active plugins wpbruiser by mihai chelaru โ€“ advanced woo search by illid โ€“ all in one seo by all in one seo team โ€“ chaty by premio โ€“ code snippets by code snippets pro โ€“ gdpr cookie consent by webtoffee โ€“ disable real mime check by sergey biryukov โ€“ elementor pro by elementor com โ€“ elementor by elementor com โ€“ iks menu by iksstudio โ€“ id back to top by idenovasi โ€“ ubermenu the ultimate wordpress mega menu by chris mavricos sevenspark โ€“ updraftplus backup restore by updraftplus com davidanderson โ€“ woocommerce weight based shipping by weightbasedshipping com โ€“ widget options by widget options team โ€“ stock sync for woocommerce pro by lauri karisola wooelements com โ€“ woocommerce eu vat assistant by aelia โ€“ woocommerce stripe gateway by woocommerce โ€“ woocommerce pdf invoices packing slips by ewout fernhout โ€“ woocommerce stock manager by storeapps โ€“ woocommerce by automattic โ€“ inactive plugins dropin plugins maintenance php maintenance php settings api enabled โœ” force ssl โ€“ currency eur โ‚ฌ currency position left thousand separator decimal separator number of decimals taxonomies product types external external grouped grouped simple simple variable variable taxonomies product visibility exclude from catalog exclude from catalog exclude from search exclude from search featured featured outofstock outofstock rated rated rated rated rated rated rated rated rated rated connected to woocommerce com โ€“ wc pages shop base shop cart cart checkout checkout my account my account terms and conditions elementor theme name supermarket ecommerce version author url child theme โŒ โ€“ if you are modifying woocommerce on a parent theme that you did not build personally we recommend using a child theme see how to create a child theme woocommerce support โœ” templates overrides supermarket ecommerce woocommerce archive product php supermarket ecommerce woocommerce checkout form checkout php supermarket ecommerce woocommerce global wrapper end php supermarket ecommerce woocommerce global wrapper start php supermarket ecommerce woocommerce single product php action scheduler complete oldest newest pending oldest newest status report information generated at
1
231,642
7,641,630,614
IssuesEvent
2018-05-08 06:08:31
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
opened
Incorrect order of formal parameters are allowed for functions
Priority/Highest Severity/Blocker component/Compiler
**Description:** Order of the formal parameters should be as follows: - Required parameters - Defaultable parameters - Rest parameter In the below sample order of the formal parameters of function `foo` is not in the correct order. However it doesn't throw a compile time error. ```ballerina function main (string... args) { foo(45); } function foo(string s = "hello", int a) { io:println(s); io:println(a); } ```
1.0
Incorrect order of formal parameters are allowed for functions - **Description:** Order of the formal parameters should be as follows: - Required parameters - Defaultable parameters - Rest parameter In the below sample order of the formal parameters of function `foo` is not in the correct order. However it doesn't throw a compile time error. ```ballerina function main (string... args) { foo(45); } function foo(string s = "hello", int a) { io:println(s); io:println(a); } ```
priority
incorrect order of formal parameters are allowed for functions description order of the formal parameters should be as follows required parameters defaultable parameters rest parameter in the below sample order of the formal parameters of function foo is not in the correct order however it doesn t throw a compile time error ballerina function main string args foo function foo string s hello int a io println s io println a
1
283,342
8,718,965,979
IssuesEvent
2018-12-07 22:19:40
goharbor/harbor
https://api.github.com/repos/goharbor/harbor
reopened
Test LDAP server error without save configuration
priority/high target/1.7.0
[Steps] 1, admin - configuration 2, change auth to LDAP, and set up the ldap server. 3, test it with error response. 4, save it, test it with success response ![image](https://user-images.githubusercontent.com/2841473/49501209-4d04f000-f8ad-11e8-8837-d66d9ea1fc38.png) [Root cause] the ladp search password is missed in the payload of the request.
1.0
Test LDAP server error without save configuration - [Steps] 1, admin - configuration 2, change auth to LDAP, and set up the ldap server. 3, test it with error response. 4, save it, test it with success response ![image](https://user-images.githubusercontent.com/2841473/49501209-4d04f000-f8ad-11e8-8837-d66d9ea1fc38.png) [Root cause] the ladp search password is missed in the payload of the request.
priority
test ldap server error without save configuration admin configuration change auth to ldap and set up the ldap server test it with error response save it test it with success response the ladp search password is missed in the payload of the request
1
181,674
6,663,226,677
IssuesEvent
2017-10-02 15:40:28
classifiedz/classifiedz.github.io
https://api.github.com/repos/classifiedz/classifiedz.github.io
closed
Registered account cannot login
bug High Priority
When I login with my registered account, it shows "These credentials do not match our records." We have to figure out why this happened.
1.0
Registered account cannot login - When I login with my registered account, it shows "These credentials do not match our records." We have to figure out why this happened.
priority
registered account cannot login when i login with my registered account it shows these credentials do not match our records we have to figure out why this happened
1
316,205
9,638,705,123
IssuesEvent
2019-05-16 11:52:50
epam/cloud-pipeline
https://api.github.com/repos/epam/cloud-pipeline
closed
Allow to terminate paused runs
kind/enhancement priority/high state/verify sys/core sys/gui
**BACKGROUND** At the moment, when a run is paused - the user cannot terminate/stop it before resuming. This is the expected behavior. But in certain "bad" cases - it is not possible to resume a run: 1. An underlying instance is terminated outside of the Cloud Pipeline 2. Docker image was removed from the registry 3. And other cases that are not yet uncovered This introduces a number of stale runs, that just sit there in the PAUSED state and nobody can remove them. **TODOs** 1. For the PAUSED runs we shall allow OWNER or the run and the ADMIN users to terminate it 2. Termination of the PAUSED run - shall drop the underlying cloud instance and mark the run as STOPPED 3. From the GUI perspective - we shall not hide the STOP button, but replace it with TERMINATE, when a run is PAUSED (it shall not be available for the interim phases, e.g. PAUSING/RESUMING)
1.0
Allow to terminate paused runs - **BACKGROUND** At the moment, when a run is paused - the user cannot terminate/stop it before resuming. This is the expected behavior. But in certain "bad" cases - it is not possible to resume a run: 1. An underlying instance is terminated outside of the Cloud Pipeline 2. Docker image was removed from the registry 3. And other cases that are not yet uncovered This introduces a number of stale runs, that just sit there in the PAUSED state and nobody can remove them. **TODOs** 1. For the PAUSED runs we shall allow OWNER or the run and the ADMIN users to terminate it 2. Termination of the PAUSED run - shall drop the underlying cloud instance and mark the run as STOPPED 3. From the GUI perspective - we shall not hide the STOP button, but replace it with TERMINATE, when a run is PAUSED (it shall not be available for the interim phases, e.g. PAUSING/RESUMING)
priority
allow to terminate paused runs background at the moment when a run is paused the user cannot terminate stop it before resuming this is the expected behavior but in certain bad cases it is not possible to resume a run an underlying instance is terminated outside of the cloud pipeline docker image was removed from the registry and other cases that are not yet uncovered this introduces a number of stale runs that just sit there in the paused state and nobody can remove them todos for the paused runs we shall allow owner or the run and the admin users to terminate it termination of the paused run shall drop the underlying cloud instance and mark the run as stopped from the gui perspective we shall not hide the stop button but replace it with terminate when a run is paused it shall not be available for the interim phases e g pausing resuming
1
768,040
26,950,939,032
IssuesEvent
2023-02-08 11:37:54
Public-Health-Scotland/source-linkage-files
https://api.github.com/repos/Public-Health-Scotland/source-linkage-files
closed
Investigate non-matching SPSS vs R test outputs
bug Priority: High checking
Please check and update the spreadsheet on teams with the findings. Check instances where the SPSS and R versions do not match cost tests: - [x] Outpatients - [x] CMH - [x] Care Home
1.0
Investigate non-matching SPSS vs R test outputs - Please check and update the spreadsheet on teams with the findings. Check instances where the SPSS and R versions do not match cost tests: - [x] Outpatients - [x] CMH - [x] Care Home
priority
investigate non matching spss vs r test outputs please check and update the spreadsheet on teams with the findings check instances where the spss and r versions do not match cost tests outpatients cmh care home
1
688,480
23,584,019,078
IssuesEvent
2022-08-23 10:00:45
ut-issl/c2a-core
https://api.github.com/repos/ut-issl/c2a-core
closed
MOBC - AOBC SILSใ‚’ๆง‹็ฏ‰ใ—๏ผŒC2A้–“้€šไฟก็”จใฎใƒ†ใ‚นใƒˆใ‚’ๆ•ดๅ‚™ใ™ใ‚‹
priority::high
## ๆฆ‚่ฆ C2A้–“้€šไฟก็”จใฎใƒ†ใ‚นใƒˆใ‚’ๆ•ดๅ‚™ใ™ใ‚‹ ใคใ„ใงใซ๏ผŒMOBC (minimum user) - AOBC (2nd obc) ใจใฎSILS็’ฐๅขƒใ‚‚ๆง‹็ฏ‰ใ™ใ‚‹ ใƒ†ใ‚นใƒˆๆ–นๆณ•ใ‚‚ไธๅฏงใซๆ›ธใ ## closeๆกไปถ ใงใใŸใ‚‰
1.0
MOBC - AOBC SILSใ‚’ๆง‹็ฏ‰ใ—๏ผŒC2A้–“้€šไฟก็”จใฎใƒ†ใ‚นใƒˆใ‚’ๆ•ดๅ‚™ใ™ใ‚‹ - ## ๆฆ‚่ฆ C2A้–“้€šไฟก็”จใฎใƒ†ใ‚นใƒˆใ‚’ๆ•ดๅ‚™ใ™ใ‚‹ ใคใ„ใงใซ๏ผŒMOBC (minimum user) - AOBC (2nd obc) ใจใฎSILS็’ฐๅขƒใ‚‚ๆง‹็ฏ‰ใ™ใ‚‹ ใƒ†ใ‚นใƒˆๆ–นๆณ•ใ‚‚ไธๅฏงใซๆ›ธใ ## closeๆกไปถ ใงใใŸใ‚‰
priority
mobc aobc silsใ‚’ๆง‹็ฏ‰ใ—๏ผŒ ๆฆ‚่ฆ ใคใ„ใงใซ๏ผŒmobc minimum user aobc obc ใจใฎsils็’ฐๅขƒใ‚‚ๆง‹็ฏ‰ใ™ใ‚‹ ใƒ†ใ‚นใƒˆๆ–นๆณ•ใ‚‚ไธๅฏงใซๆ›ธใ closeๆกไปถ ใงใใŸใ‚‰
1
719,332
24,756,006,339
IssuesEvent
2022-10-21 17:48:09
carpentries/styles
https://api.github.com/repos/carpentries/styles
closed
[workflows] r-lib/actions/setup-r version needs to be updated
high priority
Originally reported in https://github.com/datacarpentry/spreadsheet-ecology-lesson/pull/317#issuecomment-1285542583, https://github.com/carpentries/styles/blob/6b8bbad1187c34b7ed321ed5d4fea0866cddadfc/.github/workflows/website.yml#L51-L53 and https://github.com/carpentries/styles/blob/6b8bbad1187c34b7ed321ed5d4fea0866cddadfc/.github/workflows/template.yml#L134-L136 are failing because `r-lib/actions/setup-r@master` no longer exists. ## Impact This impacts all repositories that were built with the styles template or the remote theme. - In all repositories, all pull requests will fail checks. - In repositories that use R Markdown, the lessons will fail to build. - Good news: In all other repositories, the **lessons will continue to build**. ## Solution Both of these actions calls need to be updated to use ```yaml uses: r-lib/actions/setup-r@v2 with: use-public-rspm: true install-r: false ``` Because this is an urgent update that involves six lines of workflow code that is not _normally_ touched by the maintainers, the deployment of this fix is better handled through mass pull requests that target these specific items in the workflow YAML. The reasoning for this is that we have estimated previously that manually making the pull requests to merge styles into these repositories would take roughly a day of full time work, assuming nothing goes wrong. Figuring out how to write GitHub API calls to fix this solution quickly is a much better use of time even if it creates merge conflicts with styles down the line.
1.0
[workflows] r-lib/actions/setup-r version needs to be updated - Originally reported in https://github.com/datacarpentry/spreadsheet-ecology-lesson/pull/317#issuecomment-1285542583, https://github.com/carpentries/styles/blob/6b8bbad1187c34b7ed321ed5d4fea0866cddadfc/.github/workflows/website.yml#L51-L53 and https://github.com/carpentries/styles/blob/6b8bbad1187c34b7ed321ed5d4fea0866cddadfc/.github/workflows/template.yml#L134-L136 are failing because `r-lib/actions/setup-r@master` no longer exists. ## Impact This impacts all repositories that were built with the styles template or the remote theme. - In all repositories, all pull requests will fail checks. - In repositories that use R Markdown, the lessons will fail to build. - Good news: In all other repositories, the **lessons will continue to build**. ## Solution Both of these actions calls need to be updated to use ```yaml uses: r-lib/actions/setup-r@v2 with: use-public-rspm: true install-r: false ``` Because this is an urgent update that involves six lines of workflow code that is not _normally_ touched by the maintainers, the deployment of this fix is better handled through mass pull requests that target these specific items in the workflow YAML. The reasoning for this is that we have estimated previously that manually making the pull requests to merge styles into these repositories would take roughly a day of full time work, assuming nothing goes wrong. Figuring out how to write GitHub API calls to fix this solution quickly is a much better use of time even if it creates merge conflicts with styles down the line.
priority
r lib actions setup r version needs to be updated originally reported in and are failing because r lib actions setup r master no longer exists impact this impacts all repositories that were built with the styles template or the remote theme in all repositories all pull requests will fail checks in repositories that use r markdown the lessons will fail to build good news in all other repositories the lessons will continue to build solution both of these actions calls need to be updated to use yaml uses r lib actions setup r with use public rspm true install r false because this is an urgent update that involves six lines of workflow code that is not normally touched by the maintainers the deployment of this fix is better handled through mass pull requests that target these specific items in the workflow yaml the reasoning for this is that we have estimated previously that manually making the pull requests to merge styles into these repositories would take roughly a day of full time work assuming nothing goes wrong figuring out how to write github api calls to fix this solution quickly is a much better use of time even if it creates merge conflicts with styles down the line
1
315,484
9,621,232,398
IssuesEvent
2019-05-14 10:08:00
UKHackspaceFoundation/website
https://api.github.com/repos/UKHackspaceFoundation/website
closed
Allow individual members to join
High Priority in progress
People should be able to join the organisation as individual members, including: * General signup (I think this still needs to include postal address, etc) with agreement to code of conduct * Details of why we should let the person in as a member * GoCardless direct debit mandate setup We then need a way for admins to approve new memberships. Once a membership is approved, we need a method of scheduling direct debit payments (which can be used by organisation memberships too). We need to give 3 days' notice of a payment being taken. I think we decided that an individual membership would be ยฃ20/year, but we should make this configurable.
1.0
Allow individual members to join - People should be able to join the organisation as individual members, including: * General signup (I think this still needs to include postal address, etc) with agreement to code of conduct * Details of why we should let the person in as a member * GoCardless direct debit mandate setup We then need a way for admins to approve new memberships. Once a membership is approved, we need a method of scheduling direct debit payments (which can be used by organisation memberships too). We need to give 3 days' notice of a payment being taken. I think we decided that an individual membership would be ยฃ20/year, but we should make this configurable.
priority
allow individual members to join people should be able to join the organisation as individual members including general signup i think this still needs to include postal address etc with agreement to code of conduct details of why we should let the person in as a member gocardless direct debit mandate setup we then need a way for admins to approve new memberships once a membership is approved we need a method of scheduling direct debit payments which can be used by organisation memberships too we need to give days notice of a payment being taken i think we decided that an individual membership would be ยฃ year but we should make this configurable
1
162,466
6,153,970,306
IssuesEvent
2017-06-28 11:25:22
fossasia/open-event-orga-server
https://api.github.com/repos/fossasia/open-event-orga-server
closed
Handle invalid data while change data types during migrations
bug database has-PR Priority: High Priority: URGENT refactor
**Related error:** ``` sqlalchemy.exc.IntegrityError: (psycopg2.IntegrityError) column "pdf_url" contains null values [SQL: 'ALTER TABLE ticket_holders ADD COLUMN pdf_url VARCHAR NOT NULL'] ``` Due to this, migration are not running on `api.eventyay.com`. Ensure existing data (valid or not) is handled properly and also ensure columns have proper null setting. (for example here, make pdf_url nullable.) Similarly check and ensure other columns as well **Full log attached for reference:** ``` 2017-06-28T05:49:05.921467228Z run_migrations_online() 2017-06-28T05:49:05.921470119Z File "migrations/env.py", line 151, in run_migrations_online 2017-06-28T05:49:05.921473075Z context.run_migrations() 2017-06-28T05:49:05.921475912Z File "<string>", line 8, in run_migrations 2017-06-28T05:49:05.921479088Z File "/usr/local/lib/python2.7/site-packages/alembic/runtime/environment.py", line 817, in run_migrations 2017-06-28T05:49:05.922610188Z self.get_context().run_migrations(**kw) 2017-06-28T05:49:05.922642643Z File "/usr/local/lib/python2.7/site-packages/alembic/runtime/migration.py", line 329, in run_migrations 2017-06-28T05:49:05.923243091Z step.migration_fn(**kw) 2017-06-28T05:49:05.923274018Z File "/opev/open_event/migrations/versions/c6b183975be9_.py", line 20, in upgrade 2017-06-28T05:49:05.923338965Z op.add_column('ticket_holders', sa.Column('pdf_url', sa.String(), nullable=False)) 2017-06-28T05:49:05.923364414Z File "<string>", line 8, in add_column 2017-06-28T05:49:05.923427981Z File "<string>", line 3, in add_column 2017-06-28T05:49:05.923478561Z File "/usr/local/lib/python2.7/site-packages/alembic/operations/ops.py", line 1551, in add_column 2017-06-28T05:49:05.926653441Z return operations.invoke(op) 2017-06-28T05:49:05.926712969Z File "/usr/local/lib/python2.7/site-packages/alembic/operations/base.py", line 318, in invoke 2017-06-28T05:49:05.927353504Z return fn(self, operation) 2017-06-28T05:49:05.927386093Z File "/usr/local/lib/python2.7/site-packages/alembic/operations/toimpl.py", line 123, in add_column 2017-06-28T05:49:05.927969743Z schema=schema 2017-06-28T05:49:05.927998827Z File "/usr/local/lib/python2.7/site-packages/alembic/ddl/impl.py", line 172, in add_column 2017-06-28T05:49:05.92861825Z self._exec(base.AddColumn(table_name, column, schema=schema)) 2017-06-28T05:49:05.92864919Z File "/usr/local/lib/python2.7/site-packages/alembic/ddl/impl.py", line 118, in _exec 2017-06-28T05:49:05.928742734Z return conn.execute(construct, *multiparams, **params) 2017-06-28T05:49:05.928786028Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 945, in execute 2017-06-28T05:49:05.929323557Z return meth(self, multiparams, params) 2017-06-28T05:49:05.929333376Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/sql/ddl.py", line 68, in _execute_on_connection 2017-06-28T05:49:05.929336744Z return connection._execute_ddl(self, multiparams, params) 2017-06-28T05:49:05.929339861Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1002, in _execute_ddl 2017-06-28T05:49:05.929512493Z compiled 2017-06-28T05:49:05.929538983Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1189, in _execute_context 2017-06-28T05:49:05.929768208Z context) 2017-06-28T05:49:05.929810133Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1402, in _handle_dbapi_exception 2017-06-28T05:49:05.930067186Z exc_info 2017-06-28T05:49:05.930093798Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/util/compat.py", line 203, in raise_from_cause 2017-06-28T05:49:05.930194023Z reraise(type(exception), exception, tb=exc_tb, cause=cause) 2017-06-28T05:49:05.930237111Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1182, in _execute_context 2017-06-28T05:49:05.931176835Z context) 2017-06-28T05:49:05.931187737Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/default.py", line 470, in do_execute 2017-06-28T05:49:05.931191369Z cursor.execute(statement, parameters) 2017-06-28T05:49:05.931334485Z sqlalchemy.exc.IntegrityError: (psycopg2.IntegrityError) column "pdf_url" contains null values 2017-06-28T05:49:05.931342263Z [SQL: 'ALTER TABLE ticket_holders ADD COLUMN pdf_url VARCHAR NOT NULL'] ```
2.0
Handle invalid data while change data types during migrations - **Related error:** ``` sqlalchemy.exc.IntegrityError: (psycopg2.IntegrityError) column "pdf_url" contains null values [SQL: 'ALTER TABLE ticket_holders ADD COLUMN pdf_url VARCHAR NOT NULL'] ``` Due to this, migration are not running on `api.eventyay.com`. Ensure existing data (valid or not) is handled properly and also ensure columns have proper null setting. (for example here, make pdf_url nullable.) Similarly check and ensure other columns as well **Full log attached for reference:** ``` 2017-06-28T05:49:05.921467228Z run_migrations_online() 2017-06-28T05:49:05.921470119Z File "migrations/env.py", line 151, in run_migrations_online 2017-06-28T05:49:05.921473075Z context.run_migrations() 2017-06-28T05:49:05.921475912Z File "<string>", line 8, in run_migrations 2017-06-28T05:49:05.921479088Z File "/usr/local/lib/python2.7/site-packages/alembic/runtime/environment.py", line 817, in run_migrations 2017-06-28T05:49:05.922610188Z self.get_context().run_migrations(**kw) 2017-06-28T05:49:05.922642643Z File "/usr/local/lib/python2.7/site-packages/alembic/runtime/migration.py", line 329, in run_migrations 2017-06-28T05:49:05.923243091Z step.migration_fn(**kw) 2017-06-28T05:49:05.923274018Z File "/opev/open_event/migrations/versions/c6b183975be9_.py", line 20, in upgrade 2017-06-28T05:49:05.923338965Z op.add_column('ticket_holders', sa.Column('pdf_url', sa.String(), nullable=False)) 2017-06-28T05:49:05.923364414Z File "<string>", line 8, in add_column 2017-06-28T05:49:05.923427981Z File "<string>", line 3, in add_column 2017-06-28T05:49:05.923478561Z File "/usr/local/lib/python2.7/site-packages/alembic/operations/ops.py", line 1551, in add_column 2017-06-28T05:49:05.926653441Z return operations.invoke(op) 2017-06-28T05:49:05.926712969Z File "/usr/local/lib/python2.7/site-packages/alembic/operations/base.py", line 318, in invoke 2017-06-28T05:49:05.927353504Z return fn(self, operation) 2017-06-28T05:49:05.927386093Z File "/usr/local/lib/python2.7/site-packages/alembic/operations/toimpl.py", line 123, in add_column 2017-06-28T05:49:05.927969743Z schema=schema 2017-06-28T05:49:05.927998827Z File "/usr/local/lib/python2.7/site-packages/alembic/ddl/impl.py", line 172, in add_column 2017-06-28T05:49:05.92861825Z self._exec(base.AddColumn(table_name, column, schema=schema)) 2017-06-28T05:49:05.92864919Z File "/usr/local/lib/python2.7/site-packages/alembic/ddl/impl.py", line 118, in _exec 2017-06-28T05:49:05.928742734Z return conn.execute(construct, *multiparams, **params) 2017-06-28T05:49:05.928786028Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 945, in execute 2017-06-28T05:49:05.929323557Z return meth(self, multiparams, params) 2017-06-28T05:49:05.929333376Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/sql/ddl.py", line 68, in _execute_on_connection 2017-06-28T05:49:05.929336744Z return connection._execute_ddl(self, multiparams, params) 2017-06-28T05:49:05.929339861Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1002, in _execute_ddl 2017-06-28T05:49:05.929512493Z compiled 2017-06-28T05:49:05.929538983Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1189, in _execute_context 2017-06-28T05:49:05.929768208Z context) 2017-06-28T05:49:05.929810133Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1402, in _handle_dbapi_exception 2017-06-28T05:49:05.930067186Z exc_info 2017-06-28T05:49:05.930093798Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/util/compat.py", line 203, in raise_from_cause 2017-06-28T05:49:05.930194023Z reraise(type(exception), exception, tb=exc_tb, cause=cause) 2017-06-28T05:49:05.930237111Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/base.py", line 1182, in _execute_context 2017-06-28T05:49:05.931176835Z context) 2017-06-28T05:49:05.931187737Z File "/usr/local/lib/python2.7/site-packages/sqlalchemy/engine/default.py", line 470, in do_execute 2017-06-28T05:49:05.931191369Z cursor.execute(statement, parameters) 2017-06-28T05:49:05.931334485Z sqlalchemy.exc.IntegrityError: (psycopg2.IntegrityError) column "pdf_url" contains null values 2017-06-28T05:49:05.931342263Z [SQL: 'ALTER TABLE ticket_holders ADD COLUMN pdf_url VARCHAR NOT NULL'] ```
priority
handle invalid data while change data types during migrations related error sqlalchemy exc integrityerror integrityerror column pdf url contains null values due to this migration are not running on api eventyay com ensure existing data valid or not is handled properly and also ensure columns have proper null setting for example here make pdf url nullable similarly check and ensure other columns as well full log attached for reference run migrations online file migrations env py line in run migrations online context run migrations file line in run migrations file usr local lib site packages alembic runtime environment py line in run migrations self get context run migrations kw file usr local lib site packages alembic runtime migration py line in run migrations step migration fn kw file opev open event migrations versions py line in upgrade op add column ticket holders sa column pdf url sa string nullable false file line in add column file line in add column file usr local lib site packages alembic operations ops py line in add column return operations invoke op file usr local lib site packages alembic operations base py line in invoke return fn self operation file usr local lib site packages alembic operations toimpl py line in add column schema schema file usr local lib site packages alembic ddl impl py line in add column self exec base addcolumn table name column schema schema file usr local lib site packages alembic ddl impl py line in exec return conn execute construct multiparams params file usr local lib site packages sqlalchemy engine base py line in execute return meth self multiparams params file usr local lib site packages sqlalchemy sql ddl py line in execute on connection return connection execute ddl self multiparams params file usr local lib site packages sqlalchemy engine base py line in execute ddl compiled file usr local lib site packages sqlalchemy engine base py line in execute context context file usr local lib site packages sqlalchemy engine base py line in handle dbapi exception exc info file usr local lib site packages sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file usr local lib site packages sqlalchemy engine base py line in execute context context file usr local lib site packages sqlalchemy engine default py line in do execute cursor execute statement parameters sqlalchemy exc integrityerror integrityerror column pdf url contains null values
1
517,502
15,014,971,313
IssuesEvent
2021-02-01 07:35:03
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
groups.google.com - desktop site instead of mobile site
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 87.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:87.0) Gecko/87.0 Firefox/87.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/66525 --> **URL**: https://groups.google.com/forum/#!forum/firebase-talk **Browser / Version**: Firefox Mobile 87.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes Internet Explorer **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: https://groups.google.com/forum/#!forum/firebase-talk%7c%20category%3acommunication%20%7c%20category%3aassignment_workflow%20%7c%20category%3agrading_workflow%20%7c%20category%3astudent_questions)%20(detail%3aplatform%3aweb%20%7c%20detail%3aplatform%3aa,another one gets fail <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/19447dc0-6e4a-4720-914e-64f6fc44e871.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with โค๏ธ_
1.0
groups.google.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 87.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:87.0) Gecko/87.0 Firefox/87.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/66525 --> **URL**: https://groups.google.com/forum/#!forum/firebase-talk **Browser / Version**: Firefox Mobile 87.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes Internet Explorer **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: https://groups.google.com/forum/#!forum/firebase-talk%7c%20category%3acommunication%20%7c%20category%3aassignment_workflow%20%7c%20category%3agrading_workflow%20%7c%20category%3astudent_questions)%20(detail%3aplatform%3aweb%20%7c%20detail%3aplatform%3aa,another one gets fail <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/1/19447dc0-6e4a-4720-914e-64f6fc44e871.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with โค๏ธ_
priority
groups google com desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes internet explorer problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce one gets fail view the screenshot img alt screenshot src browser configuration none from with โค๏ธ
1
194,381
6,894,401,441
IssuesEvent
2017-11-23 09:51:55
hackersatcambridge/hac-website
https://api.github.com/repos/hackersatcambridge/hac-website
opened
Model component for Workshops processing
backend enhancement priority high
## What the enhancement is Make a new Workshop model that can represent all of the information in https://github.com/hackersatcambridge/workshop-example (e.g. description.md, setup_instructions.md etc.) ## Why it could be helpful We'll need this for #40
1.0
Model component for Workshops processing - ## What the enhancement is Make a new Workshop model that can represent all of the information in https://github.com/hackersatcambridge/workshop-example (e.g. description.md, setup_instructions.md etc.) ## Why it could be helpful We'll need this for #40
priority
model component for workshops processing what the enhancement is make a new workshop model that can represent all of the information in e g description md setup instructions md etc why it could be helpful we ll need this for
1
284,419
8,738,258,522
IssuesEvent
2018-12-12 02:14:51
uvasomrc/ithriv
https://api.github.com/repos/uvasomrc/ithriv
closed
Help pages: Add images and update text
0 - High Priority
Grab updated content and images from the resources listed at https://portal.ithriv.org/#/category/1028
1.0
Help pages: Add images and update text - Grab updated content and images from the resources listed at https://portal.ithriv.org/#/category/1028
priority
help pages add images and update text grab updated content and images from the resources listed at
1
436,334
12,550,348,953
IssuesEvent
2020-06-06 10:47:22
cpeditor/cpeditor
https://api.github.com/repos/cpeditor/cpeditor
closed
Ask for whether to override an existing snippet when entering a used name
enhancement good first issue help wanted high_priority
**Is your feature request related to a problem? Please describe.** Now the user is forced to choose a new name for a snippet if the entered name is the same as an existing snippet. But the user may want to override the old snippet. **Describe the solution you'd like** Ask for whether to override an existing snippet when entering a used name. **Describe alternatives you've considered** N/A **Additional context** N/A
1.0
Ask for whether to override an existing snippet when entering a used name - **Is your feature request related to a problem? Please describe.** Now the user is forced to choose a new name for a snippet if the entered name is the same as an existing snippet. But the user may want to override the old snippet. **Describe the solution you'd like** Ask for whether to override an existing snippet when entering a used name. **Describe alternatives you've considered** N/A **Additional context** N/A
priority
ask for whether to override an existing snippet when entering a used name is your feature request related to a problem please describe now the user is forced to choose a new name for a snippet if the entered name is the same as an existing snippet but the user may want to override the old snippet describe the solution you d like ask for whether to override an existing snippet when entering a used name describe alternatives you ve considered n a additional context n a
1
472,950
13,633,866,687
IssuesEvent
2020-09-24 22:21:10
joe27g/EnhancedDiscord
https://api.github.com/repos/joe27g/EnhancedDiscord
closed
Can't watch streams on Linux
category: core priority: high status: in progress type: bug
**Steps to reproduce** 1. Enter a voice channel when one or more people are streaming 2. Pop out the player in a new window 3. Click one of the stream thumbnails 4. It shows the loading animation indefinitely 5. Closing the window or ending the call will cause Discord to crash This does not happen and streams are watchable when the player is still inside the main window. This does not happen when ED is not injected. This still happens with all plugins disabled. **Desktop info** OS: Ubuntu 20.04.1 Window manager: i3 ED: master (`71f79a0`) Discord version: > Stable 65705 (d72a93f) > Host 0.0.11 > Linux 64-Bit (5.4.0-42-Generic)
1.0
Can't watch streams on Linux - **Steps to reproduce** 1. Enter a voice channel when one or more people are streaming 2. Pop out the player in a new window 3. Click one of the stream thumbnails 4. It shows the loading animation indefinitely 5. Closing the window or ending the call will cause Discord to crash This does not happen and streams are watchable when the player is still inside the main window. This does not happen when ED is not injected. This still happens with all plugins disabled. **Desktop info** OS: Ubuntu 20.04.1 Window manager: i3 ED: master (`71f79a0`) Discord version: > Stable 65705 (d72a93f) > Host 0.0.11 > Linux 64-Bit (5.4.0-42-Generic)
priority
can t watch streams on linux steps to reproduce enter a voice channel when one or more people are streaming pop out the player in a new window click one of the stream thumbnails it shows the loading animation indefinitely closing the window or ending the call will cause discord to crash this does not happen and streams are watchable when the player is still inside the main window this does not happen when ed is not injected this still happens with all plugins disabled desktop info os ubuntu window manager ed master discord version stable host linux bit generic
1
467,971
13,459,106,167
IssuesEvent
2020-09-09 11:44:42
slsdetectorgroup/slsDetectorPackage
https://api.github.com/repos/slsdetectorgroup/slsDetectorPackage
closed
M3 Dead-time free mode
action - Enhancement priority - Super high status - On it
<!-- Preview changes before submitting --> <!-- Please fill out everything with an *, as this report will be discarded otherwise --> <!-- This is a comment, the syntax is a bit different from c++ or bash --> ##### *Detector type: <!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard --> MIII ##### *Software Package Version: <!-- developer, 4.2.0, 4.1.1, etc --> ##### Priority: <!-- Super Low, Low, Medium, High, Super High --> Medium ##### *State the feature: <!-- A clear and concise description of what the feature is --> Allowing for turning on the dead time-free mode that allows for parallelizing the execution of exposure, readout and streaming of data. The min. exp. time is 18.8us for run & rdo clk = 125Mhz (division factor 8). ##### Is your feature request related to a problem. Please describe: <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> ##### Describe the solution you'd like: <!-- A clear and concise description of what you want to happen --> A command for enable dead-time free mode : enabling 0 bit at the address 0x580 For checking out errors look at the 4th bit of the same address. If 1 then the exposure was too short. ##### Describe alternatives you've considered: <!-- A clear and concise description of any alternative solutions or features you've considered --> ##### Additional context: <!-- Add any other context about the feature here --> Please find the register description below: Description of the dead-time free controller: section 12 : https://www.overleaf.com/read/tmsbgfwznqpj @anberga
1.0
M3 Dead-time free mode - <!-- Preview changes before submitting --> <!-- Please fill out everything with an *, as this report will be discarded otherwise --> <!-- This is a comment, the syntax is a bit different from c++ or bash --> ##### *Detector type: <!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard --> MIII ##### *Software Package Version: <!-- developer, 4.2.0, 4.1.1, etc --> ##### Priority: <!-- Super Low, Low, Medium, High, Super High --> Medium ##### *State the feature: <!-- A clear and concise description of what the feature is --> Allowing for turning on the dead time-free mode that allows for parallelizing the execution of exposure, readout and streaming of data. The min. exp. time is 18.8us for run & rdo clk = 125Mhz (division factor 8). ##### Is your feature request related to a problem. Please describe: <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> ##### Describe the solution you'd like: <!-- A clear and concise description of what you want to happen --> A command for enable dead-time free mode : enabling 0 bit at the address 0x580 For checking out errors look at the 4th bit of the same address. If 1 then the exposure was too short. ##### Describe alternatives you've considered: <!-- A clear and concise description of any alternative solutions or features you've considered --> ##### Additional context: <!-- Add any other context about the feature here --> Please find the register description below: Description of the dead-time free controller: section 12 : https://www.overleaf.com/read/tmsbgfwznqpj @anberga
priority
dead time free mode detector type miii software package version priority medium state the feature allowing for turning on the dead time free mode that allows for parallelizing the execution of exposure readout and streaming of data the min exp time is for run rdo clk division factor is your feature request related to a problem please describe describe the solution you d like a command for enable dead time free mode enabling bit at the address for checking out errors look at the bit of the same address if then the exposure was too short describe alternatives you ve considered additional context please find the register description below description of the dead time free controller section anberga
1
597,184
18,157,325,979
IssuesEvent
2021-09-27 04:35:09
Azordev/dasher-user
https://api.github.com/repos/Azordev/dasher-user
opened
Define folder and files structure
help wanted question priority high
There is a lot of ways to structure folders. Some usual folder and files we can expect are: - components: shared atomic components + index.js + ComponentName.jsx + ComponentName.style.jsx + ComponentName.docs.mdx + ComponentName * ChildComponent.jsx - styles: shared and global styles, including themes + index.js + themes.js + global.css + variables.css + other.css - services: external API services code + index.js + Service * index.js * file.js - utils: reusable js functions + index.js + file.js - pages: pages for each routes, and the router + router.js + AuthRedirect.jsx + Home.jsx + Page.jsx - layouts: reusable layouts for mobile and desktop + index.js + SinglePane.js + SideBar.js - templates: shared groups of components that works together + index.js + Form.js + Table.js What about this project? Any idea how to structure folders and files?
1.0
Define folder and files structure - There is a lot of ways to structure folders. Some usual folder and files we can expect are: - components: shared atomic components + index.js + ComponentName.jsx + ComponentName.style.jsx + ComponentName.docs.mdx + ComponentName * ChildComponent.jsx - styles: shared and global styles, including themes + index.js + themes.js + global.css + variables.css + other.css - services: external API services code + index.js + Service * index.js * file.js - utils: reusable js functions + index.js + file.js - pages: pages for each routes, and the router + router.js + AuthRedirect.jsx + Home.jsx + Page.jsx - layouts: reusable layouts for mobile and desktop + index.js + SinglePane.js + SideBar.js - templates: shared groups of components that works together + index.js + Form.js + Table.js What about this project? Any idea how to structure folders and files?
priority
define folder and files structure there is a lot of ways to structure folders some usual folder and files we can expect are components shared atomic components index js componentname jsx componentname style jsx componentname docs mdx componentname childcomponent jsx styles shared and global styles including themes index js themes js global css variables css other css services external api services code index js service index js file js utils reusable js functions index js file js pages pages for each routes and the router router js authredirect jsx home jsx page jsx layouts reusable layouts for mobile and desktop index js singlepane js sidebar js templates shared groups of components that works together index js form js table js what about this project any idea how to structure folders and files
1
561,649
16,621,521,179
IssuesEvent
2021-06-03 02:14:51
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Trying to create patient when doing regular dispensing gives RSOD
Bug: development Priority: high
## Describe the bug May have broken with one of the upgrades or recent changes. https://user-images.githubusercontent.com/65875762/120146642-7e21c980-c239-11eb-92e2-2d996c835393.mp4 ### To reproduce Steps to reproduce the behaviour: 1. Go to Dispensing 2. Click on New Patient 3. Observe error ### Expected behaviour Should open new patient modal ### Proposed Solution Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase. ### Version and device info - App version: 8.1.0 (develop) - Tablet model: N/A - OS version: N/A ### Additional context New patient creation works fine under vaccinations module
1.0
Trying to create patient when doing regular dispensing gives RSOD - ## Describe the bug May have broken with one of the upgrades or recent changes. https://user-images.githubusercontent.com/65875762/120146642-7e21c980-c239-11eb-92e2-2d996c835393.mp4 ### To reproduce Steps to reproduce the behaviour: 1. Go to Dispensing 2. Click on New Patient 3. Observe error ### Expected behaviour Should open new patient modal ### Proposed Solution Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase. ### Version and device info - App version: 8.1.0 (develop) - Tablet model: N/A - OS version: N/A ### Additional context New patient creation works fine under vaccinations module
priority
trying to create patient when doing regular dispensing gives rsod describe the bug may have broken with one of the upgrades or recent changes to reproduce steps to reproduce the behaviour go to dispensing click on new patient observe error expected behaviour should open new patient modal proposed solution leave if you don t know how to fix implement edit this issue description and explain here if you know the best path of implementing the fix within the codebase version and device info app version develop tablet model n a os version n a additional context new patient creation works fine under vaccinations module
1
733,106
25,288,221,317
IssuesEvent
2022-11-16 21:16:26
Automattic/woocommerce-payments
https://api.github.com/repos/Automattic/woocommerce-payments
opened
Expose charge ID in order for use in dispute evidence
type: bug priority: high
### Describe the bug Submitting evidence for a "Duplicate" dispute type requires [entering the charge ID](https://woocommerce.com/document/woocommerce-payments/fraud-and-disputes/managing-disputes-with-woocommerce-payments/#section-6) (ch_) for the other successful charge. However, that charge ID is not exposed in the orderโ€”rather, the pi_ is shown. <img width="950" alt="image" src="https://user-images.githubusercontent.com/35117248/202295846-6d8402e4-6d59-45a1-8dfe-00af16a10452.png"> ### Additional context p1668628113535099-slack-C3NCP7ZJ6
1.0
Expose charge ID in order for use in dispute evidence - ### Describe the bug Submitting evidence for a "Duplicate" dispute type requires [entering the charge ID](https://woocommerce.com/document/woocommerce-payments/fraud-and-disputes/managing-disputes-with-woocommerce-payments/#section-6) (ch_) for the other successful charge. However, that charge ID is not exposed in the orderโ€”rather, the pi_ is shown. <img width="950" alt="image" src="https://user-images.githubusercontent.com/35117248/202295846-6d8402e4-6d59-45a1-8dfe-00af16a10452.png"> ### Additional context p1668628113535099-slack-C3NCP7ZJ6
priority
expose charge id in order for use in dispute evidence describe the bug submitting evidence for a duplicate dispute type requires ch for the other successful charge however that charge id is not exposed in the orderโ€”rather the pi is shown img width alt image src additional context slack
1
718,284
24,710,165,546
IssuesEvent
2022-10-19 23:26:44
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Add unread @-mention indicator to Recent topics
help wanted in progress priority: high area: recent-topics release goal
In #21637 / #22583, we added an `@` in the left sidebar to indicate which streams/topics have unread @-mentions. We should do the same in Recent topics. The `@` should be to the left of the unread counter, with styling to match the `@` in the left sidebar. To match the left sidebar, we should not show `@` for private messages even after #20890 is integrated. This will likely be a follow-up.
1.0
Add unread @-mention indicator to Recent topics - In #21637 / #22583, we added an `@` in the left sidebar to indicate which streams/topics have unread @-mentions. We should do the same in Recent topics. The `@` should be to the left of the unread counter, with styling to match the `@` in the left sidebar. To match the left sidebar, we should not show `@` for private messages even after #20890 is integrated. This will likely be a follow-up.
priority
add unread mention indicator to recent topics in we added an in the left sidebar to indicate which streams topics have unread mentions we should do the same in recent topics the should be to the left of the unread counter with styling to match the in the left sidebar to match the left sidebar we should not show for private messages even after is integrated this will likely be a follow up
1
543,264
15,879,311,373
IssuesEvent
2021-04-09 12:17:18
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
[Doc] No mention of what values need to specify for grant types when using 'OAuth/OpenID Connect' admin APIs
Affected/5.4.0 Priority/High Severity/Major docs
When automating OIDC/Oauth2 flows by refering [1], user has to specify parameters which included in request payload under 'Configuring OAuth/OpenID Connect' section. But to specify Grant types, there's no mention of values which need to be filled within the payload. According to the Admin API (OAuthAdminService), following values can be used within '<xsd1:grantTypes>?</xsd1:grantTypes>' - refresh_token - urn:ietf:params:oauth:grant-type:saml2-bearer - implicit - password - client_credentials - iwa:ntlm, authorization_code - urn:ietf:params:oauth:grant-type:jwt-bearer [1] https://docs.wso2.com/display/IS540/Service+Provider+Configurations+used+with+APIs#ServiceProviderConfigurationsusedwithAPIs-ConfiguringOAuth/OpenIDConnect
1.0
[Doc] No mention of what values need to specify for grant types when using 'OAuth/OpenID Connect' admin APIs - When automating OIDC/Oauth2 flows by refering [1], user has to specify parameters which included in request payload under 'Configuring OAuth/OpenID Connect' section. But to specify Grant types, there's no mention of values which need to be filled within the payload. According to the Admin API (OAuthAdminService), following values can be used within '<xsd1:grantTypes>?</xsd1:grantTypes>' - refresh_token - urn:ietf:params:oauth:grant-type:saml2-bearer - implicit - password - client_credentials - iwa:ntlm, authorization_code - urn:ietf:params:oauth:grant-type:jwt-bearer [1] https://docs.wso2.com/display/IS540/Service+Provider+Configurations+used+with+APIs#ServiceProviderConfigurationsusedwithAPIs-ConfiguringOAuth/OpenIDConnect
priority
no mention of what values need to specify for grant types when using oauth openid connect admin apis when automating oidc flows by refering user has to specify parameters which included in request payload under configuring oauth openid connect section but to specify grant types there s no mention of values which need to be filled within the payload according to the admin api oauthadminservice following values can be used within refresh token urn ietf params oauth grant type bearer implicit password client credentials iwa ntlm authorization code urn ietf params oauth grant type jwt bearer
1
399,886
11,762,469,419
IssuesEvent
2020-03-14 01:33:32
cpeditor/cpeditor
https://api.github.com/repos/cpeditor/cpeditor
opened
Bind files with problem URLs
enhancement high_priority
**Is your feature request related to a problem? Please describe.** Sometimes I don't remember where the file to a problem URL is saved, or I don't want to search in the file manager by myself. Also, the problem URL is lost after closing the tab. **Describe the solution you'd like** Add an option to bind files with problem URLs. When files are reopened, the problem URLs will remain. Add an option to open the old files with the problem URL from Competitive Companion. **Describe alternatives you've considered** N/A **Additional context** From #198.
1.0
Bind files with problem URLs - **Is your feature request related to a problem? Please describe.** Sometimes I don't remember where the file to a problem URL is saved, or I don't want to search in the file manager by myself. Also, the problem URL is lost after closing the tab. **Describe the solution you'd like** Add an option to bind files with problem URLs. When files are reopened, the problem URLs will remain. Add an option to open the old files with the problem URL from Competitive Companion. **Describe alternatives you've considered** N/A **Additional context** From #198.
priority
bind files with problem urls is your feature request related to a problem please describe sometimes i don t remember where the file to a problem url is saved or i don t want to search in the file manager by myself also the problem url is lost after closing the tab describe the solution you d like add an option to bind files with problem urls when files are reopened the problem urls will remain add an option to open the old files with the problem url from competitive companion describe alternatives you ve considered n a additional context from
1
787,161
27,708,663,944
IssuesEvent
2023-03-14 12:57:40
AY2223S2-CS2113-T12-1/tp
https://api.github.com/repos/AY2223S2-CS2113-T12-1/tp
closed
Add balance method
type.Task priority.High
Allows user to input the command: balance To get a total income - total expenses To solve issue #36
1.0
Add balance method - Allows user to input the command: balance To get a total income - total expenses To solve issue #36
priority
add balance method allows user to input the command balance to get a total income total expenses to solve issue
1
248,596
7,934,176,479
IssuesEvent
2018-07-08 16:03:36
byaka/VombatiDB
https://api.github.com/repos/byaka/VombatiDB
opened
ะ˜ะฝะบะพะฝัะธัั‚ะตะฝั‚ะฝะพัั‚ัŒ ะฟั€ะธ ัƒะดะฐะปะตะฝะธะธ ะพะฑัŒะตะบั‚ะฐ, ะฝะฐ ะบะพั‚ะพั€ั‹ะน ะฑั‹ะปะฐ ััั‹ะปะบะฐ
bug high-priority
![image](https://user-images.githubusercontent.com/1580811/42421595-b89db0a6-82e0-11e8-8c4a-343ee5fac970.png) ะ’ ะฟั€ะธะฒะตะดะตะฝะฝะพะผ ะฟั€ะธะผะตั€ะต `('tmp')` ััั‹ะปะฐะตั‚ัั ะฝะฐ `('user1', 'project1', 'operator1')`. ะ’ ัะปัƒั‡ะฐะต ัƒะดะฐะปะตะฝะธั ะฟะพัะปะตะดะฝะตะณะพ, ะธ ะฟะพัะปะตะดัƒัŽั‰ะตะน ะฟะพะฟั‹ั‚ะบะต ัะพะทะฐั‚ัŒ ะฝะพะฒั‹ะน ะพะฑัŒะตะบั‚ `('tmp', 'tmp1', 'tmp2')` ะพัˆะธะฑะบะฐ ะฝะตััƒั‰ะตัั‚ะฒัƒัŽั‰ะตะณะพ ั€ะพะดะธั‚ะตะปั ะฝะต ะฟั€ะพะธะทะพะนะดะตั‚ ะธ ะพะฑัŒะตะบั‚ ะดะพะฑะฐะฒะธั‚ัั, ะฒั‹ะทะฒะฐะฒ ั‚ะฐะบะธะผ ะพะฑั€ะฐะทะพะผ ะฟะพะฒั€ะตะถะดะตะฝะธะต ะธะฝะดะตะบัะฐ. ะ˜ ั…ะพั‚ั ะฟั€ะธ ะฟะตั€ะตะทะฐะฟัƒัะบะต ั†ะตะปะพัั‚ะฝะพัั‚ัŒ ะฑะฐะทั‹ ะฑัƒะดะตั‚ ะฒะพััั‚ะฐะฝะพะฒะปะตะฝะฐ (ะฟั€ะธ ะฟะตั€ะฒะพะผ ะฟะตั€ะตะทะฐะฟัƒัะบะต ะฑัƒะดะตั‚ ะพะฑะฝะฐั€ัƒะถะตะฝะพ ะฟะพะฒั€ะตะถะดะตะฝะธะต, ะฟั€ะธ ะฒั‚ะพั€ะพะผ ะฒั‹ะทั‹ะฒะฐัŽั‰ะธะน ะฟะพะฒั€ะตะถะดะตะฝะธะต ะพะฑัŒะตะบั‚ ะฑัƒะดะตั‚ ัƒะดะฐะปะตะฝ), ั‚ะฐะบะพะต ะฟะพะฒะตะดะตะฝะธะต ะฝะตะดะพะฟัƒัั‚ะธะผะพ. ะ”ะปั ัƒัั‚ั€ะฐะฝะตะฝะธั ะฟั€ะพะฑะปะตะผั‹ ะฝะตะพะฑั…ะพะดะธะผะพ ะฒ `_findInIndex()` ะฟั€ะพะธะทะฒะพะดะธั‚ัŒ ะฟั€ะพะฒะตั€ะบัƒ ะฝะฐ ััƒั‰ะตัั‚ะฒะพะฒะฐะฝะธะต ั†ะตะปะตะฒะพะณะพ ะพะฑัŒะตะบั‚ะฐ ั‡ะตั€ะตะท `resolveLink()`.
1.0
ะ˜ะฝะบะพะฝัะธัั‚ะตะฝั‚ะฝะพัั‚ัŒ ะฟั€ะธ ัƒะดะฐะปะตะฝะธะธ ะพะฑัŒะตะบั‚ะฐ, ะฝะฐ ะบะพั‚ะพั€ั‹ะน ะฑั‹ะปะฐ ััั‹ะปะบะฐ - ![image](https://user-images.githubusercontent.com/1580811/42421595-b89db0a6-82e0-11e8-8c4a-343ee5fac970.png) ะ’ ะฟั€ะธะฒะตะดะตะฝะฝะพะผ ะฟั€ะธะผะตั€ะต `('tmp')` ััั‹ะปะฐะตั‚ัั ะฝะฐ `('user1', 'project1', 'operator1')`. ะ’ ัะปัƒั‡ะฐะต ัƒะดะฐะปะตะฝะธั ะฟะพัะปะตะดะฝะตะณะพ, ะธ ะฟะพัะปะตะดัƒัŽั‰ะตะน ะฟะพะฟั‹ั‚ะบะต ัะพะทะฐั‚ัŒ ะฝะพะฒั‹ะน ะพะฑัŒะตะบั‚ `('tmp', 'tmp1', 'tmp2')` ะพัˆะธะฑะบะฐ ะฝะตััƒั‰ะตัั‚ะฒัƒัŽั‰ะตะณะพ ั€ะพะดะธั‚ะตะปั ะฝะต ะฟั€ะพะธะทะพะนะดะตั‚ ะธ ะพะฑัŒะตะบั‚ ะดะพะฑะฐะฒะธั‚ัั, ะฒั‹ะทะฒะฐะฒ ั‚ะฐะบะธะผ ะพะฑั€ะฐะทะพะผ ะฟะพะฒั€ะตะถะดะตะฝะธะต ะธะฝะดะตะบัะฐ. ะ˜ ั…ะพั‚ั ะฟั€ะธ ะฟะตั€ะตะทะฐะฟัƒัะบะต ั†ะตะปะพัั‚ะฝะพัั‚ัŒ ะฑะฐะทั‹ ะฑัƒะดะตั‚ ะฒะพััั‚ะฐะฝะพะฒะปะตะฝะฐ (ะฟั€ะธ ะฟะตั€ะฒะพะผ ะฟะตั€ะตะทะฐะฟัƒัะบะต ะฑัƒะดะตั‚ ะพะฑะฝะฐั€ัƒะถะตะฝะพ ะฟะพะฒั€ะตะถะดะตะฝะธะต, ะฟั€ะธ ะฒั‚ะพั€ะพะผ ะฒั‹ะทั‹ะฒะฐัŽั‰ะธะน ะฟะพะฒั€ะตะถะดะตะฝะธะต ะพะฑัŒะตะบั‚ ะฑัƒะดะตั‚ ัƒะดะฐะปะตะฝ), ั‚ะฐะบะพะต ะฟะพะฒะตะดะตะฝะธะต ะฝะตะดะพะฟัƒัั‚ะธะผะพ. ะ”ะปั ัƒัั‚ั€ะฐะฝะตะฝะธั ะฟั€ะพะฑะปะตะผั‹ ะฝะตะพะฑั…ะพะดะธะผะพ ะฒ `_findInIndex()` ะฟั€ะพะธะทะฒะพะดะธั‚ัŒ ะฟั€ะพะฒะตั€ะบัƒ ะฝะฐ ััƒั‰ะตัั‚ะฒะพะฒะฐะฝะธะต ั†ะตะปะตะฒะพะณะพ ะพะฑัŒะตะบั‚ะฐ ั‡ะตั€ะตะท `resolveLink()`.
priority
ะธะฝะบะพะฝัะธัั‚ะตะฝั‚ะฝะพัั‚ัŒ ะฟั€ะธ ัƒะดะฐะปะตะฝะธะธ ะพะฑัŒะตะบั‚ะฐ ะฝะฐ ะบะพั‚ะพั€ั‹ะน ะฑั‹ะปะฐ ััั‹ะปะบะฐ ะฒ ะฟั€ะธะฒะตะดะตะฝะฝะพะผ ะฟั€ะธะผะตั€ะต tmp ััั‹ะปะฐะตั‚ัั ะฝะฐ ะฒ ัะปัƒั‡ะฐะต ัƒะดะฐะปะตะฝะธั ะฟะพัะปะตะดะฝะตะณะพ ะธ ะฟะพัะปะตะดัƒัŽั‰ะตะน ะฟะพะฟั‹ั‚ะบะต ัะพะทะฐั‚ัŒ ะฝะพะฒั‹ะน ะพะฑัŒะตะบั‚ tmp ะพัˆะธะฑะบะฐ ะฝะตััƒั‰ะตัั‚ะฒัƒัŽั‰ะตะณะพ ั€ะพะดะธั‚ะตะปั ะฝะต ะฟั€ะพะธะทะพะนะดะตั‚ ะธ ะพะฑัŒะตะบั‚ ะดะพะฑะฐะฒะธั‚ัั ะฒั‹ะทะฒะฐะฒ ั‚ะฐะบะธะผ ะพะฑั€ะฐะทะพะผ ะฟะพะฒั€ะตะถะดะตะฝะธะต ะธะฝะดะตะบัะฐ ะธ ั…ะพั‚ั ะฟั€ะธ ะฟะตั€ะตะทะฐะฟัƒัะบะต ั†ะตะปะพัั‚ะฝะพัั‚ัŒ ะฑะฐะทั‹ ะฑัƒะดะตั‚ ะฒะพััั‚ะฐะฝะพะฒะปะตะฝะฐ ะฟั€ะธ ะฟะตั€ะฒะพะผ ะฟะตั€ะตะทะฐะฟัƒัะบะต ะฑัƒะดะตั‚ ะพะฑะฝะฐั€ัƒะถะตะฝะพ ะฟะพะฒั€ะตะถะดะตะฝะธะต ะฟั€ะธ ะฒั‚ะพั€ะพะผ ะฒั‹ะทั‹ะฒะฐัŽั‰ะธะน ะฟะพะฒั€ะตะถะดะตะฝะธะต ะพะฑัŒะตะบั‚ ะฑัƒะดะตั‚ ัƒะดะฐะปะตะฝ ั‚ะฐะบะพะต ะฟะพะฒะตะดะตะฝะธะต ะฝะตะดะพะฟัƒัั‚ะธะผะพ ะดะปั ัƒัั‚ั€ะฐะฝะตะฝะธั ะฟั€ะพะฑะปะตะผั‹ ะฝะตะพะฑั…ะพะดะธะผะพ ะฒ findinindex ะฟั€ะพะธะทะฒะพะดะธั‚ัŒ ะฟั€ะพะฒะตั€ะบัƒ ะฝะฐ ััƒั‰ะตัั‚ะฒะพะฒะฐะฝะธะต ั†ะตะปะตะฒะพะณะพ ะพะฑัŒะตะบั‚ะฐ ั‡ะตั€ะตะท resolvelink
1
719,252
24,753,014,180
IssuesEvent
2022-10-21 15:09:48
minetest/minetest
https://api.github.com/repos/minetest/minetest
closed
Add big red warning to bump_release.sh to check protocol version has been updated
Bug High priority
Alternatively enforce it somehow
1.0
Add big red warning to bump_release.sh to check protocol version has been updated - Alternatively enforce it somehow
priority
add big red warning to bump release sh to check protocol version has been updated alternatively enforce it somehow
1
20,774
2,628,489,215
IssuesEvent
2015-03-06 00:09:36
OSU-Net/cyder
https://api.github.com/repos/OSU-Net/cyder
closed
Users can add interfaces to systems that aren't in their Containers.
bug high priority vuln
If you look at a System that isn't in your list of Containers, the create static/dynamic interface buttons are still available. More interestingly, you can create additional interfaces on said system - and then edit them after creation. This may be limited based on allowed domains/IPs in the container though.
1.0
Users can add interfaces to systems that aren't in their Containers. - If you look at a System that isn't in your list of Containers, the create static/dynamic interface buttons are still available. More interestingly, you can create additional interfaces on said system - and then edit them after creation. This may be limited based on allowed domains/IPs in the container though.
priority
users can add interfaces to systems that aren t in their containers if you look at a system that isn t in your list of containers the create static dynamic interface buttons are still available more interestingly you can create additional interfaces on said system and then edit them after creation this may be limited based on allowed domains ips in the container though
1
95,062
3,933,585,208
IssuesEvent
2016-04-25 19:38:59
dkpro/dkpro-tc
https://api.github.com/repos/dkpro/dkpro-tc
closed
Add test cases that verify results of a pipeline
enhancement Priority-High
Essential all examples/demos run a pipeline but do not verify the result (i.e. accuracy value or something). At least one demo/example test case should access a result file i.e. `Constants.RESULTS_FILENAME`and `assert()` any numerical value. Otherwise it is impossible to notice new bugs if backend stuff is changed (i.e. #341)
1.0
Add test cases that verify results of a pipeline - Essential all examples/demos run a pipeline but do not verify the result (i.e. accuracy value or something). At least one demo/example test case should access a result file i.e. `Constants.RESULTS_FILENAME`and `assert()` any numerical value. Otherwise it is impossible to notice new bugs if backend stuff is changed (i.e. #341)
priority
add test cases that verify results of a pipeline essential all examples demos run a pipeline but do not verify the result i e accuracy value or something at least one demo example test case should access a result file i e constants results filename and assert any numerical value otherwise it is impossible to notice new bugs if backend stuff is changed i e
1
124,033
4,890,984,220
IssuesEvent
2016-11-18 15:28:40
tlatoza/SeeCodeRun
https://api.github.com/repos/tlatoza/SeeCodeRun
opened
Expressions that throw an exception should always be highlighted in red
high priority
Currently, when an expression throws an exception, it is listed as an error in the console. Hovering over the line in the console highlights the corresponding expression that is throwing the exception. This enables the user to find the error in the code. In many cases, the user may already have their cursor in the code window. It would be easier for users to simply see all expressions throwing an exception through a highlight and then be able to select an expression in the code editor to see the corresponding message in the console highlighted. https://seecode.run/#-KWs0kfSJRTz_AZKwiI3
1.0
Expressions that throw an exception should always be highlighted in red - Currently, when an expression throws an exception, it is listed as an error in the console. Hovering over the line in the console highlights the corresponding expression that is throwing the exception. This enables the user to find the error in the code. In many cases, the user may already have their cursor in the code window. It would be easier for users to simply see all expressions throwing an exception through a highlight and then be able to select an expression in the code editor to see the corresponding message in the console highlighted. https://seecode.run/#-KWs0kfSJRTz_AZKwiI3
priority
expressions that throw an exception should always be highlighted in red currently when an expression throws an exception it is listed as an error in the console hovering over the line in the console highlights the corresponding expression that is throwing the exception this enables the user to find the error in the code in many cases the user may already have their cursor in the code window it would be easier for users to simply see all expressions throwing an exception through a highlight and then be able to select an expression in the code editor to see the corresponding message in the console highlighted
1
664,282
22,264,080,341
IssuesEvent
2022-06-10 05:18:11
SCIInstitute/ShapeWorks
https://api.github.com/repos/SCIInstitute/ShapeWorks
closed
Distance between Surface Reconstruction and original/groomed surface
High Priority Feature Request QA User Study
This is a feature request to have distance between the reconstructed surface and the original image. This surface-surface distance should have options to be computed with respect to the original image (before grooming) or the groomed version. Also, the visualization of this distance by means of color maps.
1.0
Distance between Surface Reconstruction and original/groomed surface - This is a feature request to have distance between the reconstructed surface and the original image. This surface-surface distance should have options to be computed with respect to the original image (before grooming) or the groomed version. Also, the visualization of this distance by means of color maps.
priority
distance between surface reconstruction and original groomed surface this is a feature request to have distance between the reconstructed surface and the original image this surface surface distance should have options to be computed with respect to the original image before grooming or the groomed version also the visualization of this distance by means of color maps
1
651,766
21,509,747,422
IssuesEvent
2022-04-28 02:14:06
depub-team/depub.space
https://api.github.com/repos/depub-team/depub.space
closed
Detect IP and show different set of default channels accordingly
enhancement high priority
* Phase 1: TW & HK IPs will have a set of channels different from default * Write the locale to local storage * Allow to set the locale in a new setting menu * Provide 3 locale options: Cosmos, Taiwan, Hong Kong
1.0
Detect IP and show different set of default channels accordingly - * Phase 1: TW & HK IPs will have a set of channels different from default * Write the locale to local storage * Allow to set the locale in a new setting menu * Provide 3 locale options: Cosmos, Taiwan, Hong Kong
priority
detect ip and show different set of default channels accordingly phase tw hk ips will have a set of channels different from default write the locale to local storage allow to set the locale in a new setting menu provide locale options cosmos taiwan hong kong
1
357,482
10,607,184,813
IssuesEvent
2019-10-11 02:41:37
Azure/mmlspark
https://api.github.com/repos/Azure/mmlspark
closed
Incorrect NDCG reported on Validation Data
LightGBM bug high priority
**Describe the bug** I'm seeing incorrect values (unbelievably high near 1.0) of NDCG being reported in the logs while training LightGBMRanker. Computing NDCG with a [udf](http://lobotomys.blogspot.com/2016/08/normalised-discounted-cumulative-gain.html) based implementation I see ~.6 on the same dataset. I believe there are a couple bugs in the implementation that are contributing to this **1. The ranker is being trained on the validation data** The validation data is copied out of the main dataframe here [LightGBMBase.scala#L138](https://github.com/Azure/mmlspark/blob/master/src/main/scala/com/microsoft/ml/spark/lightgbm/LightGBMBase.scala#L138) but afaict `preprocessedDF` still has the validation data in it when we perform the training. In the logs below you'll see the training set > LightGBM worker generating dense dataset with 5000 rows and 80 columns the validation set > LightGBMRanker: LightGBM worker generating dense dataset with 500 rows and 80 columns I have 2 workers and 10k rows with 5% as validation data in my example, so this should be 4750 in the training set and 500 in the validation set per worker. **2. The validationData isn't sorted for ranking** When using LightGBMRanker we rely on preprocessData to sort by the group column [LightGBMRanker.scala#L75](https://github.com/Azure/mmlspark/blob/master/src/main/scala/com/microsoft/ml/spark/lightgbm/LightGBMRanker.scala#L75) but only the training dataframe ends up being sorted [LightGBMBase.scala#L140](https://github.com/Azure/mmlspark/blob/master/src/main/scala/com/microsoft/ml/spark/lightgbm/LightGBMBase.scala#L140) ``` 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker got nodes for network init: 10.222.225.231:12401,10.222.246.27:12400 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker listening on: 12400 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker generating dense dataset with 5000 rows and 80 columns 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker generateData from double array 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker LGBM_DatasetCreateFromMat 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker Validate generated dataset has the correct number of rows and cols 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker addFloatField 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker addGroupColumn 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker done generateDataset 19/10/07 21:54:44 INFO TorrentBroadcast: Started reading broadcast variable 173 19/10/07 21:54:45 INFO MemoryStore: Block broadcast_173_piece0 stored as bytes in memory (estimated size 93.4 KB, free 47.8 GB) 19/10/07 21:54:45 INFO TorrentBroadcast: Reading broadcast variable 173 took 3 ms 19/10/07 21:54:45 INFO MemoryStore: Block broadcast_173 stored as values in memory (estimated size 406.6 KB, free 47.8 GB) 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker generating dense dataset with 500 rows and 80 columns 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker generateData from double array 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker LGBM_DatasetCreateFromMat 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker Validate generated dataset has the correct number of rows and cols 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker addFloatField 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker addGroupColumn 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker done generateDataset 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM parameters: max_position=20 eval_at=1,2,3,4,5 is_pre_partition=True boosting_type=gbdt tree_learner=data_parallel num_iterations=100 learning_rate=0.1 num_leaves=31 max_bin=255 bagging_fraction=1.0 bagging_freq=0 bagging_seed=3 early_stopping_round=5 feature_fraction=1.0 max_depth=-1 min_sum_hessian_in_leaf=0.001 num_machines=2 objective=lambdarank verbosity=1 lambda_l1=0.0 lambda_l2=0.0 metric= 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM LGBM_BoosterCreate (nolock): com.microsoft.ml.lightgbm.SWIGTYPE_p_void@232f1182 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM LGBM_BoosterAddValidData 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM begin trainCore 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker calling LGBM_BoosterUpdateOneIter 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM running iteration: 0 with result: 0 and is finished: false 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@1=0.9198606271777003 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@2=0.9506376883275274 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@3=0.9610906500000013 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@4=0.9610906500000013 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@5=0.9624385691889719 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker calling LGBM_BoosterUpdateOneIter 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM running iteration: 1 with result: 0 and is finished: false 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@1=0.9372822299651568 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@2=0.965860929604282 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@3=0.9693452501617733 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@4=0.9723464805316228 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@5=0.9736943997205932 ``` **To Reproduce** ```python import mmlspark import pandas as pd import numpy as np import pyspark.sql.functions as fn from mmlspark.lightgbm import LightGBMRanker from pyspark.sql.types import * from pyspark.ml.feature import VectorAssembler from mmlspark.lightgbm import LightGBMRanker from scipy.sparse import rand ranker = LightGBMRanker() ranker.setGroupCol('search_id') ranker.setUseBarrierExecutionMode(True) ranker.setValidationIndicatorCol("is_validation") ranker.setEarlyStoppingRound(5) num_cols = 80 num_rows = 1000 * 10 def create_row(i): m = np.array(rand(1, num_cols + 2, density=.2, format='csr').todense())[0] m[0] = int(i // 10) # search id m[1] = int(i % 10 == 0) # label return m.tolist() feat_cols = ['feat_{}'.format(c) for c in range(num_cols)] df = spark.createDataFrame(pd.DataFrame([create_row(i) for i in range(num_rows)], columns=[ 'search_id', 'label'] + feat_cols )) df = df.withColumn('label', fn.col('label').cast(IntegerType())) df = df.withColumn('search_id', fn.col('search_id').cast(IntegerType())) # This should give us df = df.withColumn('is_validation', fn.expr('pmod(search_id, 100) >= 95')) assembler = VectorAssembler().setInputCols(feat_cols).setOutputCol('features') transformed = assembler.transform(df).withColumn('features', fn.expr('toDense(features)')) transformed = transformed.repartition(1, 'search_id') model = ranker.fit(transformed) ``` **Expected behavior** NDCG to be computed correctly Validation data shouldn't be included in the training set **Info (please complete the following information):** - MMLSpark Version: 035fcd91787cdc1b1b07cfb1bc7c13d5d9f5fa84 - Spark Version 2.4.3 - Spark Platform Databricks
1.0
Incorrect NDCG reported on Validation Data - **Describe the bug** I'm seeing incorrect values (unbelievably high near 1.0) of NDCG being reported in the logs while training LightGBMRanker. Computing NDCG with a [udf](http://lobotomys.blogspot.com/2016/08/normalised-discounted-cumulative-gain.html) based implementation I see ~.6 on the same dataset. I believe there are a couple bugs in the implementation that are contributing to this **1. The ranker is being trained on the validation data** The validation data is copied out of the main dataframe here [LightGBMBase.scala#L138](https://github.com/Azure/mmlspark/blob/master/src/main/scala/com/microsoft/ml/spark/lightgbm/LightGBMBase.scala#L138) but afaict `preprocessedDF` still has the validation data in it when we perform the training. In the logs below you'll see the training set > LightGBM worker generating dense dataset with 5000 rows and 80 columns the validation set > LightGBMRanker: LightGBM worker generating dense dataset with 500 rows and 80 columns I have 2 workers and 10k rows with 5% as validation data in my example, so this should be 4750 in the training set and 500 in the validation set per worker. **2. The validationData isn't sorted for ranking** When using LightGBMRanker we rely on preprocessData to sort by the group column [LightGBMRanker.scala#L75](https://github.com/Azure/mmlspark/blob/master/src/main/scala/com/microsoft/ml/spark/lightgbm/LightGBMRanker.scala#L75) but only the training dataframe ends up being sorted [LightGBMBase.scala#L140](https://github.com/Azure/mmlspark/blob/master/src/main/scala/com/microsoft/ml/spark/lightgbm/LightGBMBase.scala#L140) ``` 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker got nodes for network init: 10.222.225.231:12401,10.222.246.27:12400 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker listening on: 12400 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker generating dense dataset with 5000 rows and 80 columns 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker generateData from double array 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker LGBM_DatasetCreateFromMat 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker Validate generated dataset has the correct number of rows and cols 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker addFloatField 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker addGroupColumn 19/10/07 21:54:44 INFO LightGBMRanker: LightGBM worker done generateDataset 19/10/07 21:54:44 INFO TorrentBroadcast: Started reading broadcast variable 173 19/10/07 21:54:45 INFO MemoryStore: Block broadcast_173_piece0 stored as bytes in memory (estimated size 93.4 KB, free 47.8 GB) 19/10/07 21:54:45 INFO TorrentBroadcast: Reading broadcast variable 173 took 3 ms 19/10/07 21:54:45 INFO MemoryStore: Block broadcast_173 stored as values in memory (estimated size 406.6 KB, free 47.8 GB) 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker generating dense dataset with 500 rows and 80 columns 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker generateData from double array 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker LGBM_DatasetCreateFromMat 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker Validate generated dataset has the correct number of rows and cols 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker addFloatField 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker addGroupColumn 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker done generateDataset 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM parameters: max_position=20 eval_at=1,2,3,4,5 is_pre_partition=True boosting_type=gbdt tree_learner=data_parallel num_iterations=100 learning_rate=0.1 num_leaves=31 max_bin=255 bagging_fraction=1.0 bagging_freq=0 bagging_seed=3 early_stopping_round=5 feature_fraction=1.0 max_depth=-1 min_sum_hessian_in_leaf=0.001 num_machines=2 objective=lambdarank verbosity=1 lambda_l1=0.0 lambda_l2=0.0 metric= 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM LGBM_BoosterCreate (nolock): com.microsoft.ml.lightgbm.SWIGTYPE_p_void@232f1182 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM LGBM_BoosterAddValidData 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM begin trainCore 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker calling LGBM_BoosterUpdateOneIter 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM running iteration: 0 with result: 0 and is finished: false 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@1=0.9198606271777003 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@2=0.9506376883275274 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@3=0.9610906500000013 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@4=0.9610906500000013 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@5=0.9624385691889719 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM worker calling LGBM_BoosterUpdateOneIter 19/10/07 21:54:45 INFO LightGBMRanker: LightGBM running iteration: 1 with result: 0 and is finished: false 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@1=0.9372822299651568 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@2=0.965860929604282 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@3=0.9693452501617733 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@4=0.9723464805316228 19/10/07 21:54:45 INFO LightGBMRanker: Valid ndcg@5=0.9736943997205932 ``` **To Reproduce** ```python import mmlspark import pandas as pd import numpy as np import pyspark.sql.functions as fn from mmlspark.lightgbm import LightGBMRanker from pyspark.sql.types import * from pyspark.ml.feature import VectorAssembler from mmlspark.lightgbm import LightGBMRanker from scipy.sparse import rand ranker = LightGBMRanker() ranker.setGroupCol('search_id') ranker.setUseBarrierExecutionMode(True) ranker.setValidationIndicatorCol("is_validation") ranker.setEarlyStoppingRound(5) num_cols = 80 num_rows = 1000 * 10 def create_row(i): m = np.array(rand(1, num_cols + 2, density=.2, format='csr').todense())[0] m[0] = int(i // 10) # search id m[1] = int(i % 10 == 0) # label return m.tolist() feat_cols = ['feat_{}'.format(c) for c in range(num_cols)] df = spark.createDataFrame(pd.DataFrame([create_row(i) for i in range(num_rows)], columns=[ 'search_id', 'label'] + feat_cols )) df = df.withColumn('label', fn.col('label').cast(IntegerType())) df = df.withColumn('search_id', fn.col('search_id').cast(IntegerType())) # This should give us df = df.withColumn('is_validation', fn.expr('pmod(search_id, 100) >= 95')) assembler = VectorAssembler().setInputCols(feat_cols).setOutputCol('features') transformed = assembler.transform(df).withColumn('features', fn.expr('toDense(features)')) transformed = transformed.repartition(1, 'search_id') model = ranker.fit(transformed) ``` **Expected behavior** NDCG to be computed correctly Validation data shouldn't be included in the training set **Info (please complete the following information):** - MMLSpark Version: 035fcd91787cdc1b1b07cfb1bc7c13d5d9f5fa84 - Spark Version 2.4.3 - Spark Platform Databricks
priority
incorrect ndcg reported on validation data describe the bug i m seeing incorrect values unbelievably high near of ndcg being reported in the logs while training lightgbmranker computing ndcg with a based implementation i see on the same dataset i believe there are a couple bugs in the implementation that are contributing to this the ranker is being trained on the validation data the validation data is copied out of the main dataframe here but afaict preprocesseddf still has the validation data in it when we perform the training in the logs below you ll see the training set lightgbm worker generating dense dataset with rows and columns the validation set lightgbmranker lightgbm worker generating dense dataset with rows and columns i have workers and rows with as validation data in my example so this should be in the training set and in the validation set per worker the validationdata isn t sorted for ranking when using lightgbmranker we rely on preprocessdata to sort by the group column but only the training dataframe ends up being sorted info lightgbmranker lightgbm worker got nodes for network init info lightgbmranker lightgbm worker listening on info lightgbmranker lightgbm worker generating dense dataset with rows and columns info lightgbmranker lightgbm worker generatedata from double array info lightgbmranker lightgbm worker lgbm datasetcreatefrommat info lightgbmranker lightgbm worker validate generated dataset has the correct number of rows and cols info lightgbmranker lightgbm worker addfloatfield info lightgbmranker lightgbm worker addgroupcolumn info lightgbmranker lightgbm worker done generatedataset info torrentbroadcast started reading broadcast variable info memorystore block broadcast stored as bytes in memory estimated size kb free gb info torrentbroadcast reading broadcast variable took ms info memorystore block broadcast stored as values in memory estimated size kb free gb info lightgbmranker lightgbm worker generating dense dataset with rows and columns info lightgbmranker lightgbm worker generatedata from double array info lightgbmranker lightgbm worker lgbm datasetcreatefrommat info lightgbmranker lightgbm worker validate generated dataset has the correct number of rows and cols info lightgbmranker lightgbm worker addfloatfield info lightgbmranker lightgbm worker addgroupcolumn info lightgbmranker lightgbm worker done generatedataset info lightgbmranker lightgbm parameters max position eval at is pre partition true boosting type gbdt tree learner data parallel num iterations learning rate num leaves max bin bagging fraction bagging freq bagging seed early stopping round feature fraction max depth min sum hessian in leaf num machines objective lambdarank verbosity lambda lambda metric info lightgbmranker lightgbm lgbm boostercreate nolock com microsoft ml lightgbm swigtype p void info lightgbmranker lightgbm lgbm boosteraddvaliddata info lightgbmranker lightgbm begin traincore info lightgbmranker lightgbm worker calling lgbm boosterupdateoneiter info lightgbmranker lightgbm running iteration with result and is finished false info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker lightgbm worker calling lgbm boosterupdateoneiter info lightgbmranker lightgbm running iteration with result and is finished false info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker valid ndcg info lightgbmranker valid ndcg to reproduce python import mmlspark import pandas as pd import numpy as np import pyspark sql functions as fn from mmlspark lightgbm import lightgbmranker from pyspark sql types import from pyspark ml feature import vectorassembler from mmlspark lightgbm import lightgbmranker from scipy sparse import rand ranker lightgbmranker ranker setgroupcol search id ranker setusebarrierexecutionmode true ranker setvalidationindicatorcol is validation ranker setearlystoppinground num cols num rows def create row i m np array rand num cols density format csr todense m int i search id m int i label return m tolist feat cols df spark createdataframe pd dataframe columns feat cols df df withcolumn label fn col label cast integertype df df withcolumn search id fn col search id cast integertype this should give us df df withcolumn is validation fn expr pmod search id assembler vectorassembler setinputcols feat cols setoutputcol features transformed assembler transform df withcolumn features fn expr todense features transformed transformed repartition search id model ranker fit transformed expected behavior ndcg to be computed correctly validation data shouldn t be included in the training set info please complete the following information mmlspark version spark version spark platform databricks
1
447,759
12,892,882,205
IssuesEvent
2020-07-13 20:29:14
makerdao/community-portal
https://api.github.com/repos/makerdao/community-portal
closed
Bug - unrendered raw css displaying on 404 "Go back" button
HIGH PRIORITY bug
@MaximumCrash ## Bug description When I click "Go back" on a 404 page, the raw text for the previous page's css displays in the rendered page for a few brief frames, rendering as expected a moment later. ## Steps to replicate: 1. Merge PR #149 OR ```git checkout isaac/404``` to have access to 404 pages with a working "Go back" button. (if #149 is already merged, ignore this step) 2. ```yarn build``` 3. ```yarn serve``` 4. Navigate to any valid URI on the built site (```http://localhost:9000/```) 5. Navigate to an invalid URI in the same window/tab 6. Click the 404 page's "Go back" button. 7. You should see the css text flash on screen for a moment: <img width="1792" alt="Screen Shot 2020-07-12 at 7 00 17 PM" src="https://user-images.githubusercontent.com/21342523/87266839-2e58d900-c494-11ea-9fd6-326184f5278f.png">
1.0
Bug - unrendered raw css displaying on 404 "Go back" button - @MaximumCrash ## Bug description When I click "Go back" on a 404 page, the raw text for the previous page's css displays in the rendered page for a few brief frames, rendering as expected a moment later. ## Steps to replicate: 1. Merge PR #149 OR ```git checkout isaac/404``` to have access to 404 pages with a working "Go back" button. (if #149 is already merged, ignore this step) 2. ```yarn build``` 3. ```yarn serve``` 4. Navigate to any valid URI on the built site (```http://localhost:9000/```) 5. Navigate to an invalid URI in the same window/tab 6. Click the 404 page's "Go back" button. 7. You should see the css text flash on screen for a moment: <img width="1792" alt="Screen Shot 2020-07-12 at 7 00 17 PM" src="https://user-images.githubusercontent.com/21342523/87266839-2e58d900-c494-11ea-9fd6-326184f5278f.png">
priority
bug unrendered raw css displaying on go back button maximumcrash bug description when i click go back on a page the raw text for the previous page s css displays in the rendered page for a few brief frames rendering as expected a moment later steps to replicate merge pr or git checkout isaac to have access to pages with a working go back button if is already merged ignore this step yarn build yarn serve navigate to any valid uri on the built site navigate to an invalid uri in the same window tab click the page s go back button you should see the css text flash on screen for a moment img width alt screen shot at pm src
1
208,729
7,157,774,418
IssuesEvent
2018-01-26 21:14:05
vmware/vic
https://api.github.com/repos/vmware/vic
closed
Named volumes incorrectly detected as anonymous, getting deleted when container is removed
area/storage component/portlayer/storage kind/customer-found priority/high team/container triage/proposed-1.3.1
<!-- Issues in this repository are for VIC Engine only Issues related to other components should be submitted to their respective projects: VIC Appliance (OVA): https://github.com/vmware/vic-product/issues Container Management Portal (Admiral): https://github.com/vmware/admiral/issues Container Registry (Harbor): https://github.com/vmware/harbor/issues If you are unsure which component your issue relates to, submit it to https://github.com/vmware/vic-product/issues and we will triage it Thank you for contributing to VIC Engine! Please review this checklist before submitting your issue. [ ] Search for duplicates before logging new issues https://github.com/vmware/vic/issues?q=is%3Aopen+is%3Aissue [ ] Questions are best asked and answered on Slack https://vmwarecode.slack.com/messages/vic-engine (see https://github.com/vmware/vic/blob/master/CONTRIBUTING.md#community for details) --> #For bug reports, please include the information below: **VIC version:** 1.3.0 / Commit ID# 473375a **Deployment details:** What was the vic-machine create command used to deploy the VCH? vic-machine-windows --target=https://X.X.X.X --thumbprint=<thumbprint> --name=dev-vch --compute-resource=/AWN-DC-H5/host/AWN Production/Resources/Servers --ops-user=administrator@vs phere.local --image-store=ds://Servers-AFA --volume-store=ds://Servers-AFA/docker-volumes:default --dns-server=X.X.X.X --bridge-network=vic-bridge --bridge-network-range=172.32.0.0/12 --public-network=Manage ment --public-network-gateway=X.X.X.X --public-network-ip=X.X.X.X/24 --container-network=WWW:vic-www --container-network=Management:vic-mgmt --container-network-gateway=Management:X.X.X.X/24 --container- network-gateway=WWW:X.X.X.X/24 --container-network-ip-range=WWW:X.X.X.X/25 --container-network-ip-range=Management:X.X.X.X/29 --container-network-dns=Management:X.X.X.X --container-network-dns= WWW:X.X.X.X --memory-shares=500 --cpu-shares=500 **Steps to reproduce:** _**The image being used for testing is sebp/elk:611**_ These are the steps I have been performing when I experience the bug: 1. Follow procedure for backing up/restoring VMDK's for docker volumes (vic 1.3 documentation) 2. restart VCH to detect new volumes 3. Run container using the following: docker -H vch.awninc.local:2376 --tls run --name elk-node2 -p 9200 -p 9300 -it --rm -m 4G -e LOGSTASH_START=0 -e KIBANA_START=0 -e MAX_MAP_COUNT=262144 --memory-swap -1 --network vic-mgmt -v elk-node2-config:/etc/elasticsearch/ -v elk-node2-data:/var/lib/elasticsearch vic.awninc.local/utilities/elk:611 bash 4. Stop container by exiting bash, using 'exit' **Actual behavior:** Container is stopped and removed from VCH; 2nd named volume gets deleted (in this case, elk-node2-data) **Expected behavior:** Container stops and is removed, all named volumes remain intact. **Logs:** [logs (1).zip](https://github.com/vmware/vic/files/1648043/logs.1.zip) **Additional details as necessary:** Named volumes in question that are being attached to the containers were moved from another datastore.
1.0
Named volumes incorrectly detected as anonymous, getting deleted when container is removed - <!-- Issues in this repository are for VIC Engine only Issues related to other components should be submitted to their respective projects: VIC Appliance (OVA): https://github.com/vmware/vic-product/issues Container Management Portal (Admiral): https://github.com/vmware/admiral/issues Container Registry (Harbor): https://github.com/vmware/harbor/issues If you are unsure which component your issue relates to, submit it to https://github.com/vmware/vic-product/issues and we will triage it Thank you for contributing to VIC Engine! Please review this checklist before submitting your issue. [ ] Search for duplicates before logging new issues https://github.com/vmware/vic/issues?q=is%3Aopen+is%3Aissue [ ] Questions are best asked and answered on Slack https://vmwarecode.slack.com/messages/vic-engine (see https://github.com/vmware/vic/blob/master/CONTRIBUTING.md#community for details) --> #For bug reports, please include the information below: **VIC version:** 1.3.0 / Commit ID# 473375a **Deployment details:** What was the vic-machine create command used to deploy the VCH? vic-machine-windows --target=https://X.X.X.X --thumbprint=<thumbprint> --name=dev-vch --compute-resource=/AWN-DC-H5/host/AWN Production/Resources/Servers --ops-user=administrator@vs phere.local --image-store=ds://Servers-AFA --volume-store=ds://Servers-AFA/docker-volumes:default --dns-server=X.X.X.X --bridge-network=vic-bridge --bridge-network-range=172.32.0.0/12 --public-network=Manage ment --public-network-gateway=X.X.X.X --public-network-ip=X.X.X.X/24 --container-network=WWW:vic-www --container-network=Management:vic-mgmt --container-network-gateway=Management:X.X.X.X/24 --container- network-gateway=WWW:X.X.X.X/24 --container-network-ip-range=WWW:X.X.X.X/25 --container-network-ip-range=Management:X.X.X.X/29 --container-network-dns=Management:X.X.X.X --container-network-dns= WWW:X.X.X.X --memory-shares=500 --cpu-shares=500 **Steps to reproduce:** _**The image being used for testing is sebp/elk:611**_ These are the steps I have been performing when I experience the bug: 1. Follow procedure for backing up/restoring VMDK's for docker volumes (vic 1.3 documentation) 2. restart VCH to detect new volumes 3. Run container using the following: docker -H vch.awninc.local:2376 --tls run --name elk-node2 -p 9200 -p 9300 -it --rm -m 4G -e LOGSTASH_START=0 -e KIBANA_START=0 -e MAX_MAP_COUNT=262144 --memory-swap -1 --network vic-mgmt -v elk-node2-config:/etc/elasticsearch/ -v elk-node2-data:/var/lib/elasticsearch vic.awninc.local/utilities/elk:611 bash 4. Stop container by exiting bash, using 'exit' **Actual behavior:** Container is stopped and removed from VCH; 2nd named volume gets deleted (in this case, elk-node2-data) **Expected behavior:** Container stops and is removed, all named volumes remain intact. **Logs:** [logs (1).zip](https://github.com/vmware/vic/files/1648043/logs.1.zip) **Additional details as necessary:** Named volumes in question that are being attached to the containers were moved from another datastore.
priority
named volumes incorrectly detected as anonymous getting deleted when container is removed issues in this repository are for vic engine only issues related to other components should be submitted to their respective projects vic appliance ova container management portal admiral container registry harbor if you are unsure which component your issue relates to submit it to and we will triage it thank you for contributing to vic engine please review this checklist before submitting your issue search for duplicates before logging new issues questions are best asked and answered on slack see for details for bug reports please include the information below vic version commit id deployment details what was the vic machine create command used to deploy the vch vic machine windows target thumbprint name dev vch compute resource awn dc host awn production resources servers ops user administrator vs phere local image store ds servers afa volume store ds servers afa docker volumes default dns server x x x x bridge network vic bridge bridge network range public network manage ment public network gateway x x x x public network ip x x x x container network www vic www container network management vic mgmt container network gateway management x x x x container network gateway www x x x x container network ip range www x x x x container network ip range management x x x x container network dns management x x x x container network dns www x x x x memory shares cpu shares steps to reproduce the image being used for testing is sebp elk these are the steps i have been performing when i experience the bug follow procedure for backing up restoring vmdk s for docker volumes vic documentation restart vch to detect new volumes run container using the following docker h vch awninc local tls run name elk p p it rm m e logstash start e kibana start e max map count memory swap network vic mgmt v elk config etc elasticsearch v elk data var lib elasticsearch vic awninc local utilities elk bash stop container by exiting bash using exit actual behavior container is stopped and removed from vch named volume gets deleted in this case elk data expected behavior container stops and is removed all named volumes remain intact logs additional details as necessary named volumes in question that are being attached to the containers were moved from another datastore
1
386,416
11,438,600,699
IssuesEvent
2020-02-05 04:12:04
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Update Application owner from the publisher UI does not reflected in DB.
2.6.0 3.1.0 Priority/Highest Type/Bug
Since we have not committed the transaction properly, in ApiMgtDAO.updateApplicationOwner new owner is not getting changed in the db, Issue observed in a distributed setup. **Solution** When updating the application before the db update we have to set autoCommit to false and after the execution of the update query we need to commit the transaction.
1.0
Update Application owner from the publisher UI does not reflected in DB. - Since we have not committed the transaction properly, in ApiMgtDAO.updateApplicationOwner new owner is not getting changed in the db, Issue observed in a distributed setup. **Solution** When updating the application before the db update we have to set autoCommit to false and after the execution of the update query we need to commit the transaction.
priority
update application owner from the publisher ui does not reflected in db since we have not committed the transaction properly in apimgtdao updateapplicationowner new owner is not getting changed in the db issue observed in a distributed setup solution when updating the application before the db update we have to set autocommit to false and after the execution of the update query we need to commit the transaction
1
477,884
13,769,526,630
IssuesEvent
2020-10-07 18:45:30
NeXi-Client/NeXi-Client
https://api.github.com/repos/NeXi-Client/NeXi-Client
closed
Setup failing when trying to update
bug high priority
Pushing a update with a installer is not working Users have complained about that the Setup was just aborting when trying to update Issue is already wide spread but no fix found https://github.com/electron-userland/electron-builder/issues/4065 This needs to be fixed ASAP
1.0
Setup failing when trying to update - Pushing a update with a installer is not working Users have complained about that the Setup was just aborting when trying to update Issue is already wide spread but no fix found https://github.com/electron-userland/electron-builder/issues/4065 This needs to be fixed ASAP
priority
setup failing when trying to update pushing a update with a installer is not working users have complained about that the setup was just aborting when trying to update issue is already wide spread but no fix found this needs to be fixed asap
1
811,525
30,291,280,789
IssuesEvent
2023-07-09 10:12:19
flurrux/insta-loader
https://api.github.com/repos/flurrux/insta-loader
closed
download buttons on mainfeed are not placed correctly and fail to download images
bug high priority
here's what it looks like: <img width="376" alt="Screenshot 2023-07-08 100616" src="https://github.com/flurrux/insta-loader/assets/40806446/afe7b2cc-9339-42ea-8241-54ca1112df1c">
1.0
download buttons on mainfeed are not placed correctly and fail to download images - here's what it looks like: <img width="376" alt="Screenshot 2023-07-08 100616" src="https://github.com/flurrux/insta-loader/assets/40806446/afe7b2cc-9339-42ea-8241-54ca1112df1c">
priority
download buttons on mainfeed are not placed correctly and fail to download images here s what it looks like img width alt screenshot src
1
819,696
30,749,113,427
IssuesEvent
2023-07-28 17:29:59
Tedeapolis/development
https://api.github.com/repos/Tedeapolis/development
closed
[FEATURE]: Sterker maken militaire voertuigen
enhancement accepted high priority
### Contact Details Arjen#0001 ### Beschrijf zo duidelijk mogelijk de feature Was een oude github, maar is toen gesloten door de maker, dus nu even een nieuwe: Graag zouden wij zien dat een aantal voertuigen van defensie wat sterker gemaakt worden, aangezien het totaal niet realistisch is dat deze na een paar kogels van een pistool al uit de lucht vallen. - hunter - annihilator3 - chinook - lazer - hydra ### Wat lost deze feature op? Maakt het realistischer voor de defensievoertuigen aangezien deze normaal gesproken ook gebouwd zijn met de insteek dat ze wat tegenstand kunnen weerstaan. Zo heeft een Apache in het echt bijvoorbeeld kevlar platen aan de onderkant. Nu hoeft dit niet zo gemaakt geworden dat ze onverwoestbaar zijn, maar iets sterker zou fijn zijn.
1.0
[FEATURE]: Sterker maken militaire voertuigen - ### Contact Details Arjen#0001 ### Beschrijf zo duidelijk mogelijk de feature Was een oude github, maar is toen gesloten door de maker, dus nu even een nieuwe: Graag zouden wij zien dat een aantal voertuigen van defensie wat sterker gemaakt worden, aangezien het totaal niet realistisch is dat deze na een paar kogels van een pistool al uit de lucht vallen. - hunter - annihilator3 - chinook - lazer - hydra ### Wat lost deze feature op? Maakt het realistischer voor de defensievoertuigen aangezien deze normaal gesproken ook gebouwd zijn met de insteek dat ze wat tegenstand kunnen weerstaan. Zo heeft een Apache in het echt bijvoorbeeld kevlar platen aan de onderkant. Nu hoeft dit niet zo gemaakt geworden dat ze onverwoestbaar zijn, maar iets sterker zou fijn zijn.
priority
sterker maken militaire voertuigen contact details arjen beschrijf zo duidelijk mogelijk de feature was een oude github maar is toen gesloten door de maker dus nu even een nieuwe graag zouden wij zien dat een aantal voertuigen van defensie wat sterker gemaakt worden aangezien het totaal niet realistisch is dat deze na een paar kogels van een pistool al uit de lucht vallen hunter chinook lazer hydra wat lost deze feature op maakt het realistischer voor de defensievoertuigen aangezien deze normaal gesproken ook gebouwd zijn met de insteek dat ze wat tegenstand kunnen weerstaan zo heeft een apache in het echt bijvoorbeeld kevlar platen aan de onderkant nu hoeft dit niet zo gemaakt geworden dat ze onverwoestbaar zijn maar iets sterker zou fijn zijn
1
636,932
20,613,738,530
IssuesEvent
2022-03-07 11:08:41
owid/covid-19-data
https://api.github.com/repos/owid/covid-19-data
closed
bug(vax,hungary): current script crashes due to regex issue
bug dom:vaccinations priority:high
### Description The current script crashes due to a regex issue. The data was last updated on Mar 3. ``` 2022-03-07 10:24:26,894 ERROR VAX - cowidev.vax.incremental.hungary: โŒ 'NoneType' object has no attribute 'group' Traceback (most recent call last): File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/cmd/commons/get.py", line 39, in run module.main() File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 141, in main Hungary().export() File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 133, in export df = self.read(last_update) File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 29, in read data_, proceed = self.parse_data(soup, last_update) File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 44, in parse_data **self.parse_data_news_page(soup), File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 75, in parse_data_news_page people_vaccinated = clean_count(match.group(1)) AttributeError: 'NoneType' object has no attribute 'group' ```
1.0
bug(vax,hungary): current script crashes due to regex issue - ### Description The current script crashes due to a regex issue. The data was last updated on Mar 3. ``` 2022-03-07 10:24:26,894 ERROR VAX - cowidev.vax.incremental.hungary: โŒ 'NoneType' object has no attribute 'group' Traceback (most recent call last): File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/cmd/commons/get.py", line 39, in run module.main() File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 141, in main Hungary().export() File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 133, in export df = self.read(last_update) File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 29, in read data_, proceed = self.parse_data(soup, last_update) File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 44, in parse_data **self.parse_data_news_page(soup), File "/Users/edouard/git/covid-19-data/scripts/src/cowidev/vax/incremental/hungary.py", line 75, in parse_data_news_page people_vaccinated = clean_count(match.group(1)) AttributeError: 'NoneType' object has no attribute 'group' ```
priority
bug vax hungary current script crashes due to regex issue description the current script crashes due to a regex issue the data was last updated on mar error vax cowidev vax incremental hungary โŒ nonetype object has no attribute group traceback most recent call last file users edouard git covid data scripts src cowidev cmd commons get py line in run module main file users edouard git covid data scripts src cowidev vax incremental hungary py line in main hungary export file users edouard git covid data scripts src cowidev vax incremental hungary py line in export df self read last update file users edouard git covid data scripts src cowidev vax incremental hungary py line in read data proceed self parse data soup last update file users edouard git covid data scripts src cowidev vax incremental hungary py line in parse data self parse data news page soup file users edouard git covid data scripts src cowidev vax incremental hungary py line in parse data news page people vaccinated clean count match group attributeerror nonetype object has no attribute group
1
206,484
7,112,761,830
IssuesEvent
2018-01-17 18:05:18
vmware/vic
https://api.github.com/repos/vmware/vic
opened
1-06 failure: Creating an image failed: [POST /storage/{store_name}][500] WriteImage default &{Code:500 Message:503 Service Unavailable
component/imagec priority/high team/container
https://ci.vcna.io/vmware/vic/15671 the mysql image pull failed with the following error: ``` Jan 16 2018 19:25:34.017Z DEBUG [BEGIN] [vic/lib/imagec.WriteImage:82] a2b2035891c3e56a92af6994f39359b155bc181a8c02e25bfcfafa9eccc5a6f7 Jan 16 2018 19:25:37.426Z INFO diffID for layer c2c5c9edebfcb471b0ef4d210988a9ef7f3aaaf2c4f2ff058bdc72619fbcf94a: sha256:7800b09ed0f6d69934ddade64be6f129e71bcd28de20ab2ae7db58ce19118cab Jan 16 2018 19:25:37.430Z DEBUG [ END ] [vic/lib/imagec.FetchImageBlob:181] [13.639816422s] library/mysql/sha256:3a8e3ebdeaf554387932484bf93f67f26d4bbb20d76eeb231b9b694efbf6a2ee Jan 16 2018 19:26:49.900Z DEBUG Creating an image failed: [POST /storage/{store_name}][500] WriteImage default &{Code:500 Message:503 Service Unavailable (Failed to connect to endpoint: [N7Vmacore4Http16LocalServiceSpecE:0xffa090d0] _serverNamespace = /folder _isRedirect = false _port = 8309)} ``` [Test-Cases.Group1-Docker-Commands.1-06-Docker-Run-VCH-15671-3824-container-logs.zip](https://github.com/vmware/vic/files/1640103/Test-Cases.Group1-Docker-Commands.1-06-Docker-Run-VCH-15671-3824-container-logs.zip) Also, the final DEBUG message looks like it needs to actually be tagged with ERROR
1.0
1-06 failure: Creating an image failed: [POST /storage/{store_name}][500] WriteImage default &{Code:500 Message:503 Service Unavailable - https://ci.vcna.io/vmware/vic/15671 the mysql image pull failed with the following error: ``` Jan 16 2018 19:25:34.017Z DEBUG [BEGIN] [vic/lib/imagec.WriteImage:82] a2b2035891c3e56a92af6994f39359b155bc181a8c02e25bfcfafa9eccc5a6f7 Jan 16 2018 19:25:37.426Z INFO diffID for layer c2c5c9edebfcb471b0ef4d210988a9ef7f3aaaf2c4f2ff058bdc72619fbcf94a: sha256:7800b09ed0f6d69934ddade64be6f129e71bcd28de20ab2ae7db58ce19118cab Jan 16 2018 19:25:37.430Z DEBUG [ END ] [vic/lib/imagec.FetchImageBlob:181] [13.639816422s] library/mysql/sha256:3a8e3ebdeaf554387932484bf93f67f26d4bbb20d76eeb231b9b694efbf6a2ee Jan 16 2018 19:26:49.900Z DEBUG Creating an image failed: [POST /storage/{store_name}][500] WriteImage default &{Code:500 Message:503 Service Unavailable (Failed to connect to endpoint: [N7Vmacore4Http16LocalServiceSpecE:0xffa090d0] _serverNamespace = /folder _isRedirect = false _port = 8309)} ``` [Test-Cases.Group1-Docker-Commands.1-06-Docker-Run-VCH-15671-3824-container-logs.zip](https://github.com/vmware/vic/files/1640103/Test-Cases.Group1-Docker-Commands.1-06-Docker-Run-VCH-15671-3824-container-logs.zip) Also, the final DEBUG message looks like it needs to actually be tagged with ERROR
priority
failure creating an image failed writeimage default code message service unavailable the mysql image pull failed with the following error jan debug jan info diffid for layer jan debug library mysql jan debug creating an image failed writeimage default code message service unavailable failed to connect to endpoint servernamespace folder isredirect false port also the final debug message looks like it needs to actually be tagged with error
1
477,331
13,760,128,132
IssuesEvent
2020-10-07 05:03:29
AY2021S1-CS2103T-T09-4/tp
https://api.github.com/repos/AY2021S1-CS2103T-T09-4/tp
opened
Data Tracking
Priority.High type.DG type.Function
Track relevant user data such as: 1. Number of correctly answered flashcards 2. Previous scores 3. Average time taken in total 4. Time of quiz As well as the following subtasks: 1. Tests 2. Update DG
1.0
Data Tracking - Track relevant user data such as: 1. Number of correctly answered flashcards 2. Previous scores 3. Average time taken in total 4. Time of quiz As well as the following subtasks: 1. Tests 2. Update DG
priority
data tracking track relevant user data such as number of correctly answered flashcards previous scores average time taken in total time of quiz as well as the following subtasks tests update dg
1
246,959
7,895,876,040
IssuesEvent
2018-06-29 06:15:31
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Visualization in double precision.
Expected Use: 2 - Rare Feature Impact: 4 - High OS: All Priority: Normal Support Group: DOE/ASC
cq-id: VisIt00006556 cq-submitter: Hank Childs cq-submit-date: 09/07/05 Mike Nemanic is unable to observe phenomena in his code because we visualize in single precision. I'm going to dummy up a Silo reader that reads in double precision and then does a conversion so it is observable in single precision. But this is not a long term solution. He says that he and Paul Nowak frequently run into this problem. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Cyrus Harrison Original creation: 06/21/2010 07:16 pm Original update: 10/18/2012 05:34 pm Ticket number: 114
1.0
Visualization in double precision. - cq-id: VisIt00006556 cq-submitter: Hank Childs cq-submit-date: 09/07/05 Mike Nemanic is unable to observe phenomena in his code because we visualize in single precision. I'm going to dummy up a Silo reader that reads in double precision and then does a conversion so it is observable in single precision. But this is not a long term solution. He says that he and Paul Nowak frequently run into this problem. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Cyrus Harrison Original creation: 06/21/2010 07:16 pm Original update: 10/18/2012 05:34 pm Ticket number: 114
priority
visualization in double precision cq id cq submitter hank childs cq submit date mike nemanic is unable to observe phenomena in his code because we visualize in single precision i m going to dummy up a silo reader that reads in double precision and then does a conversion so it is observable in single precision but this is not a long term solution he says that he and paul nowak frequently run into this problem redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author cyrus harrison original creation pm original update pm ticket number
1
234,876
7,727,252,926
IssuesEvent
2018-05-25 01:28:06
facebook/prepack
https://api.github.com/repos/facebook/prepack
closed
Wrong serialization order when residual function is called during the optimized code path
bug priority: high react compiler serializer
This serializes the `Object.assign` after the `a.callFunc` call which is the wrong order: ```js function foo(a, b, c) { if (!a) { return null; } var b = Object.assign({}, c); return a.callFunc(function() { return b.foo; }); } inspect = function() { return JSON.stringify( foo({ a: { callFunc(x) { return x() }, }, b: { foo() { return 'works!'; }, }, c: {}, }) ); } this.__optimize && __optimize(foo) ``` Interestingly, if I remove ```js if (!a) { return null; } ``` The output it serializes is correct.
1.0
Wrong serialization order when residual function is called during the optimized code path - This serializes the `Object.assign` after the `a.callFunc` call which is the wrong order: ```js function foo(a, b, c) { if (!a) { return null; } var b = Object.assign({}, c); return a.callFunc(function() { return b.foo; }); } inspect = function() { return JSON.stringify( foo({ a: { callFunc(x) { return x() }, }, b: { foo() { return 'works!'; }, }, c: {}, }) ); } this.__optimize && __optimize(foo) ``` Interestingly, if I remove ```js if (!a) { return null; } ``` The output it serializes is correct.
priority
wrong serialization order when residual function is called during the optimized code path this serializes the object assign after the a callfunc call which is the wrong order js function foo a b c if a return null var b object assign c return a callfunc function return b foo inspect function return json stringify foo a callfunc x return x b foo return works c this optimize optimize foo interestingly if i remove js if a return null the output it serializes is correct
1
353,925
10,560,737,528
IssuesEvent
2019-10-04 14:29:07
infor-design/enterprise
https://api.github.com/repos/infor-design/enterprise
closed
Datagrid: Actions menu can only be opened once
[2] focus: datagrid priority: high type: regression bug :leftwards_arrow_with_hook:
**Describe the bug** If opening an actions menu in a datagrid (Actions formatter on column), and then refresh the dataset, the menu can no longer be opened for any row item. This is a high priority issue for us. ![action-menu](https://user-images.githubusercontent.com/1101770/66034514-a8256480-e509-11e9-9b7f-7c636c0c2881.gif) It was originally reported & fixed in March 2018: https://jira.infor.com/browse/SOHO-7634 **To Reproduce** The SOHO ticket above has a link to a test page you created, but I'm unable to find out where to access it now (or if it exists anymore)? http://usalvlhlpool1.infor.com/4.5.0-rc/components/datagrid/test-actions-reload.html **Expected behavior** After refreshing dataset (either by changing dataset completely or if using updateRow() etc.), the Actions menu can still be opened. **Version** - ids-enterprise: 4.21.0 **Additonal context** See https://jira.infor.com/browse/SOHO-7634 Working on 4.20.1
1.0
Datagrid: Actions menu can only be opened once - **Describe the bug** If opening an actions menu in a datagrid (Actions formatter on column), and then refresh the dataset, the menu can no longer be opened for any row item. This is a high priority issue for us. ![action-menu](https://user-images.githubusercontent.com/1101770/66034514-a8256480-e509-11e9-9b7f-7c636c0c2881.gif) It was originally reported & fixed in March 2018: https://jira.infor.com/browse/SOHO-7634 **To Reproduce** The SOHO ticket above has a link to a test page you created, but I'm unable to find out where to access it now (or if it exists anymore)? http://usalvlhlpool1.infor.com/4.5.0-rc/components/datagrid/test-actions-reload.html **Expected behavior** After refreshing dataset (either by changing dataset completely or if using updateRow() etc.), the Actions menu can still be opened. **Version** - ids-enterprise: 4.21.0 **Additonal context** See https://jira.infor.com/browse/SOHO-7634 Working on 4.20.1
priority
datagrid actions menu can only be opened once describe the bug if opening an actions menu in a datagrid actions formatter on column and then refresh the dataset the menu can no longer be opened for any row item this is a high priority issue for us it was originally reported fixed in march to reproduce the soho ticket above has a link to a test page you created but i m unable to find out where to access it now or if it exists anymore expected behavior after refreshing dataset either by changing dataset completely or if using updaterow etc the actions menu can still be opened version ids enterprise additonal context see working on
1
412,726
12,054,979,058
IssuesEvent
2020-04-15 12:11:48
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
closed
clicking on position quantity doesn't pop order form
Needed for V2 launch Priority: High type/bug
clicking on position quantity doesn't pop order form, user has a short position on YesNo market and clicking on position quantity doesn't populate order form quantity. ![Screen Shot 2020-04-13 at 6 07 43 PM](https://user-images.githubusercontent.com/3970376/79169318-bfdf7e80-7db1-11ea-9833-88a1d2bf7b88.png)
1.0
clicking on position quantity doesn't pop order form - clicking on position quantity doesn't pop order form, user has a short position on YesNo market and clicking on position quantity doesn't populate order form quantity. ![Screen Shot 2020-04-13 at 6 07 43 PM](https://user-images.githubusercontent.com/3970376/79169318-bfdf7e80-7db1-11ea-9833-88a1d2bf7b88.png)
priority
clicking on position quantity doesn t pop order form clicking on position quantity doesn t pop order form user has a short position on yesno market and clicking on position quantity doesn t populate order form quantity
1