Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
410,102
11,983,209,292
IssuesEvent
2020-04-07 14:07:21
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
closed
`ShowMoreCard` button used in developer reply has the wrong background color
component: add-on ratings contrib: welcome priority: p3
### Describe the problem and steps to reproduce it: 1. Go to https://addons-dev.allizom.org/en-US/firefox/addon/search_by_image/reviews/?src=recommended_fallback 2. Observe the "developer response" in the middle of the page ### What happened? The `ShowMoreCard` displays a "read more" button with a white background: <img width="1280" alt="Screen Shot 2020-03-31 at 14 38 17" src="https://user-images.githubusercontent.com/217628/78027415-9ceaae80-735d-11ea-8527-d8c26ea60816.png"> ### What did you expect to happen? The button background color should be blue-ish.
1.0
`ShowMoreCard` button used in developer reply has the wrong background color - ### Describe the problem and steps to reproduce it: 1. Go to https://addons-dev.allizom.org/en-US/firefox/addon/search_by_image/reviews/?src=recommended_fallback 2. Observe the "developer response" in the middle of the page ### What happened? The `ShowMoreCard` displays a "read more" button with a white background: <img width="1280" alt="Screen Shot 2020-03-31 at 14 38 17" src="https://user-images.githubusercontent.com/217628/78027415-9ceaae80-735d-11ea-8527-d8c26ea60816.png"> ### What did you expect to happen? The button background color should be blue-ish.
priority
showmorecard button used in developer reply has the wrong background color describe the problem and steps to reproduce it go to observe the developer response in the middle of the page what happened the showmorecard displays a read more button with a white background img width alt screen shot at src what did you expect to happen the button background color should be blue ish
1
133,234
18,285,818,279
IssuesEvent
2021-10-05 10:08:20
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Security Solution][Timeline]On Clicking + Add field user focus moved from the Field name to the value
bug triage_needed Team: SecuritySolution
**Describe the bug** Timeline: On Clicking + Add field user focus moved from the Field name to the value **Build Details** `Version: 7.15.0` **Steps** - Login to kibana - Create timeline - Click on + Add field in the timeline drag area - click on drop down and don't select any value from this - Observed that user focus is moved to the value field. This makes it a pain to manually add fields because you have to stop and wait for it to do the focus change before you can start typing, otherwise it will move when you are in the middle of typing the field name **Actual** Timeline: On Clicking + Add field user focus moved from the Field name to the value **Snap Shoot** ![image](https://user-images.githubusercontent.com/59917825/136000596-512fd9e0-bfe8-43ab-be6e-6d8bc1de504e.png) **Screen-Cast** https://elastic.slack.com/archives/C012TJZ72J1/p1633426485272700 **logs** - N/A
True
[Security Solution][Timeline]On Clicking + Add field user focus moved from the Field name to the value - **Describe the bug** Timeline: On Clicking + Add field user focus moved from the Field name to the value **Build Details** `Version: 7.15.0` **Steps** - Login to kibana - Create timeline - Click on + Add field in the timeline drag area - click on drop down and don't select any value from this - Observed that user focus is moved to the value field. This makes it a pain to manually add fields because you have to stop and wait for it to do the focus change before you can start typing, otherwise it will move when you are in the middle of typing the field name **Actual** Timeline: On Clicking + Add field user focus moved from the Field name to the value **Snap Shoot** ![image](https://user-images.githubusercontent.com/59917825/136000596-512fd9e0-bfe8-43ab-be6e-6d8bc1de504e.png) **Screen-Cast** https://elastic.slack.com/archives/C012TJZ72J1/p1633426485272700 **logs** - N/A
non_priority
on clicking add field user focus moved from the field name to the value describe the bug timeline on clicking add field user focus moved from the field name to the value build details version steps login to kibana create timeline click on add field in the timeline drag area click on drop down and don t select any value from this observed that user focus is moved to the value field this makes it a pain to manually add fields because you have to stop and wait for it to do the focus change before you can start typing otherwise it will move when you are in the middle of typing the field name actual timeline on clicking add field user focus moved from the field name to the value snap shoot screen cast logs n a
0
562,345
16,657,647,556
IssuesEvent
2021-06-05 20:31:04
dodona-edu/dodona
https://api.github.com/repos/dodona-edu/dodona
closed
HTML5 video tag does not support skipping forward/backward
bug low priority student
Video fragments embedded using the HTML5 tag `video` cannot be skipped forward nor backward. Example: [exercise description](https://dodona.ugent.be/nl/courses/172/series/3251/activities/357352739/) OS: Mac OS X Browser: Google Chrome Observed by: @toonijn @pverscha
1.0
HTML5 video tag does not support skipping forward/backward - Video fragments embedded using the HTML5 tag `video` cannot be skipped forward nor backward. Example: [exercise description](https://dodona.ugent.be/nl/courses/172/series/3251/activities/357352739/) OS: Mac OS X Browser: Google Chrome Observed by: @toonijn @pverscha
priority
video tag does not support skipping forward backward video fragments embedded using the tag video cannot be skipped forward nor backward example os mac os x browser google chrome observed by toonijn pverscha
1
52,057
3,020,503,464
IssuesEvent
2015-07-31 08:25:44
52North/SOS
https://api.github.com/repos/52North/SOS
closed
Client problem with feature capabilities in DescribeSensor response with SensorML 1.0.1 encoded procedure
bug enhancement medium priority
The SweText of the feature capabilities of a DescribeSensor response procedure encoded in SensorML 1.0.1 does not contain the definition which is used by the Sensor Web Client/REST-API proxy to identify featureOfInterest. Problem: The merging process replaces the field name.
1.0
Client problem with feature capabilities in DescribeSensor response with SensorML 1.0.1 encoded procedure - The SweText of the feature capabilities of a DescribeSensor response procedure encoded in SensorML 1.0.1 does not contain the definition which is used by the Sensor Web Client/REST-API proxy to identify featureOfInterest. Problem: The merging process replaces the field name.
priority
client problem with feature capabilities in describesensor response with sensorml encoded procedure the swetext of the feature capabilities of a describesensor response procedure encoded in sensorml does not contain the definition which is used by the sensor web client rest api proxy to identify featureofinterest problem the merging process replaces the field name
1
773,206
27,149,702,309
IssuesEvent
2023-02-16 23:31:46
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
opened
Migrate to Vue 3
Type: Feature Request Priority: 3 1-off tasks Type: Epic Lead: @jimchamp
### Describe the problem that you'd like solved <!-- A clear and concise description of what you want to happen. --> Open Library maintains a small number of web components that were created with Vue 2. Vue 2 will reach end of life on 31 December 2023, so we'll have to migrate to Vue 3 in the near future. This also give us the opportunity to try using Vite for our component builds. We currently use Vue CLI, and this is one of the slower steps of our build process. More importantly, Vue CLI is now in maintenance mode, and eventually will be no longer supported. This epic can be considered complete when the following tasks are finished: - [ ] Upgrade to Vue 3 - [ ] Test new build process with Vite - [ ] Migrate all of the following components: - [x] AuthorIdentifiers - [x] HelloWorld - [ ] LibraryExplorer - [ ] MergeUI - [x] ObservationForm - [ ] Update our Vue [documentation](https://github.com/internetarchive/openlibrary/wiki/Using-Vue) - [ ] Add link to this page to wiki sidebar ### Proposal & Constraints <!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? --> <!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? --> ### Additional context <!-- Add any other context or screenshots about the feature request here. --> ### Stakeholders <!-- @ tag stakeholders of this bug -->
1.0
Migrate to Vue 3 - ### Describe the problem that you'd like solved <!-- A clear and concise description of what you want to happen. --> Open Library maintains a small number of web components that were created with Vue 2. Vue 2 will reach end of life on 31 December 2023, so we'll have to migrate to Vue 3 in the near future. This also give us the opportunity to try using Vite for our component builds. We currently use Vue CLI, and this is one of the slower steps of our build process. More importantly, Vue CLI is now in maintenance mode, and eventually will be no longer supported. This epic can be considered complete when the following tasks are finished: - [ ] Upgrade to Vue 3 - [ ] Test new build process with Vite - [ ] Migrate all of the following components: - [x] AuthorIdentifiers - [x] HelloWorld - [ ] LibraryExplorer - [ ] MergeUI - [x] ObservationForm - [ ] Update our Vue [documentation](https://github.com/internetarchive/openlibrary/wiki/Using-Vue) - [ ] Add link to this page to wiki sidebar ### Proposal & Constraints <!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? --> <!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? --> ### Additional context <!-- Add any other context or screenshots about the feature request here. --> ### Stakeholders <!-- @ tag stakeholders of this bug -->
priority
migrate to vue describe the problem that you d like solved open library maintains a small number of web components that were created with vue vue will reach end of life on december so we ll have to migrate to vue in the near future this also give us the opportunity to try using vite for our component builds we currently use vue cli and this is one of the slower steps of our build process more importantly vue cli is now in maintenance mode and eventually will be no longer supported this epic can be considered complete when the following tasks are finished upgrade to vue test new build process with vite migrate all of the following components authoridentifiers helloworld libraryexplorer mergeui observationform update our vue add link to this page to wiki sidebar proposal constraints additional context stakeholders
1
112,988
9,607,837,253
IssuesEvent
2019-05-11 22:53:01
dimitri/pgloader
https://api.github.com/repos/dimitri/pgloader
closed
AWS RDS to AWS VM
How To... Question Needs more testing / information
Thanks for contributing to [pgloader](https://pgloader.io) by reporting an issue! Reporting an issue is the only way we can solve problems, fix bugs, and improve both the software and its user experience in general. The best bug reports follow those 3 simple steps: 1. show what you did, 2. show the result you got, 3. explain how the result is not what you expected. In the case of pgloader, here's the information I will need to read in your bug report. Having all of this is a big help, and often means the bug you reported can be fixed very efficiently as soon as I get to it. Please provide the following information: <!-- delete text above this line --> - [ ] pgloader --version ``` <fill pgloader version here> ``` - [ ] did you test a fresh compile from the source tree? Compiling pgloader from sources is documented in the [README](https://github.com/dimitri/pgloader#build-from-sources), it's easy to do, and if patches are to be made to fix your bug, you're going to have to build from sources to get the fix anyway… - [ ] did you search for other similar issues? - [ ] how can I reproduce the bug? Incude a self-contained pgloader command file. If you're loading from a database, consider attaching a database dump to your issue. For MySQL, use `mysqldump`. For SQLite, just send over your source file, that's easy. Maybe be the one with your production data, of course, the one with just the sample of data that allows me to reproduce your bug. When using a proprietary database system as a source, consider creating a sample database on some Cloud service or somewhere you can then give me access to, and see my email address on my GitHub profile to send me the credentials. Still open a public issue for tracking and as documentation for other users. ``` -- -- EDIT THIS FILE TO MATCH YOUR BUG REPORT -- LOAD CSV FROM INLINE with encoding 'ascii' INTO postgresql:///pgloader TARGET TABLE jordane WITH truncate, fields terminated by '|', fields not enclosed, fields escaped by backslash-quote SET work_mem to '128MB', standard_conforming_strings to 'on' BEFORE LOAD DO $$ drop table if exists jordane; $$, $$ CREATE TABLE jordane ( "NOM" character(20), "PRENOM" character(20) ) $$; BORDET|Jordane BORDET|Audrey LASTNAME|"opening quote BONNIER|testprenombe~aucouptroplong JOURDAIN|héhé¶ ``` - [ ] pgloader output you obtain ``` PASTE HERE THE OUTPUT OF THE PGLOADER COMMAND ``` - [ ] data that is being loaded, if relevant ``` PASTE HERE THE DATA THAT HAS BEEN LOADED ``` - [ ] How the data is different from what you expected, if relevant
1.0
AWS RDS to AWS VM - Thanks for contributing to [pgloader](https://pgloader.io) by reporting an issue! Reporting an issue is the only way we can solve problems, fix bugs, and improve both the software and its user experience in general. The best bug reports follow those 3 simple steps: 1. show what you did, 2. show the result you got, 3. explain how the result is not what you expected. In the case of pgloader, here's the information I will need to read in your bug report. Having all of this is a big help, and often means the bug you reported can be fixed very efficiently as soon as I get to it. Please provide the following information: <!-- delete text above this line --> - [ ] pgloader --version ``` <fill pgloader version here> ``` - [ ] did you test a fresh compile from the source tree? Compiling pgloader from sources is documented in the [README](https://github.com/dimitri/pgloader#build-from-sources), it's easy to do, and if patches are to be made to fix your bug, you're going to have to build from sources to get the fix anyway… - [ ] did you search for other similar issues? - [ ] how can I reproduce the bug? Incude a self-contained pgloader command file. If you're loading from a database, consider attaching a database dump to your issue. For MySQL, use `mysqldump`. For SQLite, just send over your source file, that's easy. Maybe be the one with your production data, of course, the one with just the sample of data that allows me to reproduce your bug. When using a proprietary database system as a source, consider creating a sample database on some Cloud service or somewhere you can then give me access to, and see my email address on my GitHub profile to send me the credentials. Still open a public issue for tracking and as documentation for other users. ``` -- -- EDIT THIS FILE TO MATCH YOUR BUG REPORT -- LOAD CSV FROM INLINE with encoding 'ascii' INTO postgresql:///pgloader TARGET TABLE jordane WITH truncate, fields terminated by '|', fields not enclosed, fields escaped by backslash-quote SET work_mem to '128MB', standard_conforming_strings to 'on' BEFORE LOAD DO $$ drop table if exists jordane; $$, $$ CREATE TABLE jordane ( "NOM" character(20), "PRENOM" character(20) ) $$; BORDET|Jordane BORDET|Audrey LASTNAME|"opening quote BONNIER|testprenombe~aucouptroplong JOURDAIN|héhé¶ ``` - [ ] pgloader output you obtain ``` PASTE HERE THE OUTPUT OF THE PGLOADER COMMAND ``` - [ ] data that is being loaded, if relevant ``` PASTE HERE THE DATA THAT HAS BEEN LOADED ``` - [ ] How the data is different from what you expected, if relevant
non_priority
aws rds to aws vm thanks for contributing to by reporting an issue reporting an issue is the only way we can solve problems fix bugs and improve both the software and its user experience in general the best bug reports follow those simple steps show what you did show the result you got explain how the result is not what you expected in the case of pgloader here s the information i will need to read in your bug report having all of this is a big help and often means the bug you reported can be fixed very efficiently as soon as i get to it please provide the following information pgloader version did you test a fresh compile from the source tree compiling pgloader from sources is documented in the it s easy to do and if patches are to be made to fix your bug you re going to have to build from sources to get the fix anyway… did you search for other similar issues how can i reproduce the bug incude a self contained pgloader command file if you re loading from a database consider attaching a database dump to your issue for mysql use mysqldump for sqlite just send over your source file that s easy maybe be the one with your production data of course the one with just the sample of data that allows me to reproduce your bug when using a proprietary database system as a source consider creating a sample database on some cloud service or somewhere you can then give me access to and see my email address on my github profile to send me the credentials still open a public issue for tracking and as documentation for other users edit this file to match your bug report load csv from inline with encoding ascii into postgresql pgloader target table jordane with truncate fields terminated by fields not enclosed fields escaped by backslash quote set work mem to standard conforming strings to on before load do drop table if exists jordane create table jordane nom character prenom character bordet jordane bordet audrey lastname opening quote bonnier testprenombe aucouptroplong jourdain héhé¶ pgloader output you obtain paste here the output of the pgloader command data that is being loaded if relevant paste here the data that has been loaded how the data is different from what you expected if relevant
0
751,168
26,231,736,536
IssuesEvent
2023-01-05 01:14:06
canonical/cn.ubuntu.com
https://api.github.com/repos/canonical/cn.ubuntu.com
closed
Empty links on Ubuntu Core特点概览
Priority: High
I found 4 links not working on https://cn.ubuntu.com/internet-of-things/core/features The current links and correct links are: - OTA 更新 - https://cn.ubuntu.com/internet-of-things/features/ota-updates >> https://cn.ubuntu.com/internet-of-things/core/features/ota-updates - 安全启动 - https://cn.ubuntu.com/internet-of-things/features/secure-boot >> https://cn.ubuntu.com/internet-of-things/core/features/secure-boot - 全盘加密 - https://cn.ubuntu.com/internet-of-things/features/full-disk-encryption >> https://cn.ubuntu.com/internet-of-things/core/features/full-disk-encryption - 恢复模式 - https://cn.ubuntu.com/internet-of-things/features/recovery >> https://cn.ubuntu.com/internet-of-things/core/features/recovery thanks ![image](https://user-images.githubusercontent.com/118874406/210477604-2e512db5-e28c-44f6-93fd-97cb7f083a95.png) ![image](https://user-images.githubusercontent.com/118874406/210477641-dc5b60cb-1613-48af-842b-b98295b82a39.png)
1.0
Empty links on Ubuntu Core特点概览 - I found 4 links not working on https://cn.ubuntu.com/internet-of-things/core/features The current links and correct links are: - OTA 更新 - https://cn.ubuntu.com/internet-of-things/features/ota-updates >> https://cn.ubuntu.com/internet-of-things/core/features/ota-updates - 安全启动 - https://cn.ubuntu.com/internet-of-things/features/secure-boot >> https://cn.ubuntu.com/internet-of-things/core/features/secure-boot - 全盘加密 - https://cn.ubuntu.com/internet-of-things/features/full-disk-encryption >> https://cn.ubuntu.com/internet-of-things/core/features/full-disk-encryption - 恢复模式 - https://cn.ubuntu.com/internet-of-things/features/recovery >> https://cn.ubuntu.com/internet-of-things/core/features/recovery thanks ![image](https://user-images.githubusercontent.com/118874406/210477604-2e512db5-e28c-44f6-93fd-97cb7f083a95.png) ![image](https://user-images.githubusercontent.com/118874406/210477641-dc5b60cb-1613-48af-842b-b98295b82a39.png)
priority
empty links on ubuntu core特点概览 i found links not working on the current links and correct links are ota 更新 安全启动 全盘加密 恢复模式 thanks
1
589,827
17,761,599,021
IssuesEvent
2021-08-29 20:00:53
nic547/TauStellwerk
https://api.github.com/repos/nic547/TauStellwerk
closed
Replace "Random User $number" default usernames
priority:medium type:maintenance size:S
Differentiating users based on a number is kinda difficult. Doing something like adjective + animal name or a similar approach would probably be better.
1.0
Replace "Random User $number" default usernames - Differentiating users based on a number is kinda difficult. Doing something like adjective + animal name or a similar approach would probably be better.
priority
replace random user number default usernames differentiating users based on a number is kinda difficult doing something like adjective animal name or a similar approach would probably be better
1
571,702
17,023,349,137
IssuesEvent
2021-07-03 01:33:46
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Public Libraries should show in Mapnik
Component: mapnik Priority: minor Resolution: fixed Type: enhancement
**[Submitted to the original trac issue database at 10.53pm, Tuesday, 13th January 2009]** Libraries (specifically a node with amenity=library) show on osmarender but not mapnik. Examples: Forbes Library is the public library of Northampton, MA, USA: http://openstreetmap.org/?lat=42.31708&lon=-72.63611&zoom=17&layers=B000FTFT There are many libraries here (Boston, MA, USA): http://openstreetmap.org/?lat=42.33889&lon=-71.09373&zoom=17&layers=0B00FTF The Reed Free Library in Surry, NH, USA even has the ref parameter set: http://openstreetmap.org/?lat=43.01949&lon=-72.32138&zoom=17&layers=0B00FTFT
1.0
Public Libraries should show in Mapnik - **[Submitted to the original trac issue database at 10.53pm, Tuesday, 13th January 2009]** Libraries (specifically a node with amenity=library) show on osmarender but not mapnik. Examples: Forbes Library is the public library of Northampton, MA, USA: http://openstreetmap.org/?lat=42.31708&lon=-72.63611&zoom=17&layers=B000FTFT There are many libraries here (Boston, MA, USA): http://openstreetmap.org/?lat=42.33889&lon=-71.09373&zoom=17&layers=0B00FTF The Reed Free Library in Surry, NH, USA even has the ref parameter set: http://openstreetmap.org/?lat=43.01949&lon=-72.32138&zoom=17&layers=0B00FTFT
priority
public libraries should show in mapnik libraries specifically a node with amenity library show on osmarender but not mapnik examples forbes library is the public library of northampton ma usa there are many libraries here boston ma usa the reed free library in surry nh usa even has the ref parameter set
1
26,505
13,039,424,841
IssuesEvent
2020-07-28 16:43:45
aivivn/d2l-vn
https://api.github.com/repos/aivivn/d2l-vn
closed
Revise "hybridize_vn" - Phần 3
chapter: computational-performance status: phase 2
Phần này được dịch bởi: **davidnvq** Nếu bạn đã dịch phần này, vui lòng bỏ qua việc revise.
True
Revise "hybridize_vn" - Phần 3 - Phần này được dịch bởi: **davidnvq** Nếu bạn đã dịch phần này, vui lòng bỏ qua việc revise.
non_priority
revise hybridize vn phần phần này được dịch bởi davidnvq nếu bạn đã dịch phần này vui lòng bỏ qua việc revise
0
163,377
25,803,200,607
IssuesEvent
2022-12-11 06:39:31
jcklpe/research-ops-hackpack
https://api.github.com/repos/jcklpe/research-ops-hackpack
opened
Create new research project: {Project}
design product management research
### Description blah blah ### Acceptance Criteria - [ ] Create a research [Epic](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign%2Cproduct+management%2Cepic&template=epic.yaml&title=%7BProject%2FTrial%7D+Epic). - [ ] Create a research project folder. - [ ] Create [project brief issue](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign&template=project-brief.yaml&title=Write+new+project+brief%3A+%7BProject%7D+). - [ ] All tickets have been created and assigned
1.0
Create new research project: {Project} - ### Description blah blah ### Acceptance Criteria - [ ] Create a research [Epic](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign%2Cproduct+management%2Cepic&template=epic.yaml&title=%7BProject%2FTrial%7D+Epic). - [ ] Create a research project folder. - [ ] Create [project brief issue](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign&template=project-brief.yaml&title=Write+new+project+brief%3A+%7BProject%7D+). - [ ] All tickets have been created and assigned
non_priority
create new research project project description blah blah acceptance criteria create a research create a research project folder create all tickets have been created and assigned
0
124,829
16,668,906,745
IssuesEvent
2021-06-07 08:28:12
blockframes/blockframes
https://api.github.com/repos/blockframes/blockframes
closed
Warning State on Avails Criteria Search Forms
Design - UI
Objective here is to add an intermediary state between hint and error, which would be warning. First use case for that warning state is for the avails form: as soon as users fill in one of the avails fields, it should be clear that they have to fill in all the other ones as well before applying their search. Research is needed to know what is possible with Angular forms
1.0
Warning State on Avails Criteria Search Forms - Objective here is to add an intermediary state between hint and error, which would be warning. First use case for that warning state is for the avails form: as soon as users fill in one of the avails fields, it should be clear that they have to fill in all the other ones as well before applying their search. Research is needed to know what is possible with Angular forms
non_priority
warning state on avails criteria search forms objective here is to add an intermediary state between hint and error which would be warning first use case for that warning state is for the avails form as soon as users fill in one of the avails fields it should be clear that they have to fill in all the other ones as well before applying their search research is needed to know what is possible with angular forms
0
492,490
14,214,384,107
IssuesEvent
2020-11-17 05:06:04
bounswe/bounswe2020group3
https://api.github.com/repos/bounswe/bounswe2020group3
closed
[Backend] Create Dockerfile
Backend Priority: Critical Type: Enhancement
* **Project: BACKEND** * **This is a: FEATURE REQUEST** * **Description of the issue** There should be a dockerfile for CI/CD configurations * **Deadline for resolution:** ASAP
1.0
[Backend] Create Dockerfile - * **Project: BACKEND** * **This is a: FEATURE REQUEST** * **Description of the issue** There should be a dockerfile for CI/CD configurations * **Deadline for resolution:** ASAP
priority
 create dockerfile project backend this is a feature request description of the issue there should be a dockerfile for ci cd configurations deadline for resolution asap
1
59,982
3,117,659,987
IssuesEvent
2015-09-04 03:51:09
framingeinstein/issues-test
https://api.github.com/repos/framingeinstein/issues-test
opened
SPK-613: Promo Code Help
priority:normal priority:normal priority:normal priority:normal priority:normal priority:normal resolution:in-progress resolution:in-progress resolution:in-progress
Hi Toby, See SRP-90: We entered a new ticket with Magento to request help on a promo code. 1 SKU that should be eligible is the only SKU not allowing a promo code to be used. Magento reviewed this issue on our front end and believes the custom theme is creating the promo code issue. Here is their feedback: Today I took a look at your site and I found that you have the special set up and running live in your production environment so I tested this out. I was able to reproduce what you have described and realized something: The system accepted the coupon, but didn't allow me the customer to have a 0.00$ choice when I got to shipping. This means the coupon process did work but your interface did not show the option on the front end. Now, the frontend has been customized and this is where the problem is. Your developers will need to look at the code for their onepage checkout page. As a test I recommend you or they disable your custom theme and test to see if it works using the native Magento theme.. I believe strongly it will work just fine. This is the strongest indicator the custom theme coding is the reason why the 0.00$ ground choice does not exist for the one sku you mentioned. If you revert to the Magento theme and it doesn't correct the behavior then yes make the backup and add the links to the backup files to this ticket and I will download this and look further into this, but if it does work using the magento theme then your developer will need to solve this custom problem.
6.0
SPK-613: Promo Code Help - Hi Toby, See SRP-90: We entered a new ticket with Magento to request help on a promo code. 1 SKU that should be eligible is the only SKU not allowing a promo code to be used. Magento reviewed this issue on our front end and believes the custom theme is creating the promo code issue. Here is their feedback: Today I took a look at your site and I found that you have the special set up and running live in your production environment so I tested this out. I was able to reproduce what you have described and realized something: The system accepted the coupon, but didn't allow me the customer to have a 0.00$ choice when I got to shipping. This means the coupon process did work but your interface did not show the option on the front end. Now, the frontend has been customized and this is where the problem is. Your developers will need to look at the code for their onepage checkout page. As a test I recommend you or they disable your custom theme and test to see if it works using the native Magento theme.. I believe strongly it will work just fine. This is the strongest indicator the custom theme coding is the reason why the 0.00$ ground choice does not exist for the one sku you mentioned. If you revert to the Magento theme and it doesn't correct the behavior then yes make the backup and add the links to the backup files to this ticket and I will download this and look further into this, but if it does work using the magento theme then your developer will need to solve this custom problem.
priority
spk promo code help hi toby see srp we entered a new ticket with magento to request help on a promo code sku that should be eligible is the only sku not allowing a promo code to be used magento reviewed this issue on our front end and believes the custom theme is creating the promo code issue here is their feedback today i took a look at your site and i found that you have the special set up and running live in your production environment so i tested this out i was able to reproduce what you have described and realized something the system accepted the coupon but didn t allow me the customer to have a choice when i got to shipping this means the coupon process did work but your interface did not show the option on the front end now the frontend has been customized and this is where the problem is your developers will need to look at the code for their onepage checkout page as a test i recommend you or they disable your custom theme and test to see if it works using the native magento theme i believe strongly it will work just fine this is the strongest indicator the custom theme coding is the reason why the ground choice does not exist for the one sku you mentioned if you revert to the magento theme and it doesn t correct the behavior then yes make the backup and add the links to the backup files to this ticket and i will download this and look further into this but if it does work using the magento theme then your developer will need to solve this custom problem
1
769,778
27,018,820,732
IssuesEvent
2023-02-10 22:25:12
googleapis/nodejs-automl
https://api.github.com/repos/googleapis/nodejs-automl
closed
Automl Vision Object Detection Deploy Model Test: should deploy a model with a specified node count failed
type: bug priority: p1 api: automl flakybot: issue flakybot: flaky
Note: #688 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 97411a2bb514b9921bb3932543a2d895c452d5c6 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/407b55a7-9544-44ff-be2c-1f1106f810da), [Sponge](http://sponge2/407b55a7-9544-44ff-be2c-1f1106f810da) status: failed <details><summary>Test output</summary><br><pre>expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/ AssertionError: expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/ at Context.<anonymous> (test/vision_object_detection_deploy_model_node_count.test.js:39:12)</pre></details>
1.0
Automl Vision Object Detection Deploy Model Test: should deploy a model with a specified node count failed - Note: #688 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 97411a2bb514b9921bb3932543a2d895c452d5c6 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/407b55a7-9544-44ff-be2c-1f1106f810da), [Sponge](http://sponge2/407b55a7-9544-44ff-be2c-1f1106f810da) status: failed <details><summary>Test output</summary><br><pre>expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/ AssertionError: expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/ at Context.<anonymous> (test/vision_object_detection_deploy_model_node_count.test.js:39:12)</pre></details>
priority
automl vision object detection deploy model test should deploy a model with a specified node count failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output expected unauthenticated request had inval… to match not found assertionerror expected unauthenticated request had inval… to match not found at context test vision object detection deploy model node count test js
1
304,756
9,335,313,785
IssuesEvent
2019-03-28 18:15:52
robotframework/SeleniumLibrary
https://api.github.com/repos/robotframework/SeleniumLibrary
closed
Set Browser Implicit Wait weird behavior
bug priority: medium
* Using the keyword Set Browser Implicit Wait will set the implicit wait but it can not be read with Get Selenium Implicit Wait, which returns only the value set with Set Selenium Implicit Wait or when importing the library. * When you open a new tab (with js for exemple) and go back to the main tab the implicit wait is gone but it will read the last value that have been set with Set Selenium Implicit Wait. ## Steps to reproduce the issue ``` *** Test Cases *** # Open the browser and stuff... Log implicit wait before Set Browser Implicit Wait 3s Log implicit wait after set browser Set Selenium Implicit Wait 2s Log implicit wait after set 2s Execute Javascript window.open("about:blank") Select Window NEW Select Window MAIN # this will output 'after tab switch = 2 seconds' but the real value will be 0.0 Log implicit wait after tab switch *** Keyword *** Log implicit wait [Arguments] ${prefix} ${implicit_wait} = Get Selenium Implicit Wait Log ${prefix} = ${implicit_wait} console=True ``` ## Environment Browser: Google Chrome 73.0.3683.75 Browser driver: chromedriver=2.44.609538 Operating System: Windows 10 x64 Libraries - Robot Framework: 3.1 - Selenium: 3.141.0 - SeleniumLibrary: 3.3.1 - Interpreter: Python 3.7.0b4 on win32
1.0
Set Browser Implicit Wait weird behavior - * Using the keyword Set Browser Implicit Wait will set the implicit wait but it can not be read with Get Selenium Implicit Wait, which returns only the value set with Set Selenium Implicit Wait or when importing the library. * When you open a new tab (with js for exemple) and go back to the main tab the implicit wait is gone but it will read the last value that have been set with Set Selenium Implicit Wait. ## Steps to reproduce the issue ``` *** Test Cases *** # Open the browser and stuff... Log implicit wait before Set Browser Implicit Wait 3s Log implicit wait after set browser Set Selenium Implicit Wait 2s Log implicit wait after set 2s Execute Javascript window.open("about:blank") Select Window NEW Select Window MAIN # this will output 'after tab switch = 2 seconds' but the real value will be 0.0 Log implicit wait after tab switch *** Keyword *** Log implicit wait [Arguments] ${prefix} ${implicit_wait} = Get Selenium Implicit Wait Log ${prefix} = ${implicit_wait} console=True ``` ## Environment Browser: Google Chrome 73.0.3683.75 Browser driver: chromedriver=2.44.609538 Operating System: Windows 10 x64 Libraries - Robot Framework: 3.1 - Selenium: 3.141.0 - SeleniumLibrary: 3.3.1 - Interpreter: Python 3.7.0b4 on win32
priority
set browser implicit wait weird behavior using the keyword set browser implicit wait will set the implicit wait but it can not be read with get selenium implicit wait which returns only the value set with set selenium implicit wait or when importing the library when you open a new tab with js for exemple and go back to the main tab the implicit wait is gone but it will read the last value that have been set with set selenium implicit wait steps to reproduce the issue test cases open the browser and stuff log implicit wait before set browser implicit wait log implicit wait after set browser set selenium implicit wait log implicit wait after set execute javascript window open about blank select window new select window main this will output after tab switch seconds but the real value will be log implicit wait after tab switch keyword log implicit wait prefix implicit wait get selenium implicit wait log prefix implicit wait console true environment browser google chrome browser driver chromedriver operating system windows libraries robot framework selenium seleniumlibrary interpreter python on
1
282,822
30,889,436,543
IssuesEvent
2023-08-04 02:43:18
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2023-28772 (Medium) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2023-28772 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.13.3. lib/seq_buf.c has a seq_buf_putmem_hex buffer overflow. <p>Publish Date: 2023-03-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28772>CVE-2023-28772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-28772">https://www.linuxkernelcves.com/cves/CVE-2023-28772</a></p> <p>Release Date: 2023-03-23</p> <p>Fix Resolution: v4.4.276,v4.9.276,v4.14.240,v4.19.198,v5.4.133,v5.10.51,v5.12.18,v5.13.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-28772 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2023-28772 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.13.3. lib/seq_buf.c has a seq_buf_putmem_hex buffer overflow. <p>Publish Date: 2023-03-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28772>CVE-2023-28772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-28772">https://www.linuxkernelcves.com/cves/CVE-2023-28772</a></p> <p>Release Date: 2023-03-23</p> <p>Fix Resolution: v4.4.276,v4.9.276,v4.14.240,v4.19.198,v5.4.133,v5.10.51,v5.12.18,v5.13.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files lib seq buf c lib seq buf c vulnerability details an issue was discovered in the linux kernel before lib seq buf c has a seq buf putmem hex buffer overflow publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
277,830
8,633,238,693
IssuesEvent
2018-11-22 13:17:39
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Eco Forum Unsecure Login.
High Priority
The official eco forum (http://ecoforum.strangeloopgames.com) uses insecure http for logging into the forum, Anyone who would be logging into the forum would be sending their username and password combination in cleartext over the internet allowing anyone sniffing traffic to steal an admin password for the forum. You can get a FREE SSL cert through lets encrypt.
1.0
Eco Forum Unsecure Login. - The official eco forum (http://ecoforum.strangeloopgames.com) uses insecure http for logging into the forum, Anyone who would be logging into the forum would be sending their username and password combination in cleartext over the internet allowing anyone sniffing traffic to steal an admin password for the forum. You can get a FREE SSL cert through lets encrypt.
priority
eco forum unsecure login the official eco forum uses insecure http for logging into the forum anyone who would be logging into the forum would be sending their username and password combination in cleartext over the internet allowing anyone sniffing traffic to steal an admin password for the forum you can get a free ssl cert through lets encrypt
1
239,072
26,201,215,106
IssuesEvent
2023-01-03 17:38:40
MValle21/circonus-unified-agent
https://api.github.com/repos/MValle21/circonus-unified-agent
opened
CVE-2020-26160 (High) detected in github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible
security vulnerability
## CVE-2020-26160 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible</b></p></summary> <p>ARCHIVE - Golang implementation of JSON Web Tokens (JWT). This project is now maintained at:</p> <p>Library home page: <a href="https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip">https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip</a></p> <p> Dependency Hierarchy: - :x: **github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jwt-go before 4.0.0-preview1 allows attackers to bypass intended access restrictions in situations with []string{} for m["aud"] (which is allowed by the specification). Because the type assertion fails, "" is the value of aud. This is a security problem if the JWT token is presented to a service that lacks its own audience check. <p>Publish Date: 2020-09-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-26160>CVE-2020-26160</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-26160">https://nvd.nist.gov/vuln/detail/CVE-2020-26160</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: v4.0.0-preview1</p> </p> </details> <p></p>
True
CVE-2020-26160 (High) detected in github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible - ## CVE-2020-26160 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible</b></p></summary> <p>ARCHIVE - Golang implementation of JSON Web Tokens (JWT). This project is now maintained at:</p> <p>Library home page: <a href="https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip">https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip</a></p> <p> Dependency Hierarchy: - :x: **github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jwt-go before 4.0.0-preview1 allows attackers to bypass intended access restrictions in situations with []string{} for m["aud"] (which is allowed by the specification). Because the type assertion fails, "" is the value of aud. This is a security problem if the JWT token is presented to a service that lacks its own audience check. <p>Publish Date: 2020-09-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-26160>CVE-2020-26160</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-26160">https://nvd.nist.gov/vuln/detail/CVE-2020-26160</a></p> <p>Release Date: 2020-09-30</p> <p>Fix Resolution: v4.0.0-preview1</p> </p> </details> <p></p>
non_priority
cve high detected in github com dgrijalva jwt go incompatible cve high severity vulnerability vulnerable library github com dgrijalva jwt go incompatible archive golang implementation of json web tokens jwt this project is now maintained at library home page a href dependency hierarchy x github com dgrijalva jwt go incompatible vulnerable library found in base branch master vulnerability details jwt go before allows attackers to bypass intended access restrictions in situations with string for m which is allowed by the specification because the type assertion fails is the value of aud this is a security problem if the jwt token is presented to a service that lacks its own audience check publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
175,540
21,313,848,145
IssuesEvent
2022-04-16 01:08:57
Nivaskumark/kernel_v4.1.15
https://api.github.com/repos/Nivaskumark/kernel_v4.1.15
opened
CVE-2019-3459 (Medium) detected in linuxlinux-4.6
security vulnerability
## CVE-2019-3459 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A heap address information leak while using L2CAP_GET_CONF_OPT was discovered in the Linux kernel before 5.1-rc1. <p>Publish Date: 2019-04-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3459>CVE-2019-3459</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459</a></p> <p>Release Date: 2019-04-11</p> <p>Fix Resolution: v5.1-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-3459 (Medium) detected in linuxlinux-4.6 - ## CVE-2019-3459 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A heap address information leak while using L2CAP_GET_CONF_OPT was discovered in the Linux kernel before 5.1-rc1. <p>Publish Date: 2019-04-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3459>CVE-2019-3459</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459</a></p> <p>Release Date: 2019-04-11</p> <p>Fix Resolution: v5.1-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files net bluetooth core c net bluetooth core c vulnerability details a heap address information leak while using get conf opt was discovered in the linux kernel before publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
350,772
31,932,305,166
IssuesEvent
2023-09-19 08:15:59
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix jax_lax_operators.test_jax_expand_dims
JAX Frontend Sub Task Failing Test
| | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix jax_lax_operators.test_jax_expand_dims - | | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
non_priority
fix jax lax operators test jax expand dims numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
0
474,379
13,658,034,303
IssuesEvent
2020-09-28 06:56:08
AY2021S1-CS2103T-T11-3/tp
https://api.github.com/repos/AY2021S1-CS2103T-T11-3/tp
closed
Add feature: add finance records
priority.High type.Story
As a forgetful business owner, I can save my transaction history, so that I can track my financials easily. As a small business owner, I can view a summary of my finances, so that I can plan the next steps of my business.
1.0
Add feature: add finance records - As a forgetful business owner, I can save my transaction history, so that I can track my financials easily. As a small business owner, I can view a summary of my finances, so that I can plan the next steps of my business.
priority
add feature add finance records as a forgetful business owner i can save my transaction history so that i can track my financials easily as a small business owner i can view a summary of my finances so that i can plan the next steps of my business
1
101,684
31,394,752,004
IssuesEvent
2023-08-26 19:48:06
expo/expo
https://api.github.com/repos/expo/expo
opened
Android 13 (Media Library Permissions)
needs validation Development Builds
### Summary When invoking [`await MediaLibrary.requestPermissionsAsync()`](https://docs.expo.dev/versions/latest/sdk/media-library/#medialibraryrequestpermissionsasyncwriteonly) on Android 13, two permission prompts appear asking for music/audio followed by photos/videos. According to the android developer docs in Android 13 or higher they have introduced more [granular permissions](https://developer.android.com/about/versions/13/behavior-changes-13#granular-media-permissions). I'm interested in only presenting the user with a single prompt (as I don't need to ask for music/audio) which appears would be [`READ_MEDIA_IMAGES`](https://developer.android.com/reference/android/Manifest.permission#READ_MEDIA_IMAGES), but adding that permission doesn't seem to do the trick. I also don't see that permission in the list of [available permissions](https://docs.expo.dev/versions/latest/config/app/#permissions). Lastly it looks like this was [addressed](https://github.com/expo/expo/pull/20907) in the past, but perhaps I'm missing something. **Screenshots** <img src="https://github.com/expo/expo/assets/7604441/290f8bd9-f380-4384-968f-67a66c889b22" width="200"/> <img src="https://github.com/expo/expo/assets/7604441/8599599c-80ff-469e-9956-aeafe903ed6f" width="200"/> <img src="https://github.com/expo/expo/assets/7604441/9ae1af49-4999-4eea-97ee-c2967711e786" width="200"/> ### Managed or bare workflow? managed ### What platform(s) does this occur on? Android ### Package versions ```json "expo": "~49.0.8", "expo-dev-client": "~2.4.8", "expo-media-library": "~15.4.1", "expo-status-bar": "~1.6.0", "expo-updates": "~0.18.12", "react": "18.2.0", "react-native": "0.72.4" ``` ### Environment ``` expo-env-info 1.0.5 environment info: System: OS: macOS 13.5 Shell: 5.9 - /bin/zsh Binaries: Node: 16.16.0 - /usr/local/bin/node Yarn: 1.22.19 - ~/.yarn/bin/yarn npm: 9.8.1 - /usr/local/bin/npm Managers: CocoaPods: 1.11.3 - /opt/homebrew/bin/pod SDKs: iOS SDK: Platforms: DriverKit 22.4, iOS 16.4, macOS 13.3, tvOS 16.4, watchOS 9.4 IDEs: Android Studio: 2022.3 AI-223.8836.35.2231.10406996 Xcode: 14.3.1/14E300c - /usr/bin/xcodebuild npmPackages: expo: ~49.0.8 => 49.0.8 react: 18.2.0 => 18.2.0 react-native: 0.72.4 => 0.72.4 npmGlobalPackages: eas-cli: 4.1.2 expo-cli: 6.3.10 Expo Workflow: managed ``` ### Reproducible demo https://github.com/iM-GeeKy/media-permissions ### Stacktrace (if a crash is involved) _No response_
1.0
Android 13 (Media Library Permissions) - ### Summary When invoking [`await MediaLibrary.requestPermissionsAsync()`](https://docs.expo.dev/versions/latest/sdk/media-library/#medialibraryrequestpermissionsasyncwriteonly) on Android 13, two permission prompts appear asking for music/audio followed by photos/videos. According to the android developer docs in Android 13 or higher they have introduced more [granular permissions](https://developer.android.com/about/versions/13/behavior-changes-13#granular-media-permissions). I'm interested in only presenting the user with a single prompt (as I don't need to ask for music/audio) which appears would be [`READ_MEDIA_IMAGES`](https://developer.android.com/reference/android/Manifest.permission#READ_MEDIA_IMAGES), but adding that permission doesn't seem to do the trick. I also don't see that permission in the list of [available permissions](https://docs.expo.dev/versions/latest/config/app/#permissions). Lastly it looks like this was [addressed](https://github.com/expo/expo/pull/20907) in the past, but perhaps I'm missing something. **Screenshots** <img src="https://github.com/expo/expo/assets/7604441/290f8bd9-f380-4384-968f-67a66c889b22" width="200"/> <img src="https://github.com/expo/expo/assets/7604441/8599599c-80ff-469e-9956-aeafe903ed6f" width="200"/> <img src="https://github.com/expo/expo/assets/7604441/9ae1af49-4999-4eea-97ee-c2967711e786" width="200"/> ### Managed or bare workflow? managed ### What platform(s) does this occur on? Android ### Package versions ```json "expo": "~49.0.8", "expo-dev-client": "~2.4.8", "expo-media-library": "~15.4.1", "expo-status-bar": "~1.6.0", "expo-updates": "~0.18.12", "react": "18.2.0", "react-native": "0.72.4" ``` ### Environment ``` expo-env-info 1.0.5 environment info: System: OS: macOS 13.5 Shell: 5.9 - /bin/zsh Binaries: Node: 16.16.0 - /usr/local/bin/node Yarn: 1.22.19 - ~/.yarn/bin/yarn npm: 9.8.1 - /usr/local/bin/npm Managers: CocoaPods: 1.11.3 - /opt/homebrew/bin/pod SDKs: iOS SDK: Platforms: DriverKit 22.4, iOS 16.4, macOS 13.3, tvOS 16.4, watchOS 9.4 IDEs: Android Studio: 2022.3 AI-223.8836.35.2231.10406996 Xcode: 14.3.1/14E300c - /usr/bin/xcodebuild npmPackages: expo: ~49.0.8 => 49.0.8 react: 18.2.0 => 18.2.0 react-native: 0.72.4 => 0.72.4 npmGlobalPackages: eas-cli: 4.1.2 expo-cli: 6.3.10 Expo Workflow: managed ``` ### Reproducible demo https://github.com/iM-GeeKy/media-permissions ### Stacktrace (if a crash is involved) _No response_
non_priority
android media library permissions summary when invoking on android two permission prompts appear asking for music audio followed by photos videos according to the android developer docs in android or higher they have introduced more i m interested in only presenting the user with a single prompt as i don t need to ask for music audio which appears would be but adding that permission doesn t seem to do the trick i also don t see that permission in the list of lastly it looks like this was in the past but perhaps i m missing something screenshots managed or bare workflow managed what platform s does this occur on android package versions json expo expo dev client expo media library expo status bar expo updates react react native environment expo env info environment info system os macos shell bin zsh binaries node usr local bin node yarn yarn bin yarn npm usr local bin npm managers cocoapods opt homebrew bin pod sdks ios sdk platforms driverkit ios macos tvos watchos ides android studio ai xcode usr bin xcodebuild npmpackages expo react react native npmglobalpackages eas cli expo cli expo workflow managed reproducible demo stacktrace if a crash is involved no response
0
133,685
10,855,606,209
IssuesEvent
2019-11-13 18:46:14
rancher/k3s
https://api.github.com/repos/rancher/k3s
closed
Upgrade to v0.10.0-rc2 to use default kubelet directory duplicates pod entry in old as well as in new directories.
[zube]: To Test
**Version:** Upgrade from v0.9.1 to v0.10.0-rc2 **Describe the bug** 1. Upgrade to v0.10.0-rc2 has all pod entries from /var/lib/rancher/k3s/agent/kubelet/pods/ copied onto /var/lib/kubelet except helm and the pods are restarted. 2. Delete a pod removes in the new default dir and not from the old dir. 3. Helm pod is present only in old dir and not copied to default dir as job is in Completed state. **To Reproduce** 1. Install v0.9.1 `curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.9.1 sh -s - server` Create a deployment. Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/ Upgrade to rc2 `curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.10.0-rc2 sh -s - server` Verify the pods status `kubectl get pods -A` Verify the entries in /var/lib/kubelet Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/ Scale up the pods Verify the new entries in /var/lib/kubelet As expected the new pods are not in /var/lib/rancher/k3s/agent/kubelet/pods/ Scale down the replicas to 0. Pod entries are removed from new default dir and continues to appear in old dir. 5. Helm is not copied to new dir `ls /var/lib/kubelet/pods/container/` |grep helm **Expected behavior** Pods created after upgrade should be in new default dir Pods created before upgrade continue to remain in old dir All pods are in Running state Deleting a pod removes it from respective dir **Actual behavior** Pods created after upgrade should be in new default dir Pods created before upgrade continue to remain in old dir as well as in new dir All pods are in running state. Deleting a pod removes it from new dir only *** Additional info *** ``` ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods total 28 drwxr-x--- 5 root root 4096 Oct 21 18:19 ccd0be75-173e-4df4-aa1e-57979ca63a93 drwxr-x--- 5 root root 4096 Oct 21 18:19 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6 drwxr-x--- 5 root root 4096 Oct 21 18:19 946d4695-d059-470c-a074-4089bece0944 drwxr-x--- 5 root root 4096 Oct 21 18:19 0456318c-2e28-47f4-95e2-ccdc7b712b8b drwxr-x--- 5 root root 4096 Oct 21 18:19 7e056156-0b88-442d-a841-1df67def1ae6 drwxr-x--- 5 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf drwxr-x--- 5 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb ls -ltr /var/lib/kubelet/pods total 24 drwxr-x--- 4 root root 4096 Oct 21 18:20 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6 drwxr-x--- 4 root root 4096 Oct 21 18:20 0456318c-2e28-47f4-95e2-ccdc7b712b8b drwxr-x--- 4 root root 4096 Oct 21 18:20 7e056156-0b88-442d-a841-1df67def1ae6 drwxr-x--- 4 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb drwxr-x--- 4 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf drwxr-x--- 5 root root 4096 Oct 21 18:21 ccd0be75-173e-4df4-aa1e-57979ca63a93 ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods/946d4695-d059-470c-a074-4089bece0944/containers/helm/ total 0 ls -ltr /var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944 ls: cannot access '/var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944': No such file or directory ```
1.0
Upgrade to v0.10.0-rc2 to use default kubelet directory duplicates pod entry in old as well as in new directories. - **Version:** Upgrade from v0.9.1 to v0.10.0-rc2 **Describe the bug** 1. Upgrade to v0.10.0-rc2 has all pod entries from /var/lib/rancher/k3s/agent/kubelet/pods/ copied onto /var/lib/kubelet except helm and the pods are restarted. 2. Delete a pod removes in the new default dir and not from the old dir. 3. Helm pod is present only in old dir and not copied to default dir as job is in Completed state. **To Reproduce** 1. Install v0.9.1 `curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.9.1 sh -s - server` Create a deployment. Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/ Upgrade to rc2 `curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.10.0-rc2 sh -s - server` Verify the pods status `kubectl get pods -A` Verify the entries in /var/lib/kubelet Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/ Scale up the pods Verify the new entries in /var/lib/kubelet As expected the new pods are not in /var/lib/rancher/k3s/agent/kubelet/pods/ Scale down the replicas to 0. Pod entries are removed from new default dir and continues to appear in old dir. 5. Helm is not copied to new dir `ls /var/lib/kubelet/pods/container/` |grep helm **Expected behavior** Pods created after upgrade should be in new default dir Pods created before upgrade continue to remain in old dir All pods are in Running state Deleting a pod removes it from respective dir **Actual behavior** Pods created after upgrade should be in new default dir Pods created before upgrade continue to remain in old dir as well as in new dir All pods are in running state. Deleting a pod removes it from new dir only *** Additional info *** ``` ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods total 28 drwxr-x--- 5 root root 4096 Oct 21 18:19 ccd0be75-173e-4df4-aa1e-57979ca63a93 drwxr-x--- 5 root root 4096 Oct 21 18:19 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6 drwxr-x--- 5 root root 4096 Oct 21 18:19 946d4695-d059-470c-a074-4089bece0944 drwxr-x--- 5 root root 4096 Oct 21 18:19 0456318c-2e28-47f4-95e2-ccdc7b712b8b drwxr-x--- 5 root root 4096 Oct 21 18:19 7e056156-0b88-442d-a841-1df67def1ae6 drwxr-x--- 5 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf drwxr-x--- 5 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb ls -ltr /var/lib/kubelet/pods total 24 drwxr-x--- 4 root root 4096 Oct 21 18:20 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6 drwxr-x--- 4 root root 4096 Oct 21 18:20 0456318c-2e28-47f4-95e2-ccdc7b712b8b drwxr-x--- 4 root root 4096 Oct 21 18:20 7e056156-0b88-442d-a841-1df67def1ae6 drwxr-x--- 4 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb drwxr-x--- 4 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf drwxr-x--- 5 root root 4096 Oct 21 18:21 ccd0be75-173e-4df4-aa1e-57979ca63a93 ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods/946d4695-d059-470c-a074-4089bece0944/containers/helm/ total 0 ls -ltr /var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944 ls: cannot access '/var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944': No such file or directory ```
non_priority
upgrade to to use default kubelet directory duplicates pod entry in old as well as in new directories version upgrade from to describe the bug upgrade to has all pod entries from var lib rancher agent kubelet pods copied onto var lib kubelet except helm and the pods are restarted delete a pod removes in the new default dir and not from the old dir helm pod is present only in old dir and not copied to default dir as job is in completed state to reproduce install curl sfl install version sh s server create a deployment verify the entries in var lib rancher agent kubelet pods upgrade to curl sfl install version sh s server verify the pods status kubectl get pods a verify the entries in var lib kubelet verify the entries in var lib rancher agent kubelet pods scale up the pods verify the new entries in var lib kubelet as expected the new pods are not in var lib rancher agent kubelet pods scale down the replicas to pod entries are removed from new default dir and continues to appear in old dir helm is not copied to new dir ls var lib kubelet pods container grep helm expected behavior pods created after upgrade should be in new default dir pods created before upgrade continue to remain in old dir all pods are in running state deleting a pod removes it from respective dir actual behavior pods created after upgrade should be in new default dir pods created before upgrade continue to remain in old dir as well as in new dir all pods are in running state deleting a pod removes it from new dir only additional info ls ltr var lib rancher agent kubelet pods total drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct ls ltr var lib kubelet pods total drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct ls ltr var lib rancher agent kubelet pods containers helm total ls ltr var lib kubelet pods ls cannot access var lib kubelet pods no such file or directory
0
50,997
13,188,026,176
IssuesEvent
2020-08-13 05:20:36
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[g4-tankresponse] memory leaks (Trac #1796)
Migrated from Trac combo simulation defect
potential memory leaks found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1796">https://code.icecube.wisc.edu/ticket/1796</a>, reported by kjmeagher and owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2016-08-10T22:54:14", "description": "potential memory leaks found by static analysis: \nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "invalid", "_ts": "1470869654585407", "component": "combo simulation", "summary": "[g4-tankresponse] memory leaks", "priority": "normal", "keywords": "", "time": "2016-07-27T07:57:17", "milestone": "Long-Term Future", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
1.0
[g4-tankresponse] memory leaks (Trac #1796) - potential memory leaks found by static analysis: http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1796">https://code.icecube.wisc.edu/ticket/1796</a>, reported by kjmeagher and owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2016-08-10T22:54:14", "description": "potential memory leaks found by static analysis: \nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "invalid", "_ts": "1470869654585407", "component": "combo simulation", "summary": "[g4-tankresponse] memory leaks", "priority": "normal", "keywords": "", "time": "2016-07-27T07:57:17", "milestone": "Long-Term Future", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
non_priority
memory leaks trac potential memory leaks found by static analysis migrated from json status closed changetime description potential memory leaks found by static analysis n reporter kjmeagher cc resolution invalid ts component combo simulation summary memory leaks priority normal keywords time milestone long term future owner jgonzalez type defect
0
43,509
2,889,826,327
IssuesEvent
2015-06-13 20:00:42
damonkohler/android-scripting
https://api.github.com/repos/damonkohler/android-scripting
closed
Display additional info before installing interpretter
auto-migrated LowHangingFruit Priority-Medium Type-Enhancement
``` Interpretter APKs should display version and total download size. ``` Original issue reported on code.google.com by `MeanEYE.rcf` on 7 Jul 2010 at 10:51
1.0
Display additional info before installing interpretter - ``` Interpretter APKs should display version and total download size. ``` Original issue reported on code.google.com by `MeanEYE.rcf` on 7 Jul 2010 at 10:51
priority
display additional info before installing interpretter interpretter apks should display version and total download size original issue reported on code google com by meaneye rcf on jul at
1
187,839
6,761,563,228
IssuesEvent
2017-10-25 02:35:19
opencollective/opencollective
https://api.github.com/repos/opencollective/opencollective
closed
Donor can't make a donation: logged in, system prompts her to login again
high priority
Thank you for taking the time to report an issue 🙏 The easier it is for us to reproduce it, the faster we can solve it. So please try to be as complete as possible when filing your issue. *** From user: > The donation form beginning is a bit confusing for her: She’s already logged in, but placing her email in the form prompts her to log in again. Is this normal? URL: https://opencollective.com/buttercup/donate Logged in as: https://opencollective.com/ida-mitchell Error message if any: Expected result: being able to make a donation Browser: Bonus point if you can add a screenshot :-) <img width="1328" alt="pastedgraphic-3" src="https://user-images.githubusercontent.com/3671070/31959680-5bdf7e36-b8c3-11e7-94c3-5194243b1169.png"> Thank you and have a wonderful day/evening! (and sorry for inconvenience! We'll do our best to solve this ASAP)
1.0
Donor can't make a donation: logged in, system prompts her to login again - Thank you for taking the time to report an issue 🙏 The easier it is for us to reproduce it, the faster we can solve it. So please try to be as complete as possible when filing your issue. *** From user: > The donation form beginning is a bit confusing for her: She’s already logged in, but placing her email in the form prompts her to log in again. Is this normal? URL: https://opencollective.com/buttercup/donate Logged in as: https://opencollective.com/ida-mitchell Error message if any: Expected result: being able to make a donation Browser: Bonus point if you can add a screenshot :-) <img width="1328" alt="pastedgraphic-3" src="https://user-images.githubusercontent.com/3671070/31959680-5bdf7e36-b8c3-11e7-94c3-5194243b1169.png"> Thank you and have a wonderful day/evening! (and sorry for inconvenience! We'll do our best to solve this ASAP)
priority
donor can t make a donation logged in system prompts her to login again thank you for taking the time to report an issue 🙏 the easier it is for us to reproduce it the faster we can solve it so please try to be as complete as possible when filing your issue from user the donation form beginning is a bit confusing for her she’s already logged in but placing her email in the form prompts her to log in again is this normal url logged in as error message if any expected result being able to make a donation browser bonus point if you can add a screenshot img width alt pastedgraphic src thank you and have a wonderful day evening and sorry for inconvenience we ll do our best to solve this asap
1
3,124
8,972,202,604
IssuesEvent
2019-01-29 17:39:59
firecracker-microvm/firecracker
https://api.github.com/repos/firecracker-microvm/firecracker
closed
Add Try_Read Functionality to EventFd Implementation
Contribute: Good First Issue Priority: Low Program: Architecture Quality: Improvement
We are currently using EventFds via the implementation from the sys_util crate, inherited from crosvm. One potentially significant downside is that it only exposes a blocking `read()` API. There are places in our code where we use `read()` knowing that an event should definitely be available, because the logic is programmed that way, but if bugs are introduces, or something goes terribly wrong otherwise, the code will simply hang on the `read()` (not even panic). Maybe it would make sense to add some sort of `try_read()` functionality, and do checked reads in this context. There will be some additional overhead involved, but it should not be significant.
1.0
Add Try_Read Functionality to EventFd Implementation - We are currently using EventFds via the implementation from the sys_util crate, inherited from crosvm. One potentially significant downside is that it only exposes a blocking `read()` API. There are places in our code where we use `read()` knowing that an event should definitely be available, because the logic is programmed that way, but if bugs are introduces, or something goes terribly wrong otherwise, the code will simply hang on the `read()` (not even panic). Maybe it would make sense to add some sort of `try_read()` functionality, and do checked reads in this context. There will be some additional overhead involved, but it should not be significant.
non_priority
add try read functionality to eventfd implementation we are currently using eventfds via the implementation from the sys util crate inherited from crosvm one potentially significant downside is that it only exposes a blocking read api there are places in our code where we use read knowing that an event should definitely be available because the logic is programmed that way but if bugs are introduces or something goes terribly wrong otherwise the code will simply hang on the read not even panic maybe it would make sense to add some sort of try read functionality and do checked reads in this context there will be some additional overhead involved but it should not be significant
0
19,773
5,932,300,454
IssuesEvent
2017-05-24 08:59:48
HGustavs/LenaSYS
https://api.github.com/repos/HGustavs/LenaSYS
closed
missleading text in code viewer dialog
All CodeViewer Needs Fixing
**In the picture below, should it really say "delete" and not "cancel"?** ![asd](https://cloud.githubusercontent.com/assets/26705142/26297324/af293390-3ed2-11e7-8255-24ef22cf6806.png)
1.0
missleading text in code viewer dialog - **In the picture below, should it really say "delete" and not "cancel"?** ![asd](https://cloud.githubusercontent.com/assets/26705142/26297324/af293390-3ed2-11e7-8255-24ef22cf6806.png)
non_priority
missleading text in code viewer dialog in the picture below should it really say delete and not cancel
0
57,893
16,131,877,731
IssuesEvent
2021-04-29 06:39:32
cython/cython
https://api.github.com/repos/cython/cython
closed
Default arguments in methods are not preserved for introspection
Python Semantics defect
Tried on the latest master: ``` # test.pyx def run(a, b=1): return a + b cdef class A: def run(self, a, b=1): return a + b # app.py from test import A, run import inspect a = A() print(inspect.signature(run)) # ok - prints (a, b=1) print(inspect.signature(a.run)) # not ok - prints (a, b) print(inspect.signature(A.run)) # not ok - prints (self, a, b) ```
1.0
Default arguments in methods are not preserved for introspection - Tried on the latest master: ``` # test.pyx def run(a, b=1): return a + b cdef class A: def run(self, a, b=1): return a + b # app.py from test import A, run import inspect a = A() print(inspect.signature(run)) # ok - prints (a, b=1) print(inspect.signature(a.run)) # not ok - prints (a, b) print(inspect.signature(A.run)) # not ok - prints (self, a, b) ```
non_priority
default arguments in methods are not preserved for introspection tried on the latest master test pyx def run a b return a b cdef class a def run self a b return a b app py from test import a run import inspect a a print inspect signature run ok prints a b print inspect signature a run not ok prints a b print inspect signature a run not ok prints self a b
0
71,401
3,356,465,895
IssuesEvent
2015-11-18 20:37:16
psouza4/mediacentermaster
https://api.github.com/repos/psouza4/mediacentermaster
closed
Create a file blacklist for parser
Affects-General-Usability Component-Functionality Feature-Downloads Priority-Medium Type-FeatureRequest
Create a user editable configuration file (JSON, XML,etc) to be processed for a blacklist of files to be deleted and ignored while parsing movies and tv shows. The blacklist will be a string value, and depending on code implementation, may support .Net regular expressions.
1.0
Create a file blacklist for parser - Create a user editable configuration file (JSON, XML,etc) to be processed for a blacklist of files to be deleted and ignored while parsing movies and tv shows. The blacklist will be a string value, and depending on code implementation, may support .Net regular expressions.
priority
create a file blacklist for parser create a user editable configuration file json xml etc to be processed for a blacklist of files to be deleted and ignored while parsing movies and tv shows the blacklist will be a string value and depending on code implementation may support net regular expressions
1
121,229
17,648,388,093
IssuesEvent
2021-08-20 09:37:23
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Security Solution]Rule name is clickable and navigated to rule page contradicting TGrid Form AC
bug Team: SecuritySolution
**Describe the bug** Rule name is clickable and navigated to rule page contradicting TGrid Form AC **Build Details** ``` Version:7.15.0 Commit:d791226d9385122f33f4a5ca38fa5369012fbec3 Build:43636 ``` **Browsers** all **Precondition** - Install the Endpoint Security on the kibana **Steps to Reproduce** 1. Generate few Alerts. 2. Observed that Rule name value is clickable and let user to its rule page which is contradicting the updated TGrid form of the alert list. ![image](https://user-images.githubusercontent.com/59917825/130213536-387176c0-2c48-4a92-b7ef-198d000b56b4.png) **Actual Result** Rule name is clickable and navigated to rule page contradicting TGrid Form AC **Expected Result** Rule name should not be clickable **Screen-Shoot** ![image](https://user-images.githubusercontent.com/59917825/130213595-d8dd62eb-f736-4e06-87dd-86f4eae48303.png) **logs** N/A
True
[Security Solution]Rule name is clickable and navigated to rule page contradicting TGrid Form AC - **Describe the bug** Rule name is clickable and navigated to rule page contradicting TGrid Form AC **Build Details** ``` Version:7.15.0 Commit:d791226d9385122f33f4a5ca38fa5369012fbec3 Build:43636 ``` **Browsers** all **Precondition** - Install the Endpoint Security on the kibana **Steps to Reproduce** 1. Generate few Alerts. 2. Observed that Rule name value is clickable and let user to its rule page which is contradicting the updated TGrid form of the alert list. ![image](https://user-images.githubusercontent.com/59917825/130213536-387176c0-2c48-4a92-b7ef-198d000b56b4.png) **Actual Result** Rule name is clickable and navigated to rule page contradicting TGrid Form AC **Expected Result** Rule name should not be clickable **Screen-Shoot** ![image](https://user-images.githubusercontent.com/59917825/130213595-d8dd62eb-f736-4e06-87dd-86f4eae48303.png) **logs** N/A
non_priority
rule name is clickable and navigated to rule page contradicting tgrid form ac describe the bug rule name is clickable and navigated to rule page contradicting tgrid form ac build details version commit build browsers all precondition install the endpoint security on the kibana steps to reproduce generate few alerts observed that rule name value is clickable and let user to its rule page which is contradicting the updated tgrid form of the alert list actual result rule name is clickable and navigated to rule page contradicting tgrid form ac expected result rule name should not be clickable screen shoot logs n a
0
64,443
26,735,622,277
IssuesEvent
2023-01-30 09:15:07
flipperdevices/flipperzero-firmware
https://api.github.com/repos/flipperdevices/flipperzero-firmware
closed
Show SD card CID in SD card info
feature request core+services
### Describe the enhancement you're suggesting. The card info shows filesystem type and size but not info on the physical hardware. When there are troubles with some cards it would be nice to be able to see which card type is used. ### Anything else? _No response_
1.0
Show SD card CID in SD card info - ### Describe the enhancement you're suggesting. The card info shows filesystem type and size but not info on the physical hardware. When there are troubles with some cards it would be nice to be able to see which card type is used. ### Anything else? _No response_
non_priority
show sd card cid in sd card info describe the enhancement you re suggesting the card info shows filesystem type and size but not info on the physical hardware when there are troubles with some cards it would be nice to be able to see which card type is used anything else no response
0
119,065
15,395,579,778
IssuesEvent
2021-03-03 19:25:27
emory-libraries/blacklight-catalog
https://api.github.com/repos/emory-libraries/blacklight-catalog
opened
Review current bootstrap CSS and recommend changes needed for current theming and branding
UI Design View (Display and Navigation)
As the product owner, I would like to begin exploring what is needed to begin theming and branding the Blacklight application. The current bootstrap used for Emory Digital Collections is based off the pattern library for Emory Libraries website. Based on initial design of wireframes for header, footer, homepage, facets, and single item display, I would like to understand from a UI design perspective if any modifications are needed to the Bootstrap CSS. This will allow us to begin development of needed components of the header and footer in the coming sprints. Acceptance criteria: - [ ] Review the [Bootstrap Mapping for Emory Digital Collections](https://docs.google.com/spreadsheets/d/1-sbatCXlB7ysNHPrj3vT2mBy4nufx1ovlG7UwX2sdo8/edit?usp=sharing) - [ ] Determine if any new components are needed at this point for the Blacklight Catalog based on current designs for header, footer, facets, home page and single item view page. - [ ] Document and review recommendations with product owner.
1.0
Review current bootstrap CSS and recommend changes needed for current theming and branding - As the product owner, I would like to begin exploring what is needed to begin theming and branding the Blacklight application. The current bootstrap used for Emory Digital Collections is based off the pattern library for Emory Libraries website. Based on initial design of wireframes for header, footer, homepage, facets, and single item display, I would like to understand from a UI design perspective if any modifications are needed to the Bootstrap CSS. This will allow us to begin development of needed components of the header and footer in the coming sprints. Acceptance criteria: - [ ] Review the [Bootstrap Mapping for Emory Digital Collections](https://docs.google.com/spreadsheets/d/1-sbatCXlB7ysNHPrj3vT2mBy4nufx1ovlG7UwX2sdo8/edit?usp=sharing) - [ ] Determine if any new components are needed at this point for the Blacklight Catalog based on current designs for header, footer, facets, home page and single item view page. - [ ] Document and review recommendations with product owner.
non_priority
review current bootstrap css and recommend changes needed for current theming and branding as the product owner i would like to begin exploring what is needed to begin theming and branding the blacklight application the current bootstrap used for emory digital collections is based off the pattern library for emory libraries website based on initial design of wireframes for header footer homepage facets and single item display i would like to understand from a ui design perspective if any modifications are needed to the bootstrap css this will allow us to begin development of needed components of the header and footer in the coming sprints acceptance criteria review the determine if any new components are needed at this point for the blacklight catalog based on current designs for header footer facets home page and single item view page document and review recommendations with product owner
0
228,517
7,552,437,773
IssuesEvent
2018-04-19 00:17:30
leo-project/leofs
https://api.github.com/repos/leo-project/leofs
closed
[eleveldb] Make log files less fragmented
Improve Priority-MIDDLE survey v1.4
As reported on https://github.com/leo-project/leofs/issues/940 by @vstax, eleveldb have a fragmentation problem for its .log files. This can be fixed by calling fallocate() or posix_fallocate() when creating .log file. and also we have to check how this fix would affect the performance of synced writes to .log file for safe.
1.0
[eleveldb] Make log files less fragmented - As reported on https://github.com/leo-project/leofs/issues/940 by @vstax, eleveldb have a fragmentation problem for its .log files. This can be fixed by calling fallocate() or posix_fallocate() when creating .log file. and also we have to check how this fix would affect the performance of synced writes to .log file for safe.
priority
make log files less fragmented as reported on by vstax eleveldb have a fragmentation problem for its log files this can be fixed by calling fallocate or posix fallocate when creating log file and also we have to check how this fix would affect the performance of synced writes to log file for safe
1
301,839
22,777,115,619
IssuesEvent
2022-07-08 15:27:48
orchest/orchest
https://api.github.com/repos/orchest/orchest
closed
Improve documentation of installation targets
improvement documentation
**Describe the problem this improvement solves** At the moment, we are documenting two different installation methods: - minikube - `kubectl` and, in addition, our convenience script. However, it would be good to add more clear separation between all these methods, and add other interesting targets such as EKS (see #930, #885). **Describe the solution you'd like** A very clear and easy to navigate installation page in which users can pick the method that best suits them.
1.0
Improve documentation of installation targets - **Describe the problem this improvement solves** At the moment, we are documenting two different installation methods: - minikube - `kubectl` and, in addition, our convenience script. However, it would be good to add more clear separation between all these methods, and add other interesting targets such as EKS (see #930, #885). **Describe the solution you'd like** A very clear and easy to navigate installation page in which users can pick the method that best suits them.
non_priority
improve documentation of installation targets describe the problem this improvement solves at the moment we are documenting two different installation methods minikube kubectl and in addition our convenience script however it would be good to add more clear separation between all these methods and add other interesting targets such as eks see describe the solution you d like a very clear and easy to navigate installation page in which users can pick the method that best suits them
0
240,645
18,363,637,398
IssuesEvent
2021-10-09 17:14:49
KylieScharf/flask_portfolio
https://api.github.com/repos/KylieScharf/flask_portfolio
opened
edit journal
documentation
Final Journals - [x] kylie and khushi [journal](https://docs.google.com/document/d/1eyTgQhMv7jFi28SIGlwOs3k95a9KklfJFCS0O9pXCzA/edit?usp=sharing) - [x] Kevin and Hamza and Daniel [journal](https://docs.google.com/document/d/1fy-J_PVrvafykD-OTxMm-Pe0L_UKrYUTC4-DP4vGXEw/edit?usp=sharing) Add 3.5 and 3.6 notes to journal - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Add test corrections to journal - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Add TPT notes to journal - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Add TT notes to journal (if any) - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Finish all journal entries - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel
1.0
edit journal - Final Journals - [x] kylie and khushi [journal](https://docs.google.com/document/d/1eyTgQhMv7jFi28SIGlwOs3k95a9KklfJFCS0O9pXCzA/edit?usp=sharing) - [x] Kevin and Hamza and Daniel [journal](https://docs.google.com/document/d/1fy-J_PVrvafykD-OTxMm-Pe0L_UKrYUTC4-DP4vGXEw/edit?usp=sharing) Add 3.5 and 3.6 notes to journal - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Add test corrections to journal - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Add TPT notes to journal - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Add TT notes to journal (if any) - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel Finish all journal entries - [x] kylie - [x] khushi - [x] kevin - [x] hamza - [x] daniel
non_priority
edit journal final journals kylie and khushi kevin and hamza and daniel add and notes to journal kylie khushi kevin hamza daniel add test corrections to journal kylie khushi kevin hamza daniel add tpt notes to journal kylie khushi kevin hamza daniel add tt notes to journal if any kylie khushi kevin hamza daniel finish all journal entries kylie khushi kevin hamza daniel
0
151,617
23,848,813,285
IssuesEvent
2022-09-06 16:00:04
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
No comments pane
*as-designed
Type: <b>Bug</b> I cannot find the [comments pane](https://code.visualstudio.com/updates/v1_71#_comments) anywhere. It doesn't matter which setting I use for it to appear. ![image](https://user-images.githubusercontent.com/623736/188651459-6653dd8e-ec3a-470e-a65f-c87b1bc1d22e.png) VS Code version: Code 1.71.0 (784b0177c56c607789f9638da7b6bf3230d47a8c, 2022-09-01T07:25:10.472Z) OS version: Linux x64 5.15.60-1-MANJARO Modes: Sandboxed: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i7-4790 CPU @ 3.60GHz (8 x 3870)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: disabled_software<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off| |Load (avg)|0, 0, 1| |Memory (System)|15.52GB (11.27GB free)| |Process Argv|/a/bitbean/smg-auto/dealsmgr/api-server| |Screen Reader|no| |VM|0%| |DESKTOP_SESSION|xfce| |XDG_CURRENT_DESKTOP|XFCE| |XDG_SESSION_DESKTOP|xfce| |XDG_SESSION_TYPE|x11| </details><details><summary>Extensions (32)</summary> Extension|Author (truncated)|Version ---|---|--- gitdiffer|aag|0.0.3 Bookmarks|ale|13.3.1 better-toml|bun|0.3.2 dart-code|Dar|3.48.1 flutter|Dar|3.48.0 vscode-eslint|dba|2.2.6 vscode-deno|den|3.13.1 githistory|don|0.6.19 permute-lines|ear|1.1.0 EditorConfig|Edi|0.16.4 prettier-vscode|esb|9.8.0 vscode-solution-explorer|fer|0.7.1 go|gol|0.35.2 css-to-jss|inf|1.0.7 plantuml|jeb|2.17.4 code-eol|jef|1.0.12 docthis|joe|0.7.1 docomment|k--|0.1.31 csharpfixformat|Leo|0.0.84 csharp|ms-|1.25.0 python|ms-|2022.14.0 vscode-pylance|ms-|2022.8.50 jupyter|ms-|2022.8.1002431955 jupyter-keymap|ms-|1.0.0 jupyter-renderers|ms-|1.0.9 remote-containers|ms-|0.251.0 remote-ssh|ms-|0.84.0 remote-ssh-edit|ms-|0.80.0 remote-wsl|ms-|0.66.3 vscode-remote-extensionpack|ms-|0.21.0 gitlink|qez|1.2.4 vscode-yaml|red|1.10.1 </details> <!-- generated by issue reporter -->
1.0
No comments pane - Type: <b>Bug</b> I cannot find the [comments pane](https://code.visualstudio.com/updates/v1_71#_comments) anywhere. It doesn't matter which setting I use for it to appear. ![image](https://user-images.githubusercontent.com/623736/188651459-6653dd8e-ec3a-470e-a65f-c87b1bc1d22e.png) VS Code version: Code 1.71.0 (784b0177c56c607789f9638da7b6bf3230d47a8c, 2022-09-01T07:25:10.472Z) OS version: Linux x64 5.15.60-1-MANJARO Modes: Sandboxed: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i7-4790 CPU @ 3.60GHz (8 x 3870)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: disabled_software<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off| |Load (avg)|0, 0, 1| |Memory (System)|15.52GB (11.27GB free)| |Process Argv|/a/bitbean/smg-auto/dealsmgr/api-server| |Screen Reader|no| |VM|0%| |DESKTOP_SESSION|xfce| |XDG_CURRENT_DESKTOP|XFCE| |XDG_SESSION_DESKTOP|xfce| |XDG_SESSION_TYPE|x11| </details><details><summary>Extensions (32)</summary> Extension|Author (truncated)|Version ---|---|--- gitdiffer|aag|0.0.3 Bookmarks|ale|13.3.1 better-toml|bun|0.3.2 dart-code|Dar|3.48.1 flutter|Dar|3.48.0 vscode-eslint|dba|2.2.6 vscode-deno|den|3.13.1 githistory|don|0.6.19 permute-lines|ear|1.1.0 EditorConfig|Edi|0.16.4 prettier-vscode|esb|9.8.0 vscode-solution-explorer|fer|0.7.1 go|gol|0.35.2 css-to-jss|inf|1.0.7 plantuml|jeb|2.17.4 code-eol|jef|1.0.12 docthis|joe|0.7.1 docomment|k--|0.1.31 csharpfixformat|Leo|0.0.84 csharp|ms-|1.25.0 python|ms-|2022.14.0 vscode-pylance|ms-|2022.8.50 jupyter|ms-|2022.8.1002431955 jupyter-keymap|ms-|1.0.0 jupyter-renderers|ms-|1.0.9 remote-containers|ms-|0.251.0 remote-ssh|ms-|0.84.0 remote-ssh-edit|ms-|0.80.0 remote-wsl|ms-|0.66.3 vscode-remote-extensionpack|ms-|0.21.0 gitlink|qez|1.2.4 vscode-yaml|red|1.10.1 </details> <!-- generated by issue reporter -->
non_priority
no comments pane type bug i cannot find the anywhere it doesn t matter which setting i use for it to appear vs code version code os version linux manjaro modes sandboxed no system info item value cpus intel r core tm cpu x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled multiple raster threads enabled on opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode disabled software video encode disabled software vulkan disabled off webgl enabled enabled webgpu disabled off load avg memory system free process argv a bitbean smg auto dealsmgr api server screen reader no vm desktop session xfce xdg current desktop xfce xdg session desktop xfce xdg session type extensions extension author truncated version gitdiffer aag bookmarks ale better toml bun dart code dar flutter dar vscode eslint dba vscode deno den githistory don permute lines ear editorconfig edi prettier vscode esb vscode solution explorer fer go gol css to jss inf plantuml jeb code eol jef docthis joe docomment k csharpfixformat leo csharp ms python ms vscode pylance ms jupyter ms jupyter keymap ms jupyter renderers ms remote containers ms remote ssh ms remote ssh edit ms remote wsl ms vscode remote extensionpack ms gitlink qez vscode yaml red
0
204,081
15,398,716,972
IssuesEvent
2021-03-04 00:37:16
nucleus-security/Test-repo
https://api.github.com/repos/nucleus-security/Test-repo
opened
Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [Medium] - CentOS Security Update for kernel(CESA-2017:2863)
Testing
Source: QUALYS Finding Description: CentOS has released security update for kernel to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6 Impact: This vulnerability could be exploited to gain complete access to sensitive information. Malicious users could also use this vulnerability to change all the contents or configuration on the system. Additionally this vulnerability can also be used to cause a complete denial of service and could render the resource completely unavailable.</p> Target(s): Asset name: 45.55.254.143 IP: 45.55.254.143 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">centos 6</a> for updates and patch information. <p>Patch:<br /> Following are links for downloading patches to fix the vulnerabilities: </p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">CESA-2017:2863: centos 6</a></p> References: ID:256316 CVE:CVE-2017-7541 Category:CentOS PCI Flagged:1 Vendor References:CESA-2017:2863 centos 6 Bugtraq IDs:99955 Severity: Medium Date Discovered: 2020-01-07 14:35:48 Nucleus Notification Rules Triggered: GitHub Rule Project Name: Ticketing Rules now apply to all vulnerabilities
1.0
Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [Medium] - CentOS Security Update for kernel(CESA-2017:2863) - Source: QUALYS Finding Description: CentOS has released security update for kernel to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6 Impact: This vulnerability could be exploited to gain complete access to sensitive information. Malicious users could also use this vulnerability to change all the contents or configuration on the system. Additionally this vulnerability can also be used to cause a complete denial of service and could render the resource completely unavailable.</p> Target(s): Asset name: 45.55.254.143 IP: 45.55.254.143 Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">centos 6</a> for updates and patch information. <p>Patch:<br /> Following are links for downloading patches to fix the vulnerabilities: </p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">CESA-2017:2863: centos 6</a></p> References: ID:256316 CVE:CVE-2017-7541 Category:CentOS PCI Flagged:1 Vendor References:CESA-2017:2863 centos 6 Bugtraq IDs:99955 Severity: Medium Date Discovered: 2020-01-07 14:35:48 Nucleus Notification Rules Triggered: GitHub Rule Project Name: Ticketing Rules now apply to all vulnerabilities
non_priority
nucleus project ticketing rules now apply to all vulnerabilities centos security update for kernel cesa source qualys finding description centos has released security update for kernel to fix the vulnerabilities affected products centos impact this vulnerability could be exploited to gain complete access to sensitive information malicious users could also use this vulnerability to change all the contents or configuration on the system additionally this vulnerability can also be used to cause a complete denial of service and could render the resource completely unavailable target s asset name ip solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory for updates and patch information patch following are links for downloading patches to fix the vulnerabilities references id cve cve category centos pci flagged vendor references cesa centos bugtraq ids severity medium date discovered nucleus notification rules triggered github rule project name ticketing rules now apply to all vulnerabilities
0
148,354
23,343,120,061
IssuesEvent
2022-08-09 15:30:29
Atri-Labs/atrilabs-engine
https://api.github.com/repos/Atri-Labs/atrilabs-engine
closed
Change UI design of actions bar
design
1. Send data 2. Send file - send own or different component 3. Internal navigation 4. External navigation
1.0
Change UI design of actions bar - 1. Send data 2. Send file - send own or different component 3. Internal navigation 4. External navigation
non_priority
change ui design of actions bar send data send file send own or different component internal navigation external navigation
0
96,668
20,053,529,803
IssuesEvent
2022-02-03 09:36:23
creativecommons/project_creativecommons.org
https://api.github.com/repos/creativecommons/project_creativecommons.org
opened
Determine whether we still (intend to) use Engaging Networks
🟩 priority: low ✨ goal: improvement 💻 aspect: code 🚦 status: awaiting triage
There is a Gravityforms Engaging Networks addon defined as a dependency of our staging website. However, the legacy website does not contain the same plugin. ## Task - [ ] determine whether we still use or intend to use the Engaging Networks service - [ ] decide on whether or not to include the Gravityforms Engaging Networks addon based on our intentions
1.0
Determine whether we still (intend to) use Engaging Networks - There is a Gravityforms Engaging Networks addon defined as a dependency of our staging website. However, the legacy website does not contain the same plugin. ## Task - [ ] determine whether we still use or intend to use the Engaging Networks service - [ ] decide on whether or not to include the Gravityforms Engaging Networks addon based on our intentions
non_priority
determine whether we still intend to use engaging networks there is a gravityforms engaging networks addon defined as a dependency of our staging website however the legacy website does not contain the same plugin task determine whether we still use or intend to use the engaging networks service decide on whether or not to include the gravityforms engaging networks addon based on our intentions
0
157,626
13,697,764,116
IssuesEvent
2020-10-01 04:01:24
SE-Group4/PhotoGallery
https://api.github.com/repos/SE-Group4/PhotoGallery
opened
[1 mark] Create Backlog for All Three Sprints
documentation
[1 mark] Backlog for all three sprints. Please identify team member(s) to whom each backlog item has been assigned to. Identify the Sprint I backlog items that you were not been able to complete by the end of the Sprint I. A backlog is basically a prioritized list of concrete tasks assigned to individual members of the team to help implement the requested features. Each task shouldn’t generally take more than 4-5 hours to complete per one or two team members.
1.0
[1 mark] Create Backlog for All Three Sprints - [1 mark] Backlog for all three sprints. Please identify team member(s) to whom each backlog item has been assigned to. Identify the Sprint I backlog items that you were not been able to complete by the end of the Sprint I. A backlog is basically a prioritized list of concrete tasks assigned to individual members of the team to help implement the requested features. Each task shouldn’t generally take more than 4-5 hours to complete per one or two team members.
non_priority
create backlog for all three sprints backlog for all three sprints please identify team member s to whom each backlog item has been assigned to identify the sprint i backlog items that you were not been able to complete by the end of the sprint i a backlog is basically a prioritized list of concrete tasks assigned to individual members of the team to help implement the requested features each task shouldn’t generally take more than hours to complete per one or two team members
0
343,460
10,330,819,606
IssuesEvent
2019-09-02 15:40:35
bbc/simorgh
https://api.github.com/repos/bbc/simorgh
closed
Add cookie check to mPulse beacon
high priority simorgh-core-stream
**Is your feature request related to a problem? Please describe.** If a user has disabled performance cookies, the mPulse beacon will still be shown, therefore not respecting a user's privacy settings. **Describe the solution you'd like** Check that the user has enabled cookies (exact perferences/type of cookie to check TBD) before rendering the mPulse beacon. Check the Dynatrace Beacon in RN to see which cookie was required to be set before loading the beacon. **Describe alternatives you've considered** N/A **Testing notes** [Tester to complete] Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc **Additional context** Add any other context or screenshots about the feature request here. - [x] Initially labelled with ["Refinement needed"](https://github.com/bbc/simorgh/labels/Refinement%20Needed)
1.0
Add cookie check to mPulse beacon - **Is your feature request related to a problem? Please describe.** If a user has disabled performance cookies, the mPulse beacon will still be shown, therefore not respecting a user's privacy settings. **Describe the solution you'd like** Check that the user has enabled cookies (exact perferences/type of cookie to check TBD) before rendering the mPulse beacon. Check the Dynatrace Beacon in RN to see which cookie was required to be set before loading the beacon. **Describe alternatives you've considered** N/A **Testing notes** [Tester to complete] Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc **Additional context** Add any other context or screenshots about the feature request here. - [x] Initially labelled with ["Refinement needed"](https://github.com/bbc/simorgh/labels/Refinement%20Needed)
priority
add cookie check to mpulse beacon is your feature request related to a problem please describe if a user has disabled performance cookies the mpulse beacon will still be shown therefore not respecting a user s privacy settings describe the solution you d like check that the user has enabled cookies exact perferences type of cookie to check tbd before rendering the mpulse beacon check the dynatrace beacon in rn to see which cookie was required to be set before loading the beacon describe alternatives you ve considered n a testing notes dev insight will cypress tests be required or are unit tests sufficient will there be any potential regression etc additional context add any other context or screenshots about the feature request here initially labelled with
1
324,570
9,905,480,169
IssuesEvent
2019-06-27 11:42:33
kudobuilder/kudo
https://api.github.com/repos/kudobuilder/kudo
closed
CreateOrUpdate function fix
component/operator kind/bug priority/high
In the plan controller, the line: ``` result, err := controllerutil.CreateOrUpdate(context.TODO(), r.Client, obj, func(runtime.Object) error { return nil }) ``` needs to be fixed. The last argument of this function is supposed to capture the modifications to the object pulled from the server. Need to replace it with something like this from instance_controller.go ``` did, err := controllerutil.CreateOrUpdate(context.TODO(), mgr.GetClient(), current, func(o runtime.Object) error { t := true o.(*maestrov1alpha1.PlanExecution).Spec.Suspend = &t return nil }) ```
1.0
CreateOrUpdate function fix - In the plan controller, the line: ``` result, err := controllerutil.CreateOrUpdate(context.TODO(), r.Client, obj, func(runtime.Object) error { return nil }) ``` needs to be fixed. The last argument of this function is supposed to capture the modifications to the object pulled from the server. Need to replace it with something like this from instance_controller.go ``` did, err := controllerutil.CreateOrUpdate(context.TODO(), mgr.GetClient(), current, func(o runtime.Object) error { t := true o.(*maestrov1alpha1.PlanExecution).Spec.Suspend = &t return nil }) ```
priority
createorupdate function fix in the plan controller the line result err controllerutil createorupdate context todo r client obj func runtime object error return nil needs to be fixed the last argument of this function is supposed to capture the modifications to the object pulled from the server need to replace it with something like this from instance controller go did err controllerutil createorupdate context todo mgr getclient current func o runtime object error t true o planexecution spec suspend t return nil
1
499,526
14,449,290,741
IssuesEvent
2020-12-08 07:51:35
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Incorrect stability check
platform:azure priority-3-normal reproduction needed type:bug
**What Renovate type, platform and version are you using?** Self Hosted, Azure DevOps **Describe the bug** Renovate doesn't create PRs even if the update met stability days requirements. **Relevant debug logs** ``` DEBUG: Updated 2 package files (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: No updated lock files in branch (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: 2 file(s) to commit (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: Committing files to branch renovate/microsoft-applicationinsights-packages (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) INFO: Branch created (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) "commitSha": "067fd1f" DEBUG: Updating renovate/stability-days status check state to green (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: setBranchStatus(renovate/microsoft-applicationinsights-packages, renovate/stability-days, Updates have met stability days requirement, green, https://docs.renovatebot.com/) (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: Branch status pending (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) "commitSha": "067fd1f" ``` **To Reproduce** 1. Set stabilityDays: X 2. Set prCreation: "not-pending" 3. Add dependency that pass stabilityDays period. 4. Run renovate **Additional context** I think it's somehow related to this bug #4884
1.0
Incorrect stability check - **What Renovate type, platform and version are you using?** Self Hosted, Azure DevOps **Describe the bug** Renovate doesn't create PRs even if the update met stability days requirements. **Relevant debug logs** ``` DEBUG: Updated 2 package files (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: No updated lock files in branch (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: 2 file(s) to commit (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: Committing files to branch renovate/microsoft-applicationinsights-packages (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) INFO: Branch created (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) "commitSha": "067fd1f" DEBUG: Updating renovate/stability-days status check state to green (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: setBranchStatus(renovate/microsoft-applicationinsights-packages, renovate/stability-days, Updates have met stability days requirement, green, https://docs.renovatebot.com/) (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) DEBUG: Branch status pending (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages) "commitSha": "067fd1f" ``` **To Reproduce** 1. Set stabilityDays: X 2. Set prCreation: "not-pending" 3. Add dependency that pass stabilityDays period. 4. Run renovate **Additional context** I think it's somehow related to this bug #4884
priority
incorrect stability check what renovate type platform and version are you using self hosted azure devops describe the bug renovate doesn t create prs even if the update met stability days requirements relevant debug logs debug updated package files repository mycompany projectx branch renovate microsoft applicationinsights packages debug no updated lock files in branch repository mycompany projectx branch renovate microsoft applicationinsights packages debug file s to commit repository mycompany projectx branch renovate microsoft applicationinsights packages debug committing files to branch renovate microsoft applicationinsights packages repository mycompany projectx branch renovate microsoft applicationinsights packages info branch created repository mycompany projectx branch renovate microsoft applicationinsights packages commitsha debug updating renovate stability days status check state to green repository mycompany projectx branch renovate microsoft applicationinsights packages debug setbranchstatus renovate microsoft applicationinsights packages renovate stability days updates have met stability days requirement green repository mycompany projectx branch renovate microsoft applicationinsights packages debug branch status pending repository mycompany projectx branch renovate microsoft applicationinsights packages commitsha to reproduce set stabilitydays x set prcreation not pending add dependency that pass stabilitydays period run renovate additional context i think it s somehow related to this bug
1
214,558
7,274,378,728
IssuesEvent
2018-02-21 09:48:24
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
clientauth.jwt jar version mentioned in the doc is different from the actual jar version created
Affected/5.5.0-Alpha Priority/High Type/Docs
clientauth.jwt jar version mentioned in the doc [1] is different from the actual jar version created. [1] https://docs.wso2.com/display/IS550/Private+Key+JWT+Client+Authentication+for+OIDC It says, Place the target/org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.2-SNAPSHOT.jar in the <IS_HOME>/repository/component/dropins directory. But the actual jar created is as, org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.3-SNAPSHOT.jar
1.0
clientauth.jwt jar version mentioned in the doc is different from the actual jar version created - clientauth.jwt jar version mentioned in the doc [1] is different from the actual jar version created. [1] https://docs.wso2.com/display/IS550/Private+Key+JWT+Client+Authentication+for+OIDC It says, Place the target/org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.2-SNAPSHOT.jar in the <IS_HOME>/repository/component/dropins directory. But the actual jar created is as, org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.3-SNAPSHOT.jar
priority
clientauth jwt jar version mentioned in the doc is different from the actual jar version created clientauth jwt jar version mentioned in the doc is different from the actual jar version created it says place the target org carbon identity token handler clientauth jwt snapshot jar in the repository component dropins directory but the actual jar created is as org carbon identity token handler clientauth jwt snapshot jar
1
372,812
11,028,587,565
IssuesEvent
2019-12-06 12:03:10
coder3101/cp-editor2
https://api.github.com/repos/coder3101/cp-editor2
closed
Showing compiler warnings
enhancement high_priority linux macOs windows
**Is your feature request related to a problem? Please describe.** I can't see the compiler warnings. **Describe the solution you'd like** Show compiler warnings after compiling. Maybe add a setting of whether to show the warnings or not. (However, this can be done by adding `-w` in the compile command.) **Describe alternatives you've considered** N/A **Additional context** N/A
1.0
Showing compiler warnings - **Is your feature request related to a problem? Please describe.** I can't see the compiler warnings. **Describe the solution you'd like** Show compiler warnings after compiling. Maybe add a setting of whether to show the warnings or not. (However, this can be done by adding `-w` in the compile command.) **Describe alternatives you've considered** N/A **Additional context** N/A
priority
showing compiler warnings is your feature request related to a problem please describe i can t see the compiler warnings describe the solution you d like show compiler warnings after compiling maybe add a setting of whether to show the warnings or not however this can be done by adding w in the compile command describe alternatives you ve considered n a additional context n a
1
303,193
22,958,859,177
IssuesEvent
2022-07-19 13:52:23
fga-eps-mds/UnbFlow
https://api.github.com/repos/fga-eps-mds/UnbFlow
closed
Docs: Escrever historias de usuário
documentation
# Descrição Criar as histórias de usuário com base nos epicos, features e no prototipo de média fidelidade ## Critério de aceitação - [x] Escrever histórias de usuário que serão realizadas durante as issues
1.0
Docs: Escrever historias de usuário - # Descrição Criar as histórias de usuário com base nos epicos, features e no prototipo de média fidelidade ## Critério de aceitação - [x] Escrever histórias de usuário que serão realizadas durante as issues
non_priority
docs escrever historias de usuário descrição criar as histórias de usuário com base nos epicos features e no prototipo de média fidelidade critério de aceitação escrever histórias de usuário que serão realizadas durante as issues
0
186,220
6,734,466,543
IssuesEvent
2017-10-18 18:08:56
octobercms/october
https://api.github.com/repos/octobercms/october
closed
Using Artisan::add in init.php does not work
Priority: Medium Status: Review Needed Type: Unconfirmed Bug
https://octobercms.com/docs/console/development states that you should be able to write `Artisan::add(new Acme\Blog\Console\MyCommand);` in `init.php` and the command will be added Instead I get an exception ``` [Symfony\Component\Debug\Exception\FatalErrorException] Call to undefined method October\Rain\Foundation\Console\Kernel::add() ```
1.0
Using Artisan::add in init.php does not work - https://octobercms.com/docs/console/development states that you should be able to write `Artisan::add(new Acme\Blog\Console\MyCommand);` in `init.php` and the command will be added Instead I get an exception ``` [Symfony\Component\Debug\Exception\FatalErrorException] Call to undefined method October\Rain\Foundation\Console\Kernel::add() ```
priority
using artisan add in init php does not work states that you should be able to write artisan add new acme blog console mycommand in init php and the command will be added instead i get an exception call to undefined method october rain foundation console kernel add
1
18,777
24,678,890,034
IssuesEvent
2022-10-18 19:25:58
dtcenter/MET
https://api.github.com/repos/dtcenter/MET
opened
Investigate `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml
type: bug alert: NEED ACCOUNT KEY requestor: METplus Team MET: PreProcessing Tools (Point) priority: high
## Describe the Problem ## During review of #2294 for issue #2276, a problem was discovered in the output of the `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml. The output file created by this test (HourlyData_20220312.nc) contains values of Infinity (`Inf`). While the GHA run for that PR did increase the occurrence of Inf in the output, the problem existed prior to those code changes. This issue is to investigate the source of the `Inf` values appearing in the output, and fix the code to avoid them. ### Expected Behavior ### The output of ascii2nc should never contain a value of infinity. The code should be enhanced by adding more error checking to avoid them. Perhaps, they should be reported as bad data value (i.e. -9999) rather than `Inf`? ### Environment ### Describe your runtime environment: *1. Visible in the output of GHA and in the output of the MET nightly build on seneca.* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Log on to 'seneca'* *2. Go to NB area:* ``` cd /d1/projects/MET/MET_regression/develop/NB20221018 ``` *3. Dump to ascii: ``` Rscript MET-develop/scripts/Rscripts/pntnc2ascii.R MET-develop/test_output/ascii2nc/airnow/HourlyData_20220312.nc > HourlyData_20220312.txt ``` *4. See error in columns 6 and 9 of the output:* ``` grep Inf HourlyData_20220312.txt | wc -l 33 ``` *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Organization** level **Project** for support of the current coordinated release - [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
1.0
Investigate `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml - ## Describe the Problem ## During review of #2294 for issue #2276, a problem was discovered in the output of the `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml. The output file created by this test (HourlyData_20220312.nc) contains values of Infinity (`Inf`). While the GHA run for that PR did increase the occurrence of Inf in the output, the problem existed prior to those code changes. This issue is to investigate the source of the `Inf` values appearing in the output, and fix the code to avoid them. ### Expected Behavior ### The output of ascii2nc should never contain a value of infinity. The code should be enhanced by adding more error checking to avoid them. Perhaps, they should be reported as bad data value (i.e. -9999) rather than `Inf`? ### Environment ### Describe your runtime environment: *1. Visible in the output of GHA and in the output of the MET nightly build on seneca.* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Log on to 'seneca'* *2. Go to NB area:* ``` cd /d1/projects/MET/MET_regression/develop/NB20221018 ``` *3. Dump to ascii: ``` Rscript MET-develop/scripts/Rscripts/pntnc2ascii.R MET-develop/test_output/ascii2nc/airnow/HourlyData_20220312.nc > HourlyData_20220312.txt ``` *4. See error in columns 6 and 9 of the output:* ``` grep Inf HourlyData_20220312.txt | wc -l 33 ``` *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Organization** level **Project** for support of the current coordinated release - [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
non_priority
investigate airnow hourly test in unit xml describe the problem during review of for issue a problem was discovered in the output of the airnow hourly test in unit xml the output file created by this test hourlydata nc contains values of infinity inf while the gha run for that pr did increase the occurrence of inf in the output the problem existed prior to those code changes this issue is to investigate the source of the inf values appearing in the output and fix the code to avoid them expected behavior the output of should never contain a value of infinity the code should be enhanced by adding more error checking to avoid them perhaps they should be reported as bad data value i e rather than inf environment describe your runtime environment visible in the output of gha and in the output of the met nightly build on seneca to reproduce describe the steps to reproduce the behavior log on to seneca go to nb area cd projects met met regression develop dump to ascii rscript met develop scripts rscripts r met develop test output airnow hourlydata nc hourlydata txt see error in columns and of the output grep inf hourlydata txt wc l post relevant sample data following these instructions relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and linked issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version close this issue
0
566,801
16,830,786,287
IssuesEvent
2021-06-18 04:14:50
CertifaiAI/classifai
https://api.github.com/repos/CertifaiAI/classifai
opened
Backend language support
bug low priority
**Describe the bug** As ClassifAI is supporting multiple languages, The title of UI prompted from backend is still in english. **Expected behavior** The language should be dynamic in backend as well **Screenshots** ![image](https://user-images.githubusercontent.com/76682536/122505277-9746ba80-d02e-11eb-9dca-f4fab40b8d7f.png) **Desktop (please complete the following information):** - OS: Windows - Browser chrome - Version v2.0.0 alpha1
1.0
Backend language support - **Describe the bug** As ClassifAI is supporting multiple languages, The title of UI prompted from backend is still in english. **Expected behavior** The language should be dynamic in backend as well **Screenshots** ![image](https://user-images.githubusercontent.com/76682536/122505277-9746ba80-d02e-11eb-9dca-f4fab40b8d7f.png) **Desktop (please complete the following information):** - OS: Windows - Browser chrome - Version v2.0.0 alpha1
priority
backend language support describe the bug as classifai is supporting multiple languages the title of ui prompted from backend is still in english expected behavior the language should be dynamic in backend as well screenshots desktop please complete the following information os windows browser chrome version
1
288,020
24,882,269,162
IssuesEvent
2022-10-28 03:01:02
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Orçamento - Legislação - Lagamar
generalization test development template - Memory (66) tag - Orçamento subtag - Legislação
DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Lagamar.
1.0
Teste de generalizacao para a tag Orçamento - Legislação - Lagamar - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Lagamar.
non_priority
teste de generalizacao para a tag orçamento legislação lagamar dod realizar o teste de generalização do validador da tag orçamento legislação para o município de lagamar
0
622,689
19,654,052,470
IssuesEvent
2022-01-10 10:34:47
pgfmc/Core
https://api.github.com/repos/pgfmc/Core
closed
Fix when AFK deactivates
bug low priority
What needs to change: - AFK toggle on teleport - Don't AFK toggle on Y level descend
1.0
Fix when AFK deactivates - What needs to change: - AFK toggle on teleport - Don't AFK toggle on Y level descend
priority
fix when afk deactivates what needs to change afk toggle on teleport don t afk toggle on y level descend
1
461,080
13,223,211,077
IssuesEvent
2020-08-17 16:49:19
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Apps can't be installed in latest master from Dashboard
[zube]: Next Up alpha-priority/0 kind/bug-qa
**What kind of request is this:** **Steps to reproduce:** - Install Rancher HA ( k8s `master-head (08/10/2020)` _502b1839e_) - Navigate to the cluster Dashboard and Apps **Result:** ![Screen Shot 2020-08-10 at 9 52 27 AM](https://user-images.githubusercontent.com/27659/89809384-54ec4d00-daf0-11ea-987e-f7fd92d93bbb.png) ![Screen Shot 2020-08-10 at 9 53 41 AM](https://user-images.githubusercontent.com/27659/89809399-5c135b00-daf0-11ea-8a9b-fead5501c7ca.png) **Other details that may be helpful:** **Environment information** - Rancher version: `master-head (08/10/2020)` _502b1839e_ - Installation option: HA k8s
1.0
Apps can't be installed in latest master from Dashboard - **What kind of request is this:** **Steps to reproduce:** - Install Rancher HA ( k8s `master-head (08/10/2020)` _502b1839e_) - Navigate to the cluster Dashboard and Apps **Result:** ![Screen Shot 2020-08-10 at 9 52 27 AM](https://user-images.githubusercontent.com/27659/89809384-54ec4d00-daf0-11ea-987e-f7fd92d93bbb.png) ![Screen Shot 2020-08-10 at 9 53 41 AM](https://user-images.githubusercontent.com/27659/89809399-5c135b00-daf0-11ea-8a9b-fead5501c7ca.png) **Other details that may be helpful:** **Environment information** - Rancher version: `master-head (08/10/2020)` _502b1839e_ - Installation option: HA k8s
priority
apps can t be installed in latest master from dashboard what kind of request is this steps to reproduce install rancher ha master head navigate to the cluster dashboard and apps result other details that may be helpful environment information rancher version master head installation option ha
1
95,033
3,933,560,384
IssuesEvent
2016-04-25 19:33:40
ghutchis/avogadro
https://api.github.com/repos/ghutchis/avogadro
closed
Avogadro crashes when exporting bitmap graphics on Mac.
auto-migrated high priority v_1.1.0
Once correctly opened and visualized a .log or a .cube g09 file, I cannot export the visualized image using File->Export->Graphics... because the program suddenly crashes. Only the vectorial image can be exported, but the color are strongly altered. Reported by: *anonymous
1.0
Avogadro crashes when exporting bitmap graphics on Mac. - Once correctly opened and visualized a .log or a .cube g09 file, I cannot export the visualized image using File->Export->Graphics... because the program suddenly crashes. Only the vectorial image can be exported, but the color are strongly altered. Reported by: *anonymous
priority
avogadro crashes when exporting bitmap graphics on mac once correctly opened and visualized a log or a cube file i cannot export the visualized image using file export graphics because the program suddenly crashes only the vectorial image can be exported but the color are strongly altered reported by anonymous
1
382,846
11,339,118,576
IssuesEvent
2020-01-23 00:43:55
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
record_function decorator in autograd profiler does not work with RPC
high priority module: rpc
## 🐛 Bug Nested `RecordFunction`s during profiling do not work with RPC calls, due to mismanagement of some internal variables in `record_function.cpp` caused by handling `RecordFunction` objects completing in different hreads. ## To Reproduce ``` with torch.autograd.profiler.profiler() as prof: with torch.autograd.profiler.record_function("foo"): rpc.rpc_async(...) ``` will result in an error: ``` > I0122 16:32:15.967217 594065 record_function.cpp:214] Exception in RecordFunction::end(): (thread_local_func_ == thi s) || (thread_local_func_ == nullptr && threadId_ != 0) INTERNAL ASSERT FAILED at caffe2/torch/csrc/autograd/record_fu nction.cpp:235, please report a bug to PyTorch. baze: must be top of stack. If you are calling RecordFunction::end in aseparate thread, call RecordFunction::setThreadId() in the creatingthread. (end at caffe2/torch/csrc/autograd/record_ function.cpp:235) ``` This is because the handling of the thread local variables `thread_local_func` and `parent_` are not managed properly when `RecordFunction::end` is called from a separate thread. Expected behavior: The nested scopes should be profiled correctly as non-RPC code would be.
1.0
record_function decorator in autograd profiler does not work with RPC - ## 🐛 Bug Nested `RecordFunction`s during profiling do not work with RPC calls, due to mismanagement of some internal variables in `record_function.cpp` caused by handling `RecordFunction` objects completing in different hreads. ## To Reproduce ``` with torch.autograd.profiler.profiler() as prof: with torch.autograd.profiler.record_function("foo"): rpc.rpc_async(...) ``` will result in an error: ``` > I0122 16:32:15.967217 594065 record_function.cpp:214] Exception in RecordFunction::end(): (thread_local_func_ == thi s) || (thread_local_func_ == nullptr && threadId_ != 0) INTERNAL ASSERT FAILED at caffe2/torch/csrc/autograd/record_fu nction.cpp:235, please report a bug to PyTorch. baze: must be top of stack. If you are calling RecordFunction::end in aseparate thread, call RecordFunction::setThreadId() in the creatingthread. (end at caffe2/torch/csrc/autograd/record_ function.cpp:235) ``` This is because the handling of the thread local variables `thread_local_func` and `parent_` are not managed properly when `RecordFunction::end` is called from a separate thread. Expected behavior: The nested scopes should be profiled correctly as non-RPC code would be.
priority
record function decorator in autograd profiler does not work with rpc 🐛 bug nested recordfunction s during profiling do not work with rpc calls due to mismanagement of some internal variables in record function cpp caused by handling recordfunction objects completing in different hreads to reproduce with torch autograd profiler profiler as prof with torch autograd profiler record function foo rpc rpc async will result in an error record function cpp exception in recordfunction end thread local func thi s thread local func nullptr threadid internal assert failed at torch csrc autograd record fu nction cpp please report a bug to pytorch baze must be top of stack if you are calling recordfunction end in aseparate thread call recordfunction setthreadid in the creatingthread end at torch csrc autograd record function cpp this is because the handling of the thread local variables thread local func and parent are not managed properly when recordfunction end is called from a separate thread expected behavior the nested scopes should be profiled correctly as non rpc code would be
1
15,368
3,461,414,456
IssuesEvent
2015-12-20 02:02:27
ehmorris/Facebook-Mood
https://api.github.com/repos/ehmorris/Facebook-Mood
closed
Extension tests
extension-related testing
This is actually going to be kinda cool. Writing a testing module for javascript as a chrome extension.
1.0
Extension tests - This is actually going to be kinda cool. Writing a testing module for javascript as a chrome extension.
non_priority
extension tests this is actually going to be kinda cool writing a testing module for javascript as a chrome extension
0
187,005
21,993,039,292
IssuesEvent
2022-05-26 01:22:44
raindigi/GraphqlType-API-Registration
https://api.github.com/repos/raindigi/GraphqlType-API-Registration
opened
CVE-2022-24434 (High) detected in dicer-0.3.0.tgz
security vulnerability
## CVE-2022-24434 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dicer-0.3.0.tgz</b></p></summary> <p>A very fast streaming multipart parser for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz">https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz</a></p> <p>Path to dependency file: /GraphqlType-API-Registration/package.json</p> <p>Path to vulnerable library: /node_modules/dicer/package.json</p> <p> Dependency Hierarchy: - apollo-server-express-2.3.3.tgz (Root Library) - apollo-server-core-2.3.3.tgz - graphql-upload-8.0.4.tgz - busboy-0.3.0.tgz - :x: **dicer-0.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/raindigi/GraphqlType-API-Registration/commit/ddbf6735cb305c07ce82a0faefa069bd89b7581f">ddbf6735cb305c07ce82a0faefa069bd89b7581f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects all versions of package dicer. A malicious attacker can send a modified form to server, and crash the nodejs service. An attacker could sent the payload again and again so that the service continuously crashes. <p>Publish Date: 2022-05-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24434>CVE-2022-24434</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-24434">https://nvd.nist.gov/vuln/detail/CVE-2022-24434</a></p> <p>Release Date: 2022-05-20</p> <p>Fix Resolution: no_fix</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24434 (High) detected in dicer-0.3.0.tgz - ## CVE-2022-24434 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dicer-0.3.0.tgz</b></p></summary> <p>A very fast streaming multipart parser for node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz">https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz</a></p> <p>Path to dependency file: /GraphqlType-API-Registration/package.json</p> <p>Path to vulnerable library: /node_modules/dicer/package.json</p> <p> Dependency Hierarchy: - apollo-server-express-2.3.3.tgz (Root Library) - apollo-server-core-2.3.3.tgz - graphql-upload-8.0.4.tgz - busboy-0.3.0.tgz - :x: **dicer-0.3.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/raindigi/GraphqlType-API-Registration/commit/ddbf6735cb305c07ce82a0faefa069bd89b7581f">ddbf6735cb305c07ce82a0faefa069bd89b7581f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects all versions of package dicer. A malicious attacker can send a modified form to server, and crash the nodejs service. An attacker could sent the payload again and again so that the service continuously crashes. <p>Publish Date: 2022-05-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24434>CVE-2022-24434</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-24434">https://nvd.nist.gov/vuln/detail/CVE-2022-24434</a></p> <p>Release Date: 2022-05-20</p> <p>Fix Resolution: no_fix</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in dicer tgz cve high severity vulnerability vulnerable library dicer tgz a very fast streaming multipart parser for node js library home page a href path to dependency file graphqltype api registration package json path to vulnerable library node modules dicer package json dependency hierarchy apollo server express tgz root library apollo server core tgz graphql upload tgz busboy tgz x dicer tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects all versions of package dicer a malicious attacker can send a modified form to server and crash the nodejs service an attacker could sent the payload again and again so that the service continuously crashes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution no fix step up your open source security game with whitesource
0
50,519
13,187,554,369
IssuesEvent
2020-08-13 03:47:37
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
test ticket (Trac #869)
Migrated from Trac cmake defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/869 , reported by nega and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2015-02-12T06:31:40", "description": "", "reporter": "nega", "cc": "", "resolution": "invalid", "_ts": "1423722700498868", "component": "cmake", "summary": "test ticket", "priority": "normal", "keywords": "", "time": "2015-02-11T23:11:24", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
test ticket (Trac #869) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/869 , reported by nega and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2015-02-12T06:31:40", "description": "", "reporter": "nega", "cc": "", "resolution": "invalid", "_ts": "1423722700498868", "component": "cmake", "summary": "test ticket", "priority": "normal", "keywords": "", "time": "2015-02-11T23:11:24", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
non_priority
test ticket trac migrated from reported by nega and owned by nega json status closed changetime description reporter nega cc resolution invalid ts component cmake summary test ticket priority normal keywords time milestone owner nega type defect
0
810,198
30,229,778,490
IssuesEvent
2023-07-06 05:44:42
glific/mobile
https://api.github.com/repos/glific/mobile
closed
Start a flow for a collection
Priority: High
**Start a flow for a collection** Implement 'Start a flow' in Collection Chat Screen Option. When selected from options, a dialog box will appear, providing users with the ability to choose a flow and initiate it using the "Start" button or cancel the action with the "Cancel" button. **Approach** - Implement dialog box component according to design. - Implement a query to fetch all the available flows from the backend. - Implement logic to start the chosen flow when the "Start" button is pressed. Figma link: https://www.figma.com/file/SbP0ru8kiB2l334ieX1ESe/Mobile-app?type=design&node-id=754-5118&mode=design&t=DfFgoqouNYlgv68A-0 Get all flows API: https://api.glific.com/#e0f1066e-471b-47de-b83d-0003387e9032 ```sh variables = { filter: { status: "published", isActive: true, }, opts: { limit: null, offset: 0, order: "ASC", } } ``` Start a flow API: https://api.glific.com/#cb956d81-9ce5-4101-ad5f-2c65402c4171 ```sh variables = { flowId: "2054", groupId: "2", } ```
1.0
Start a flow for a collection - **Start a flow for a collection** Implement 'Start a flow' in Collection Chat Screen Option. When selected from options, a dialog box will appear, providing users with the ability to choose a flow and initiate it using the "Start" button or cancel the action with the "Cancel" button. **Approach** - Implement dialog box component according to design. - Implement a query to fetch all the available flows from the backend. - Implement logic to start the chosen flow when the "Start" button is pressed. Figma link: https://www.figma.com/file/SbP0ru8kiB2l334ieX1ESe/Mobile-app?type=design&node-id=754-5118&mode=design&t=DfFgoqouNYlgv68A-0 Get all flows API: https://api.glific.com/#e0f1066e-471b-47de-b83d-0003387e9032 ```sh variables = { filter: { status: "published", isActive: true, }, opts: { limit: null, offset: 0, order: "ASC", } } ``` Start a flow API: https://api.glific.com/#cb956d81-9ce5-4101-ad5f-2c65402c4171 ```sh variables = { flowId: "2054", groupId: "2", } ```
priority
start a flow for a collection start a flow for a collection implement start a flow in collection chat screen option when selected from options a dialog box will appear providing users with the ability to choose a flow and initiate it using the start button or cancel the action with the cancel button approach implement dialog box component according to design implement a query to fetch all the available flows from the backend implement logic to start the chosen flow when the start button is pressed figma link get all flows api sh variables filter status published isactive true opts limit null offset order asc start a flow api sh variables flowid groupid
1
665,016
22,296,070,020
IssuesEvent
2022-06-13 01:49:26
ESIPFed/Geoweaver
https://api.github.com/repos/ESIPFed/Geoweaver
closed
Improve the weaver GUI
enhancement high-priority
Some existing problems: - The words inside the circles are hardly readable - The process names often go out of bounds - The status indicators are not easy to understand the color change - The toolbar is not self explainable - The process status bars are not indicating anything - ...
1.0
Improve the weaver GUI - Some existing problems: - The words inside the circles are hardly readable - The process names often go out of bounds - The status indicators are not easy to understand the color change - The toolbar is not self explainable - The process status bars are not indicating anything - ...
priority
improve the weaver gui some existing problems the words inside the circles are hardly readable the process names often go out of bounds the status indicators are not easy to understand the color change the toolbar is not self explainable the process status bars are not indicating anything
1
273,400
8,530,235,069
IssuesEvent
2018-11-03 20:20:27
projectcalico/calico
https://api.github.com/repos/projectcalico/calico
closed
Document need for `iptables -P FORWARD ACCEPT` for IPVS until K8s fixes are merged
area/docs/content content/needed priority/P3
Currently, when using kube-proxy in IPVS mode, for NodePorts to work (when workloads are on a host other than the host initially being connected to) it is necessary to run `iptables -P FORWARD ACCEPT` on all nodes. The issue is raised upstream https://github.com/kubernetes/kubernetes/issues/59656 and the fix is up as this PR https://github.com/kubernetes/kubernetes/pull/62007. ## Expected Behavior Should document the need for the above at least until the fix is in a K8s release. ## Current Behavior Not documented. /cc @song-jiang
1.0
Document need for `iptables -P FORWARD ACCEPT` for IPVS until K8s fixes are merged - Currently, when using kube-proxy in IPVS mode, for NodePorts to work (when workloads are on a host other than the host initially being connected to) it is necessary to run `iptables -P FORWARD ACCEPT` on all nodes. The issue is raised upstream https://github.com/kubernetes/kubernetes/issues/59656 and the fix is up as this PR https://github.com/kubernetes/kubernetes/pull/62007. ## Expected Behavior Should document the need for the above at least until the fix is in a K8s release. ## Current Behavior Not documented. /cc @song-jiang
priority
document need for iptables p forward accept for ipvs until fixes are merged currently when using kube proxy in ipvs mode for nodeports to work when workloads are on a host other than the host initially being connected to it is necessary to run iptables p forward accept on all nodes the issue is raised upstream and the fix is up as this pr expected behavior should document the need for the above at least until the fix is in a release current behavior not documented cc song jiang
1
702,123
24,120,640,480
IssuesEvent
2022-09-20 18:22:51
googleapis/nodejs-analytics-data
https://api.github.com/repos/googleapis/nodejs-analytics-data
closed
Realtime report with multiple dimensions: should run realtime with multiple dimensions failed
type: bug priority: p1 api: analyticsdata flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: ae646a54bc64abff0cf92625117ffb258e303e8b buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c), [Sponge](http://sponge2/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/runRealtimeReportWithMultipleDimensions.test.js:24:28) at Context.<anonymous> (test/runRealtimeReportWithMultipleDimensions.test.js:31:20) at processImmediate (internal/timers.js:461:21)</pre></details>
1.0
Realtime report with multiple dimensions: should run realtime with multiple dimensions failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: ae646a54bc64abff0cf92625117ffb258e303e8b buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c), [Sponge](http://sponge2/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c) status: failed <details><summary>Test output</summary><br><pre>Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. Error: Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558 16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. at checkExecSyncError (child_process.js:635:11) at Object.execSync (child_process.js:671:15) at execSync (test/runRealtimeReportWithMultipleDimensions.test.js:24:28) at Context.<anonymous> (test/runRealtimeReportWithMultipleDimensions.test.js:31:20) at processImmediate (internal/timers.js:461:21)</pre></details>
priority
realtime report with multiple dimensions should run realtime with multiple dimensions failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node runrealtimereportwithmultipledimensions js unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error command failed node runrealtimereportwithmultipledimensions js unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at checkexecsyncerror child process js at object execsync child process js at execsync test runrealtimereportwithmultipledimensions test js at context test runrealtimereportwithmultipledimensions test js at processimmediate internal timers js
1
121,118
15,857,163,455
IssuesEvent
2021-04-08 04:07:14
asoffer/Icarus
https://api.github.com/repos/asoffer/Icarus
closed
Should array types allow implicit integer casts?
design decision
Currently array lengths need to be integral but will be implicitly cast to the underlying array-type's length implementation type. This is currently a `uint64_t` so any natively supported type in Icarus can convert to this losslessly (because negative lengths are prohibited). However, if Icarus starts supporting larger integer types this cast will be lossy and implicit. This issue tracks simply making a decision here. There is a nice consistency in always requiring the cast. There is also no reason to support arrays of length 2^64 or greater in any reasonable future.
1.0
Should array types allow implicit integer casts? - Currently array lengths need to be integral but will be implicitly cast to the underlying array-type's length implementation type. This is currently a `uint64_t` so any natively supported type in Icarus can convert to this losslessly (because negative lengths are prohibited). However, if Icarus starts supporting larger integer types this cast will be lossy and implicit. This issue tracks simply making a decision here. There is a nice consistency in always requiring the cast. There is also no reason to support arrays of length 2^64 or greater in any reasonable future.
non_priority
should array types allow implicit integer casts currently array lengths need to be integral but will be implicitly cast to the underlying array type s length implementation type this is currently a t so any natively supported type in icarus can convert to this losslessly because negative lengths are prohibited however if icarus starts supporting larger integer types this cast will be lossy and implicit this issue tracks simply making a decision here there is a nice consistency in always requiring the cast there is also no reason to support arrays of length or greater in any reasonable future
0
140,427
5,408,859,972
IssuesEvent
2017-03-01 01:33:53
google/roboto
https://api.github.com/repos/google/roboto
opened
Roboto Mono 'l' legibility
Priority-High
Requested change: > Change the lowercase L in Roboto Mono to a reverse S shape, like Andale Mono or Monaco have; Roboto Mono has clearer distinction between lowercase L and 1 than Courier does, but the reverse-S type of L is clearer still.
1.0
Roboto Mono 'l' legibility - Requested change: > Change the lowercase L in Roboto Mono to a reverse S shape, like Andale Mono or Monaco have; Roboto Mono has clearer distinction between lowercase L and 1 than Courier does, but the reverse-S type of L is clearer still.
priority
roboto mono l legibility requested change change the lowercase l in roboto mono to a reverse s shape like andale mono or monaco have roboto mono has clearer distinction between lowercase l and than courier does but the reverse s type of l is clearer still
1
817,370
30,639,264,360
IssuesEvent
2023-07-24 20:24:25
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
speech.snippets.transcribe_gcs_v2_test: test_transcribe_gcs_v2 failed
priority: p2 type: bug api: speech samples flakybot: issue flakybot: flaky
Note: #9423 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 16ab3cf93caddb092710ba507d537945128c884c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/297c18a8-be85-43ab-92b9-655f63a7de99), [Sponge](http://sponge2/297c18a8-be85-43ab-92b9-655f63a7de99) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "502:Bad Gateway" debug_error_string = "UNKNOWN:Error received from peer {grpc_message:"502:Bad Gateway", grpc_status:14, created_time:"2023-06-17T13:39:31.187553795+00:00"}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 191, in retry_target return target() File "/workspace/speech/snippets/transcribe_gcs_v2_test.py", line 38, in test_transcribe_gcs_v2 response = transcribe_gcs_v2.transcribe_gcs_v2( File "/workspace/speech/snippets/transcribe_gcs_v2.py", line 48, in transcribe_gcs_v2 operation = client.create_recognizer(request=request) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/speech_v2/services/speech/client.py", line 695, in create_recognizer response = rpc( File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.ServiceUnavailable: 503 502:Bad Gateway The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 349, in retry_wrapped_func return retry_target( File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 207, in retry_target raise exceptions.RetryError( google.api_core.exceptions.RetryError: Deadline of 120.0s exceeded while calling target function, last exception: 503 502:Bad Gateway</pre></details>
1.0
speech.snippets.transcribe_gcs_v2_test: test_transcribe_gcs_v2 failed - Note: #9423 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 16ab3cf93caddb092710ba507d537945128c884c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/297c18a8-be85-43ab-92b9-655f63a7de99), [Sponge](http://sponge2/297c18a8-be85-43ab-92b9-655f63a7de99) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAVAILABLE details = "502:Bad Gateway" debug_error_string = "UNKNOWN:Error received from peer {grpc_message:"502:Bad Gateway", grpc_status:14, created_time:"2023-06-17T13:39:31.187553795+00:00"}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 191, in retry_target return target() File "/workspace/speech/snippets/transcribe_gcs_v2_test.py", line 38, in test_transcribe_gcs_v2 response = transcribe_gcs_v2.transcribe_gcs_v2( File "/workspace/speech/snippets/transcribe_gcs_v2.py", line 48, in transcribe_gcs_v2 operation = client.create_recognizer(request=request) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/speech_v2/services/speech/client.py", line 695, in create_recognizer response = rpc( File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.ServiceUnavailable: 503 502:Bad Gateway The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 349, in retry_wrapped_func return retry_target( File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 207, in retry_target raise exceptions.RetryError( google.api_core.exceptions.RetryError: Deadline of 120.0s exceeded while calling target function, last exception: 503 502:Bad Gateway</pre></details>
priority
speech snippets transcribe gcs test test transcribe gcs failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace speech snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace speech snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unavailable details bad gateway debug error string unknown error received from peer grpc message bad gateway grpc status created time the above exception was the direct cause of the following exception traceback most recent call last file workspace speech snippets nox py lib site packages google api core retry py line in retry target return target file workspace speech snippets transcribe gcs test py line in test transcribe gcs response transcribe gcs transcribe gcs file workspace speech snippets transcribe gcs py line in transcribe gcs operation client create recognizer request request file workspace speech snippets nox py lib site packages google cloud speech services speech client py line in create recognizer response rpc file workspace speech snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions serviceunavailable bad gateway the above exception was the direct cause of the following exception traceback most recent call last file workspace speech snippets nox py lib site packages google api core retry py line in retry wrapped func return retry target file workspace speech snippets nox py lib site packages google api core retry py line in retry target raise exceptions retryerror google api core exceptions retryerror deadline of exceeded while calling target function last exception bad gateway
1
257,248
8,134,897,438
IssuesEvent
2018-08-19 21:14:51
python/mypy
https://api.github.com/repos/python/mypy
opened
Re-work how fine grained targets are processed
needs discussion priority-1-normal refactoring topic-fine-grained-incremental
Currently, fine grained targets are processed per updated module. This can lead to files being processed multiple times (and also a bit hard to reason, but this may be subjective). I propose to reorganise them to be processed in topologically sorted order. So the algorithm would be like this: 1. Process all edited files, calculate all fired triggers, chain them to find all invalid targets, check for blocking errors. 2. Group targets per module, order them per SCC, then per same heuristics we use to order modules within SCCs in coarse grained incremental. Within module targets are ordered by line number, this is unchanged. 3. Process targets in one module from the queue, calculate updated deps and fired triggers, update invalid targets queue (maintaining the sort order), check blockers 4. Continue step 3 until no modules left in the queue at this SCC 5. Flush error messages 6. Continue steps 3-5 until no SCCs left This way it is much less likely that we will reprocess the same module twice. This will probably give an especially significant performance gain for cold runs, where many modules are updated w.r.t. to remote cache. Also IMO this algorithm is easier to reason about (and more similar to what happens in coarse-grained mode). This idea appeared some time ago, but was postponed. Filing an issue to not forget about this.
1.0
Re-work how fine grained targets are processed - Currently, fine grained targets are processed per updated module. This can lead to files being processed multiple times (and also a bit hard to reason, but this may be subjective). I propose to reorganise them to be processed in topologically sorted order. So the algorithm would be like this: 1. Process all edited files, calculate all fired triggers, chain them to find all invalid targets, check for blocking errors. 2. Group targets per module, order them per SCC, then per same heuristics we use to order modules within SCCs in coarse grained incremental. Within module targets are ordered by line number, this is unchanged. 3. Process targets in one module from the queue, calculate updated deps and fired triggers, update invalid targets queue (maintaining the sort order), check blockers 4. Continue step 3 until no modules left in the queue at this SCC 5. Flush error messages 6. Continue steps 3-5 until no SCCs left This way it is much less likely that we will reprocess the same module twice. This will probably give an especially significant performance gain for cold runs, where many modules are updated w.r.t. to remote cache. Also IMO this algorithm is easier to reason about (and more similar to what happens in coarse-grained mode). This idea appeared some time ago, but was postponed. Filing an issue to not forget about this.
priority
re work how fine grained targets are processed currently fine grained targets are processed per updated module this can lead to files being processed multiple times and also a bit hard to reason but this may be subjective i propose to reorganise them to be processed in topologically sorted order so the algorithm would be like this process all edited files calculate all fired triggers chain them to find all invalid targets check for blocking errors group targets per module order them per scc then per same heuristics we use to order modules within sccs in coarse grained incremental within module targets are ordered by line number this is unchanged process targets in one module from the queue calculate updated deps and fired triggers update invalid targets queue maintaining the sort order check blockers continue step until no modules left in the queue at this scc flush error messages continue steps until no sccs left this way it is much less likely that we will reprocess the same module twice this will probably give an especially significant performance gain for cold runs where many modules are updated w r t to remote cache also imo this algorithm is easier to reason about and more similar to what happens in coarse grained mode this idea appeared some time ago but was postponed filing an issue to not forget about this
1
56,680
14,078,482,024
IssuesEvent
2020-11-04 13:38:02
themagicalmammal/android_kernel_samsung_s5neolte
https://api.github.com/repos/themagicalmammal/android_kernel_samsung_s5neolte
opened
CVE-2019-15926 (High) detected in linuxv3.10
security vulnerability
## CVE-2019-15926 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_s5neolte/commit/f978d7dbb980bbe5267a625da958c4226e1a8ae0">f978d7dbb980bbe5267a625da958c4226e1a8ae0</a></p> <p>Found in base branch: <b>cosmic-experimental-1.6</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.2.3. Out of bounds access exists in the functions ath6kl_wmi_pstream_timeout_event_rx and ath6kl_wmi_cac_event_rx in the file drivers/net/wireless/ath/ath6kl/wmi.c. <p>Publish Date: 2019-09-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15926>CVE-2019-15926</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3</a></p> <p>Release Date: 2019-09-04</p> <p>Fix Resolution: v5.3-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-15926 (High) detected in linuxv3.10 - ## CVE-2019-15926 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_s5neolte/commit/f978d7dbb980bbe5267a625da958c4226e1a8ae0">f978d7dbb980bbe5267a625da958c4226e1a8ae0</a></p> <p>Found in base branch: <b>cosmic-experimental-1.6</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.2.3. Out of bounds access exists in the functions ath6kl_wmi_pstream_timeout_event_rx and ath6kl_wmi_cac_event_rx in the file drivers/net/wireless/ath/ath6kl/wmi.c. <p>Publish Date: 2019-09-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15926>CVE-2019-15926</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3</a></p> <p>Release Date: 2019-09-04</p> <p>Fix Resolution: v5.3-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch cosmic experimental vulnerable source files vulnerability details an issue was discovered in the linux kernel before out of bounds access exists in the functions wmi pstream timeout event rx and wmi cac event rx in the file drivers net wireless ath wmi c publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
70,221
9,382,912,243
IssuesEvent
2019-04-05 00:32:19
certbot/certbot
https://api.github.com/repos/certbot/certbot
closed
Clarify documentation of configuration file
area: documentation
https://certbot.eff.org/docs/using.html#configuration-file says:: ``` # All flags used by the client can be configured here. Run Certbot with # "--help" to learn more about the available options. ``` This begs the question of whether there are configuration options that are not flags to the client, and it is not answered in this section. This should be clarified, either by discussing which options are configuration file only, or changing the above to something like: ``` # The possible configuration options are exactly the same as the flags # used by the client. Run Certbot with "--help" to learn more about the # available options. ```
1.0
Clarify documentation of configuration file - https://certbot.eff.org/docs/using.html#configuration-file says:: ``` # All flags used by the client can be configured here. Run Certbot with # "--help" to learn more about the available options. ``` This begs the question of whether there are configuration options that are not flags to the client, and it is not answered in this section. This should be clarified, either by discussing which options are configuration file only, or changing the above to something like: ``` # The possible configuration options are exactly the same as the flags # used by the client. Run Certbot with "--help" to learn more about the # available options. ```
non_priority
clarify documentation of configuration file says all flags used by the client can be configured here run certbot with help to learn more about the available options this begs the question of whether there are configuration options that are not flags to the client and it is not answered in this section this should be clarified either by discussing which options are configuration file only or changing the above to something like the possible configuration options are exactly the same as the flags used by the client run certbot with help to learn more about the available options
0
104,038
4,194,219,805
IssuesEvent
2016-06-25 00:18:10
UHMDCmd/DCmd
https://api.github.com/repos/UHMDCmd/DCmd
opened
new Service-Host Assignment does not show.
bug Medium Priority
When editing an application, creating a new service host assignment does not show in the grid once created or updated. Audit log shows that something had happen, but nothing is shown for it. On the Host page side, if the application has no service and try to make a new service host assignment, nothing happens. Only works if an existing service is there.
1.0
new Service-Host Assignment does not show. - When editing an application, creating a new service host assignment does not show in the grid once created or updated. Audit log shows that something had happen, but nothing is shown for it. On the Host page side, if the application has no service and try to make a new service host assignment, nothing happens. Only works if an existing service is there.
priority
new service host assignment does not show when editing an application creating a new service host assignment does not show in the grid once created or updated audit log shows that something had happen but nothing is shown for it on the host page side if the application has no service and try to make a new service host assignment nothing happens only works if an existing service is there
1
129,388
18,092,941,060
IssuesEvent
2021-09-22 05:16:59
AlexRogalskiy/typescript-tools
https://api.github.com/repos/AlexRogalskiy/typescript-tools
closed
CVE-2021-32796 (Medium) detected in xmldom-0.6.0.tgz
security vulnerability Status: Invalid
## CVE-2021-32796 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmldom-0.6.0.tgz</b></p></summary> <p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p> <p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz</a></p> <p>Path to dependency file: typescript-tools/package.json</p> <p>Path to vulnerable library: /node_modules/xmldom/package.json</p> <p> Dependency Hierarchy: - :x: **xmldom-0.6.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/typescript-tools/commit/afc1c33cea76f8d57316e4485c82961b29f04d08">afc1c33cea76f8d57316e4485c82961b29f04d08</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> xmldom is an open source pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module. xmldom versions 0.6.0 and older do not correctly escape special characters when serializing elements removed from their ancestor. This may lead to unexpected syntactic changes during XML processing in some downstream applications. This issue has been resolved in version 0.7.0. As a workaround downstream applications can validate the input and reject the maliciously crafted documents. <p>Publish Date: 2021-07-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32796>CVE-2021-32796</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q">https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q</a></p> <p>Release Date: 2021-07-27</p> <p>Fix Resolution: xmldom - 0.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-32796 (Medium) detected in xmldom-0.6.0.tgz - ## CVE-2021-32796 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmldom-0.6.0.tgz</b></p></summary> <p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p> <p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz</a></p> <p>Path to dependency file: typescript-tools/package.json</p> <p>Path to vulnerable library: /node_modules/xmldom/package.json</p> <p> Dependency Hierarchy: - :x: **xmldom-0.6.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/typescript-tools/commit/afc1c33cea76f8d57316e4485c82961b29f04d08">afc1c33cea76f8d57316e4485c82961b29f04d08</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> xmldom is an open source pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module. xmldom versions 0.6.0 and older do not correctly escape special characters when serializing elements removed from their ancestor. This may lead to unexpected syntactic changes during XML processing in some downstream applications. This issue has been resolved in version 0.7.0. As a workaround downstream applications can validate the input and reject the maliciously crafted documents. <p>Publish Date: 2021-07-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32796>CVE-2021-32796</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q">https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q</a></p> <p>Release Date: 2021-07-27</p> <p>Fix Resolution: xmldom - 0.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in xmldom tgz cve medium severity vulnerability vulnerable library xmldom tgz a pure javascript standard based xml dom level core domparser and xmlserializer module library home page a href path to dependency file typescript tools package json path to vulnerable library node modules xmldom package json dependency hierarchy x xmldom tgz vulnerable library found in head commit a href found in base branch master vulnerability details xmldom is an open source pure javascript standard based xml dom level core domparser and xmlserializer module xmldom versions and older do not correctly escape special characters when serializing elements removed from their ancestor this may lead to unexpected syntactic changes during xml processing in some downstream applications this issue has been resolved in version as a workaround downstream applications can validate the input and reject the maliciously crafted documents publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xmldom step up your open source security game with whitesource
0
526,540
15,295,240,829
IssuesEvent
2021-02-24 04:21:34
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Missing translations for onboarding modal on first launch
OS/Desktop QA/Yes l10n onboarding priority/P2 release-notes/exclude
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Missing translations for onboarding modal on first launch ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Set locale to France (FR) 2. Install 1.21.65 x64 beta build 3. Launch browser, trigger onboarding modal that is not translated ## Actual result: <!--Please add screenshots if needed--> ![image](https://user-images.githubusercontent.com/17010094/108821250-3abe1c00-75b5-11eb-9043-70cd7e4f3ce6.png) ## Expected result: Translated strings when locale is not EN ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.21.62 Chromium: 88.0.4324.182 (Official Build) beta (64-bit) -- | -- Revision | `73ee5087001dcef33047c4ed650471b225dd8caf-refs/branch-heads/4324@{#2202}` OS | Linux ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? NA - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @rebron @mkarolin
1.0
Missing translations for onboarding modal on first launch - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Missing translations for onboarding modal on first launch ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Set locale to France (FR) 2. Install 1.21.65 x64 beta build 3. Launch browser, trigger onboarding modal that is not translated ## Actual result: <!--Please add screenshots if needed--> ![image](https://user-images.githubusercontent.com/17010094/108821250-3abe1c00-75b5-11eb-9043-70cd7e4f3ce6.png) ## Expected result: Translated strings when locale is not EN ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.21.62 Chromium: 88.0.4324.182 (Official Build) beta (64-bit) -- | -- Revision | `73ee5087001dcef33047c4ed650471b225dd8caf-refs/branch-heads/4324@{#2202}` OS | Linux ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? NA - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @rebron @mkarolin
priority
missing translations for onboarding modal on first launch have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description missing translations for onboarding modal on first launch steps to reproduce set locale to france fr install beta build launch browser trigger onboarding modal that is not translated actual result expected result translated strings when locale is not en reproduces how often easy brave version brave version info brave chromium   official build  beta  bit revision refs branch heads os linux version channel information can you reproduce this issue with the current release na can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc rebron mkarolin
1
305,968
9,379,024,072
IssuesEvent
2019-04-04 14:07:48
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
Investigate image quality degradation
category - terrain and imagery priority - high type - bug
Reported [on the forum](https://groups.google.com/forum/#!topic/cesium-dev/xWcXPYkh3c8) Check out the below image. You'll notice the Cesium version is washed out and slightly blurry compared to the same image in Bing Maps (click it for full resolution). At some point in the rendering pipeline (texture filtering, AA, lighting, etc..) Cesium processing is reducing imagery quality. I looked at the source files in the network debugger and confirmed that both engines are using the same exact imagery and the rendered Bing map version is identical to the source (since it's just doing img elements like traditional webmaps). This means that Cesium is definitely the culprit here. ![image](https://cloud.githubusercontent.com/assets/1495196/11547017/d15c397c-991e-11e5-9f14-80133036d9ba.png)
1.0
Investigate image quality degradation - Reported [on the forum](https://groups.google.com/forum/#!topic/cesium-dev/xWcXPYkh3c8) Check out the below image. You'll notice the Cesium version is washed out and slightly blurry compared to the same image in Bing Maps (click it for full resolution). At some point in the rendering pipeline (texture filtering, AA, lighting, etc..) Cesium processing is reducing imagery quality. I looked at the source files in the network debugger and confirmed that both engines are using the same exact imagery and the rendered Bing map version is identical to the source (since it's just doing img elements like traditional webmaps). This means that Cesium is definitely the culprit here. ![image](https://cloud.githubusercontent.com/assets/1495196/11547017/d15c397c-991e-11e5-9f14-80133036d9ba.png)
priority
investigate image quality degradation reported check out the below image you ll notice the cesium version is washed out and slightly blurry compared to the same image in bing maps click it for full resolution at some point in the rendering pipeline texture filtering aa lighting etc cesium processing is reducing imagery quality i looked at the source files in the network debugger and confirmed that both engines are using the same exact imagery and the rendered bing map version is identical to the source since it s just doing img elements like traditional webmaps this means that cesium is definitely the culprit here
1
2,782
2,643,518,606
IssuesEvent
2015-03-12 11:34:29
owncloud/core
https://api.github.com/repos/owncloud/core
closed
[oc update 8.0.2] Loading gif showing forever in update view
Design
When installing an update (from 8.0.0 to 8.0.2) and experiencing an error the loading-small.gif is displayed forever below the "3. Replace" section even though the update process terminated. Also, should this really be a .gif animation? Suggesting a replacement with a more modern looking css or js loading symbol. (http://cssload.net/ and http://tobiasahlin.com/spinkit/ etc.)
1.0
[oc update 8.0.2] Loading gif showing forever in update view - When installing an update (from 8.0.0 to 8.0.2) and experiencing an error the loading-small.gif is displayed forever below the "3. Replace" section even though the update process terminated. Also, should this really be a .gif animation? Suggesting a replacement with a more modern looking css or js loading symbol. (http://cssload.net/ and http://tobiasahlin.com/spinkit/ etc.)
non_priority
loading gif showing forever in update view when installing an update from to and experiencing an error the loading small gif is displayed forever below the replace section even though the update process terminated also should this really be a gif animation suggesting a replacement with a more modern looking css or js loading symbol and etc
0
303,876
26,235,122,000
IssuesEvent
2023-01-05 06:14:08
kubernetes-sigs/kwok
https://api.github.com/repos/kubernetes-sigs/kwok
closed
Failing `kwokctl snapshot` with runtime `nerdctl`
kind/failing-test
### Which jobs are failing? https://github.com/kubernetes-sigs/kwok/blob/88e598265a3c899bfb010088d8a6d85ebf429a57/.github/workflows/test.yaml#L137-L141 ### Which tests are failing? https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840 ``` console ================================================================================ Testing kwokctl/kwokctl_nerdctl_snapshot... github.com/docker/buildx 0.[9](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:10).1+azure-2 ed00243a0ce2a0aee75311b06e32d33b44729689 mkdir: cannot create directory ‘tmp’: File exists unpacking docker.io/local/kwok:test (sha256:5019cea8f3dd71453e676622dd7825d12c2e7ae4272560aa295d7787dfbf0245)... Loaded image: docker.io/local/kwok:testTest snapshot on nerdctl for 1.25.3 1.24.7 1.23.13 1.22.15 1.21.14 1.20.15 ------------------------------ Testing snapshot on nerdctl for 1.25.3 {"time":"2022-11-28T14:50:53.869177459Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} {"time":"2022-11-28T14:50:54.550340799Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} time="2022-11-28T14:50:54Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-25-3" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.25.3" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.25.3" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.25.3" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-etcd" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver" You can now use your cluster with: {"time":"2022-11-28T14:50:59.168789831Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-25-3"} kubectl config use-context kwok-snapshot-cluster-nerdctl-1-25-3 Thanks for using kwok! No resources found No resources found deployment.apps/fake-pod created node/fake-node created Download https://github.com/etcd-io/etcd/releases/download/v3.5.6/etcd-v3.5.6-linux-amd64.tar.gz {"time":"2022-11-28T14:51:37.882622576Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-25-3/etcd-data kwok-snapshot-cluster-nerdctl-1-25-3-etcd:/: exit status 1\ntime=\"2022-11-28T14:51:37Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/14518/ns/net\\\": no such file or directory\" id=2af2754d12fd763d179bf1e3225b62a7ea3d5a86a37eabc1bb29944a04c8ce85\ntime=\"2022-11-28T14:51:37Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} Error: Empty snapshot restore failed Expected: Actual: NAMESPACE NAME default fake-pod-78f5b8f676-gj6r9 NAME fake-node {"time":"2022-11-28T14:52:34.878035422Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} time="2022-11-28T14:52:34Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-etcd" time="2022-11-28T14:52:35Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-25-3" {"time":"2022-11-28T14:52:35.731964445Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} {"time":"2022-11-28T14:52:35.882827316Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-25-3"} ------------------------------ Testing snapshot on nerdctl for 1.24.7 {"time":"2022-11-28T14:52:35.899447477Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} {"time":"2022-11-28T14:52:37.003300129Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} time="2022-11-28T14:52:37Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-24-7" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.24.7" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.24.7" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.24.7" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-etcd" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver" You can now use your cluster with: {"time":"2022-11-28T14:52:41.660050875Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-24-7"} kubectl config use-context kwok-snapshot-cluster-nerdctl-1-24-7 Thanks for using kwok! No resources found No resources found deployment.apps/fake-pod created node/fake-node created {"time":"2022-11-28T14:53:19.502931142Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-24-7/etcd-data kwok-snapshot-cluster-nerdctl-1-24-7-etcd:/: exit status 1\ntime=\"2022-11-28T14:53:19Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/18073/ns/net\\\": no such file or directory\" id=4fb042bd8586bcd039dac994d5262e857ed2631ddcc8509367aa5ac50201a179\ntime=\"2022-11-28T14:53:19Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} Error: Empty snapshot restore failed Expected: Actual: NAMESPACE NAME default fake-pod-6f5fffcbc-hff5p NAME fake-node {"time":"2022-11-28T14:54:16.753867524Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller" time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler" time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager" time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver" time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-etcd" time="2022-11-28T14:54:17Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-24-7" {"time":"2022-11-28T14:54:17.535599245Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} {"time":"2022-11-28T14:54:17.677782486Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-24-7"} ------------------------------ Testing snapshot on nerdctl for 1.23.13 {"time":"2022-11-28T14:54:17.692811822Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} {"time":"2022-11-28T14:54:18.561574115Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} time="2022-11-28T14:54:18Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-23-13" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.23.13" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.23.13" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.23.13" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-etcd" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller" You can now use your cluster with: {"time":"2022-11-28T14:54:22.92009441Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-23-13"} kubectl config use-context kwok-snapshot-cluster-nerdctl-1-23-13 Thanks for using kwok! No resources found No resources found deployment.apps/fake-pod created node/fake-node created {"time":"2022-11-28T14:55:00.555011[10](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:11)1Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-23-13/etcd-data kwok-snapshot-cluster-nerdctl-1-23-13-etcd:/: exit status 1\ntime=\"2022-[11](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:12)-28T14:55:00Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/21546/ns/net\\\": no such file or directory\" id=37a024f06534b4c24888948cb02592bfbf348a86def52b0ef420c9b2278c5e95\ntime=\"2022-11-28T14:55:00Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} Error: Empty snapshot restore failed Expected: Actual: NAMESPACE NAME default fake-pod-6bf4bdd9cc-5jqlf NAME fake-node {"time":"2022-11-28T14:55:57.588653024Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller" time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler" time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager" time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver" time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-etcd" time="2022-11-28T14:55:58Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-23-13" {"time":"2022-11-28T14:55:58.389470291Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} {"time":"2022-11-28T14:55:58.526502308Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-23-13"} ------------------------------ Testing snapshot on nerdctl for 1.22.15 {"time":"2022-11-28T14:55:58.545860907Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"} {"time":"2022-11-28T14:55:59.42[12](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:13)21125Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"} time="2022-11-28T[14](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:15):55:59Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-22-[15](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:16)" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.22.15" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.22.15" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.22.15" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-controller-manager" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kwok-controller" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-scheduler" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-etcd" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-apiserver" You can now use your cluster with: kubectl config use-context kwok-snapshot-cluster-nerdctl-1-22-15 Thanks for using kwok! {"time":"2022-11-28T14:56:03.6658[16](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:17)766Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-22-15"} No resources found No resources found deployment.apps/fake-pod created node/fake-node created {"time":"2022-11-28T14:56:41.2365[17](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:18)354Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.[19](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:20).3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-22-15/etcd-data kwok-snapshot-cluster-nerdctl-1-22-15-etcd:/: exit status 1\ntime=\"[20](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:21)[22](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:23)-11-28T14:56:40Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/25054/ns/net\\\": no such file or directory\" id=8ec[23](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:24)8fa41f2f3a7039bf6e120f2de73ad776cc8c775db43b3301157aa430f[26](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:27)\ntime=\"2022-11-[28](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:29)T14:56:[40](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:41)Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} ``` ### Since when has it been failing? Always ### Reason for failure (if possible) _No response_ ### Anything else we need to know? _No response_
1.0
Failing `kwokctl snapshot` with runtime `nerdctl` - ### Which jobs are failing? https://github.com/kubernetes-sigs/kwok/blob/88e598265a3c899bfb010088d8a6d85ebf429a57/.github/workflows/test.yaml#L137-L141 ### Which tests are failing? https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840 ``` console ================================================================================ Testing kwokctl/kwokctl_nerdctl_snapshot... github.com/docker/buildx 0.[9](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:10).1+azure-2 ed00243a0ce2a0aee75311b06e32d33b44729689 mkdir: cannot create directory ‘tmp’: File exists unpacking docker.io/local/kwok:test (sha256:5019cea8f3dd71453e676622dd7825d12c2e7ae4272560aa295d7787dfbf0245)... Loaded image: docker.io/local/kwok:testTest snapshot on nerdctl for 1.25.3 1.24.7 1.23.13 1.22.15 1.21.14 1.20.15 ------------------------------ Testing snapshot on nerdctl for 1.25.3 {"time":"2022-11-28T14:50:53.869177459Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} {"time":"2022-11-28T14:50:54.550340799Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} time="2022-11-28T14:50:54Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-25-3" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.25.3" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.25.3" time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.25.3" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-etcd" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager" time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver" You can now use your cluster with: {"time":"2022-11-28T14:50:59.168789831Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-25-3"} kubectl config use-context kwok-snapshot-cluster-nerdctl-1-25-3 Thanks for using kwok! No resources found No resources found deployment.apps/fake-pod created node/fake-node created Download https://github.com/etcd-io/etcd/releases/download/v3.5.6/etcd-v3.5.6-linux-amd64.tar.gz {"time":"2022-11-28T14:51:37.882622576Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-25-3/etcd-data kwok-snapshot-cluster-nerdctl-1-25-3-etcd:/: exit status 1\ntime=\"2022-11-28T14:51:37Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/14518/ns/net\\\": no such file or directory\" id=2af2754d12fd763d179bf1e3225b62a7ea3d5a86a37eabc1bb29944a04c8ce85\ntime=\"2022-11-28T14:51:37Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} Error: Empty snapshot restore failed Expected: Actual: NAMESPACE NAME default fake-pod-78f5b8f676-gj6r9 NAME fake-node {"time":"2022-11-28T14:52:34.878035422Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} time="2022-11-28T14:52:34Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver" time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-etcd" time="2022-11-28T14:52:35Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-25-3" {"time":"2022-11-28T14:52:35.731964445Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"} {"time":"2022-11-28T14:52:35.882827316Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-25-3"} ------------------------------ Testing snapshot on nerdctl for 1.24.7 {"time":"2022-11-28T14:52:35.899447477Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} {"time":"2022-11-28T14:52:37.003300129Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} time="2022-11-28T14:52:37Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-24-7" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.24.7" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.24.7" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.24.7" time="2022-11-28T14:52:37Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-etcd" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller" time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver" You can now use your cluster with: {"time":"2022-11-28T14:52:41.660050875Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-24-7"} kubectl config use-context kwok-snapshot-cluster-nerdctl-1-24-7 Thanks for using kwok! No resources found No resources found deployment.apps/fake-pod created node/fake-node created {"time":"2022-11-28T14:53:19.502931142Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-24-7/etcd-data kwok-snapshot-cluster-nerdctl-1-24-7-etcd:/: exit status 1\ntime=\"2022-11-28T14:53:19Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/18073/ns/net\\\": no such file or directory\" id=4fb042bd8586bcd039dac994d5262e857ed2631ddcc8509367aa5ac50201a179\ntime=\"2022-11-28T14:53:19Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} Error: Empty snapshot restore failed Expected: Actual: NAMESPACE NAME default fake-pod-6f5fffcbc-hff5p NAME fake-node {"time":"2022-11-28T14:54:16.753867524Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller" time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler" time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager" time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver" time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-etcd" time="2022-11-28T14:54:17Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-24-7" {"time":"2022-11-28T14:54:17.535599245Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"} {"time":"2022-11-28T14:54:17.677782486Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-24-7"} ------------------------------ Testing snapshot on nerdctl for 1.23.13 {"time":"2022-11-28T14:54:17.692811822Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} {"time":"2022-11-28T14:54:18.561574115Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} time="2022-11-28T14:54:18Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-23-13" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.23.13" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.23.13" time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.23.13" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-etcd" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver" time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller" You can now use your cluster with: {"time":"2022-11-28T14:54:22.92009441Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-23-13"} kubectl config use-context kwok-snapshot-cluster-nerdctl-1-23-13 Thanks for using kwok! No resources found No resources found deployment.apps/fake-pod created node/fake-node created {"time":"2022-11-28T14:55:00.555011[10](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:11)1Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-23-13/etcd-data kwok-snapshot-cluster-nerdctl-1-23-13-etcd:/: exit status 1\ntime=\"2022-[11](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:12)-28T14:55:00Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/21546/ns/net\\\": no such file or directory\" id=37a024f06534b4c24888948cb02592bfbf348a86def52b0ef420c9b2278c5e95\ntime=\"2022-11-28T14:55:00Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} Error: Empty snapshot restore failed Expected: Actual: NAMESPACE NAME default fake-pod-6bf4bdd9cc-5jqlf NAME fake-node {"time":"2022-11-28T14:55:57.588653024Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller" time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler" time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager" time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver" time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-etcd" time="2022-11-28T14:55:58Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-23-13" {"time":"2022-11-28T14:55:58.389470291Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"} {"time":"2022-11-28T14:55:58.526502308Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-23-13"} ------------------------------ Testing snapshot on nerdctl for 1.22.15 {"time":"2022-11-28T14:55:58.545860907Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"} {"time":"2022-11-28T14:55:59.42[12](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:13)21125Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"} time="2022-11-28T[14](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:15):55:59Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-22-[15](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:16)" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_apiserver: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_scheduler: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kwok_controller: [Links]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.22.15" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.22.15" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.22.15" time="2022-11-28T14:55:59Z" level=info msg="Ensuring image local/kwok:test" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-controller-manager" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kwok-controller" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-scheduler" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-etcd" time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-apiserver" You can now use your cluster with: kubectl config use-context kwok-snapshot-cluster-nerdctl-1-22-15 Thanks for using kwok! {"time":"2022-11-28T14:56:03.6658[16](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:17)766Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-22-15"} No resources found No resources found deployment.apps/fake-pod created node/fake-node created {"time":"2022-11-28T14:56:41.2365[17](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:18)354Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.[19](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:20).3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-22-15/etcd-data kwok-snapshot-cluster-nerdctl-1-22-15-etcd:/: exit status 1\ntime=\"[20](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:21)[22](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:23)-11-28T14:56:40Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/25054/ns/net\\\": no such file or directory\" id=8ec[23](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:24)8fa41f2f3a7039bf6e120f2de73ad776cc8c775db43b3301157aa430f[26](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:27)\ntime=\"2022-11-[28](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:29)T14:56:[40](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:41)Z\" level=fatal msg=\"expected container status running, got stopped\"\n"} ``` ### Since when has it been failing? Always ### Reason for failure (if possible) _No response_ ### Anything else we need to know? _No response_
non_priority
failing kwokctl snapshot with runtime nerdctl which jobs are failing which tests are failing console testing kwokctl kwokctl nerdctl snapshot github com docker buildx mkdir cannot create directory ‘tmp’ file exists unpacking docker io local kwok test loaded image docker io local kwok testtest snapshot on nerdctl for testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg starting cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image local kwok test time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg creating container kwok snapshot cluster nerdctl kwok controller time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl kube apiserver you can now use your cluster with time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg cluster is ready cluster snapshot cluster nerdctl kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok no resources found no resources found deployment apps fake pod created node fake node created download time level error source opt hostedtoolcache go src runtime proc go msg execute exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id ntime level fatal msg expected container status running got stopped n error empty snapshot restore failed expected actual namespace name default fake pod name fake node time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg stopping cluster cluster snapshot cluster nerdctl time level info msg removing container kwok snapshot cluster nerdctl kube scheduler time level info msg removing container kwok snapshot cluster nerdctl kube controller manager time level info msg removing container kwok snapshot cluster nerdctl kwok controller time level info msg removing container kwok snapshot cluster nerdctl kube apiserver time level info msg removing container kwok snapshot cluster nerdctl etcd time level info msg removing network kwok snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg deleting cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg cluster deleted cluster snapshot cluster nerdctl testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg starting cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg ensuring image local kwok test time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kwok controller time level info msg creating container kwok snapshot cluster nerdctl kube apiserver you can now use your cluster with time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg cluster is ready cluster snapshot cluster nerdctl kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok no resources found no resources found deployment apps fake pod created node fake node created time level error source opt hostedtoolcache go src runtime proc go msg execute exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id ntime level fatal msg expected container status running got stopped n error empty snapshot restore failed expected actual namespace name default fake pod name fake node time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg stopping cluster cluster snapshot cluster nerdctl time level info msg removing container kwok snapshot cluster nerdctl kwok controller time level info msg removing container kwok snapshot cluster nerdctl kube scheduler time level info msg removing container kwok snapshot cluster nerdctl kube controller manager time level info msg removing container kwok snapshot cluster nerdctl kube apiserver time level info msg removing container kwok snapshot cluster nerdctl etcd time level info msg removing network kwok snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg deleting cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg cluster deleted cluster snapshot cluster nerdctl testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg starting cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image local kwok test time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl kube apiserver time level info msg creating container kwok snapshot cluster nerdctl kwok controller you can now use your cluster with time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg cluster is ready cluster snapshot cluster nerdctl kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok no resources found no resources found deployment apps fake pod created node fake node created time exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id ntime level fatal msg expected container status running got stopped n error empty snapshot restore failed expected actual namespace name default fake pod name fake node time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg stopping cluster cluster snapshot cluster nerdctl time level info msg removing container kwok snapshot cluster nerdctl kwok controller time level info msg removing container kwok snapshot cluster nerdctl kube scheduler time level info msg removing container kwok snapshot cluster nerdctl kube controller manager time level info msg removing container kwok snapshot cluster nerdctl kube apiserver time level info msg removing container kwok snapshot cluster nerdctl etcd time level info msg removing network kwok snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg deleting cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg cluster deleted cluster snapshot cluster nerdctl testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg ensuring image local kwok test time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl kwok controller time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kube apiserver you can now use your cluster with kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok time is ready cluster snapshot cluster nerdctl no resources found no resources found deployment apps fake pod created node fake node created time exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id level fatal msg expected container status running got stopped n since when has it been failing always reason for failure if possible no response anything else we need to know no response
0
163,259
6,194,884,665
IssuesEvent
2017-07-05 11:04:46
zero-os/0-orchestrator
https://api.github.com/repos/zero-os/0-orchestrator
closed
Rename rocksdb-version parameter to storage-engine-version
priority_major state_verification type_bug
To specify the ardb flist version we are now using `rocksdb-version`, which is a bad name because it points to the storage engine used by ardb to store the key values, while it is actually a pointer to the flist we need to start for bringing up the key value stores. `storage-engine-version` is a better name because it points more to what the flist actually delivers. > **Important** Also adapt the documentation (setup.md)
1.0
Rename rocksdb-version parameter to storage-engine-version - To specify the ardb flist version we are now using `rocksdb-version`, which is a bad name because it points to the storage engine used by ardb to store the key values, while it is actually a pointer to the flist we need to start for bringing up the key value stores. `storage-engine-version` is a better name because it points more to what the flist actually delivers. > **Important** Also adapt the documentation (setup.md)
priority
rename rocksdb version parameter to storage engine version to specify the ardb flist version we are now using rocksdb version which is a bad name because it points to the storage engine used by ardb to store the key values while it is actually a pointer to the flist we need to start for bringing up the key value stores storage engine version is a better name because it points more to what the flist actually delivers important also adapt the documentation setup md
1
369,184
10,889,408,872
IssuesEvent
2019-11-18 18:10:50
SpaceStation0/star-engine
https://api.github.com/repos/SpaceStation0/star-engine
opened
Implementation of ECS + event system
A-ecs E-medium P-priority-high T-feature
The specific details will be fleshed out once the specification has been written, but for now, this is known for sure: - The way the engine will process things will be a hybrid of a pure ECS and an event system - Once the ECS finishes processing, an event queue variable will be extracted and sent to the event system - The event system has a set of notifiers. Events will be send to notifiers, which can also emit events. - Events emitted by notifiers will be collected, and sent out to their respective notifiers - The process ends once there are no more events to process
1.0
Implementation of ECS + event system - The specific details will be fleshed out once the specification has been written, but for now, this is known for sure: - The way the engine will process things will be a hybrid of a pure ECS and an event system - Once the ECS finishes processing, an event queue variable will be extracted and sent to the event system - The event system has a set of notifiers. Events will be send to notifiers, which can also emit events. - Events emitted by notifiers will be collected, and sent out to their respective notifiers - The process ends once there are no more events to process
priority
implementation of ecs event system the specific details will be fleshed out once the specification has been written but for now this is known for sure the way the engine will process things will be a hybrid of a pure ecs and an event system once the ecs finishes processing an event queue variable will be extracted and sent to the event system the event system has a set of notifiers events will be send to notifiers which can also emit events events emitted by notifiers will be collected and sent out to their respective notifiers the process ends once there are no more events to process
1
341,394
30,584,770,052
IssuesEvent
2023-07-21 12:34:10
proarc/proarc-client
https://api.github.com/repos/proarc/proarc-client
closed
Chyba založení STT vícedílné monografie
1 chyba 6 k testování 7 návrh na zavření 6c otestováno: KNAV 6f otestováno: MKP priorita 1
Při pokusu o založení STT vícedílné monografie se záznam z katalogu nepřevzal do formuláře. Testovací záznam: sysno = 000737900, katalog knav monografie. ![image](https://github.com/proarc/proarc-client/assets/79305516/b1ad093b-6163-4d80-9204-20069b152b9a)
3.0
Chyba založení STT vícedílné monografie - Při pokusu o založení STT vícedílné monografie se záznam z katalogu nepřevzal do formuláře. Testovací záznam: sysno = 000737900, katalog knav monografie. ![image](https://github.com/proarc/proarc-client/assets/79305516/b1ad093b-6163-4d80-9204-20069b152b9a)
non_priority
chyba založení stt vícedílné monografie při pokusu o založení stt vícedílné monografie se záznam z katalogu nepřevzal do formuláře testovací záznam sysno katalog knav monografie
0
337,095
10,210,335,695
IssuesEvent
2019-08-14 14:36:10
Signbank/Global-signbank
https://api.github.com/repos/Signbank/Global-signbank
opened
Order gloss typeahead results by length
ASL blocking bug high priority
When creating a new gloss, Julie wants to search for and select the lemma 'WE'. However, it is not in the list of 5 suggestions, because longer words like 'jewelry' and 'next-week' apparently go first, and the actual thing she is looking does not 'make the list'. A solution could be to simply sort the list of matches by length; for long words you can simply continue typing if your match is not in the list, while this is not possible for short words like WE.
1.0
Order gloss typeahead results by length - When creating a new gloss, Julie wants to search for and select the lemma 'WE'. However, it is not in the list of 5 suggestions, because longer words like 'jewelry' and 'next-week' apparently go first, and the actual thing she is looking does not 'make the list'. A solution could be to simply sort the list of matches by length; for long words you can simply continue typing if your match is not in the list, while this is not possible for short words like WE.
priority
order gloss typeahead results by length when creating a new gloss julie wants to search for and select the lemma we however it is not in the list of suggestions because longer words like jewelry and next week apparently go first and the actual thing she is looking does not make the list a solution could be to simply sort the list of matches by length for long words you can simply continue typing if your match is not in the list while this is not possible for short words like we
1
415,918
28,058,792,075
IssuesEvent
2023-03-29 11:12:01
CarsOk/castillo_del_rey
https://api.github.com/repos/CarsOk/castillo_del_rey
opened
Corregir el diagrama relacional
documentation
# HISTORIA DE USUARIO ## Yo como Instructor SENA ## Quiero Corregir el diagrama relacional ## Para Tener la información correcta y no tener contratiempos al momento del desarrollo del software # Criterios de aceptación - [ ] Que tengan las relaciones - [ ] Las llaves primarias y foráneas
1.0
Corregir el diagrama relacional - # HISTORIA DE USUARIO ## Yo como Instructor SENA ## Quiero Corregir el diagrama relacional ## Para Tener la información correcta y no tener contratiempos al momento del desarrollo del software # Criterios de aceptación - [ ] Que tengan las relaciones - [ ] Las llaves primarias y foráneas
non_priority
corregir el diagrama relacional historia de usuario yo como instructor sena quiero corregir el diagrama relacional para tener la información correcta y no tener contratiempos al momento del desarrollo del software criterios de aceptación que tengan las relaciones las llaves primarias y foráneas
0
273,296
20,780,547,081
IssuesEvent
2022-03-16 14:25:15
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
opened
[Docs] Data Loading Page
area/documentation
### Description YugabyteDB data loading experience is completely different from the experience you have with single-server relational databases. For instance, application developers who use Postgres, MySQL or SQL Server still load data with multiple INSERTs which is considered an anti-pattern in YugabyteDB. Developers do this with Postgres and MySQL simply because the loading time is fast and many sample datasets that you download from the Internet use multiple INSERTs instead of COPY. Such developers will be puzzled to see how the loading time with multiple INSERTS takes 2x, 3x, 10x longer with YugabyteDB. We need to educate and guide developers on how to approach the initial data loading in YugabyteDB. This is the step where the developer experience is poor. **Data Loading page** Create a separate **Data Loading** page. Place it right after or near the **Installation** and **Getting Started** sections. The logic here is that once the developers deploys YugabyteDB and succeeds with a getting-started-level app/tutorial she will move forward with some sample data that represents a subset of the data of a company's app or a data set downloaded from the Internet. Thus, the **Data Loading** page has to be visible in the navigation tree. **Data Loading page content** The content needs to explain various data loading optimization techniques and explain why some approaches don't work as expected (for instance the multiple INSERTs case). Some techniques to mention: 1. Loading with INSERTS * Replace multiple INSERTs to a single or a few INSERT statements * Execute the INSERT statement(s) in a single transaction - use `BEGIN` insert `COMMIT` structure * Disable triggers (but don't forget to turn them back) - `ALTER TABLE <table_name> disable TRIGGER ALL` * etc. more techniques are discused here: https://docs.google.com/document/d/1jCLiHDEHiYpgVObILDC_2Ormr-Kx36YhkqHXUCVGO1Q/edit#heading=h.28rbkno9vzy9 It's important to explain why each step is necessary in YugabyteDB, how we're different and why this is OK. Remember that the performance is poor even with a single-node YugbyteDB cluster running locally if you compare to a local Postgres instance. 2. Loading with the COPY command Suggest this as an alternate (preffered?) option to the #1 above. Provide steps and instructions. You can use Franck's blog for reference: https://dev.to/yugabyte/copy-progression-in-yugabytedb-4ghb **YugabyteDB on-prem vs. Yugabyte Cloud - fighting networking latency** We can deal with a developer who uses Yugabyte DB (or Platform) in her own environment or a developer who has started with Yugabyte Cloud. For the on-prem scenarious, suggest loading the data from a location close to the YugabyteDB deployment. For Yugabyte Cloud, well, it's more complicated because the developer needs to create an instance in a cloud region, where the database is running, and from that instance do the loading. For now, at least, we can provide some basic steps.
1.0
[Docs] Data Loading Page - ### Description YugabyteDB data loading experience is completely different from the experience you have with single-server relational databases. For instance, application developers who use Postgres, MySQL or SQL Server still load data with multiple INSERTs which is considered an anti-pattern in YugabyteDB. Developers do this with Postgres and MySQL simply because the loading time is fast and many sample datasets that you download from the Internet use multiple INSERTs instead of COPY. Such developers will be puzzled to see how the loading time with multiple INSERTS takes 2x, 3x, 10x longer with YugabyteDB. We need to educate and guide developers on how to approach the initial data loading in YugabyteDB. This is the step where the developer experience is poor. **Data Loading page** Create a separate **Data Loading** page. Place it right after or near the **Installation** and **Getting Started** sections. The logic here is that once the developers deploys YugabyteDB and succeeds with a getting-started-level app/tutorial she will move forward with some sample data that represents a subset of the data of a company's app or a data set downloaded from the Internet. Thus, the **Data Loading** page has to be visible in the navigation tree. **Data Loading page content** The content needs to explain various data loading optimization techniques and explain why some approaches don't work as expected (for instance the multiple INSERTs case). Some techniques to mention: 1. Loading with INSERTS * Replace multiple INSERTs to a single or a few INSERT statements * Execute the INSERT statement(s) in a single transaction - use `BEGIN` insert `COMMIT` structure * Disable triggers (but don't forget to turn them back) - `ALTER TABLE <table_name> disable TRIGGER ALL` * etc. more techniques are discused here: https://docs.google.com/document/d/1jCLiHDEHiYpgVObILDC_2Ormr-Kx36YhkqHXUCVGO1Q/edit#heading=h.28rbkno9vzy9 It's important to explain why each step is necessary in YugabyteDB, how we're different and why this is OK. Remember that the performance is poor even with a single-node YugbyteDB cluster running locally if you compare to a local Postgres instance. 2. Loading with the COPY command Suggest this as an alternate (preffered?) option to the #1 above. Provide steps and instructions. You can use Franck's blog for reference: https://dev.to/yugabyte/copy-progression-in-yugabytedb-4ghb **YugabyteDB on-prem vs. Yugabyte Cloud - fighting networking latency** We can deal with a developer who uses Yugabyte DB (or Platform) in her own environment or a developer who has started with Yugabyte Cloud. For the on-prem scenarious, suggest loading the data from a location close to the YugabyteDB deployment. For Yugabyte Cloud, well, it's more complicated because the developer needs to create an instance in a cloud region, where the database is running, and from that instance do the loading. For now, at least, we can provide some basic steps.
non_priority
data loading page description yugabytedb data loading experience is completely different from the experience you have with single server relational databases for instance application developers who use postgres mysql or sql server still load data with multiple inserts which is considered an anti pattern in yugabytedb developers do this with postgres and mysql simply because the loading time is fast and many sample datasets that you download from the internet use multiple inserts instead of copy such developers will be puzzled to see how the loading time with multiple inserts takes longer with yugabytedb we need to educate and guide developers on how to approach the initial data loading in yugabytedb this is the step where the developer experience is poor data loading page create a separate data loading page place it right after or near the installation and getting started sections the logic here is that once the developers deploys yugabytedb and succeeds with a getting started level app tutorial she will move forward with some sample data that represents a subset of the data of a company s app or a data set downloaded from the internet thus the data loading page has to be visible in the navigation tree data loading page content the content needs to explain various data loading optimization techniques and explain why some approaches don t work as expected for instance the multiple inserts case some techniques to mention loading with inserts replace multiple inserts to a single or a few insert statements execute the insert statement s in a single transaction use begin insert commit structure disable triggers but don t forget to turn them back alter table disable trigger all etc more techniques are discused here it s important to explain why each step is necessary in yugabytedb how we re different and why this is ok remember that the performance is poor even with a single node yugbytedb cluster running locally if you compare to a local postgres instance loading with the copy command suggest this as an alternate preffered option to the above provide steps and instructions you can use franck s blog for reference yugabytedb on prem vs yugabyte cloud fighting networking latency we can deal with a developer who uses yugabyte db or platform in her own environment or a developer who has started with yugabyte cloud for the on prem scenarious suggest loading the data from a location close to the yugabytedb deployment for yugabyte cloud well it s more complicated because the developer needs to create an instance in a cloud region where the database is running and from that instance do the loading for now at least we can provide some basic steps
0
88,300
25,365,034,562
IssuesEvent
2022-11-21 05:10:07
wixtoolset/issues
https://api.github.com/repos/wixtoolset/issues
closed
Sign the vsix packages
enhancement build process
If this issue is a feature request: * Describe the scenario and benefits that the feature supports. > the user feals more comfortable that he selected the right extension from the marketplace * Describe how you're accomplishing the feature today (if possible). > at the moment you see that it is not signed. > ![vsix_sign](https://cloud.githubusercontent.com/assets/20229016/23957553/fa5c888e-099f-11e7-82de-5ab5c85a7c16.jpg) * Describe what you'd like the new feature to do. > show that the extension that will be installed belongs to the wix toolset (.NET Foundation)
1.0
Sign the vsix packages - If this issue is a feature request: * Describe the scenario and benefits that the feature supports. > the user feals more comfortable that he selected the right extension from the marketplace * Describe how you're accomplishing the feature today (if possible). > at the moment you see that it is not signed. > ![vsix_sign](https://cloud.githubusercontent.com/assets/20229016/23957553/fa5c888e-099f-11e7-82de-5ab5c85a7c16.jpg) * Describe what you'd like the new feature to do. > show that the extension that will be installed belongs to the wix toolset (.NET Foundation)
non_priority
sign the vsix packages if this issue is a feature request describe the scenario and benefits that the feature supports the user feals more comfortable that he selected the right extension from the marketplace describe how you re accomplishing the feature today if possible at the moment you see that it is not signed describe what you d like the new feature to do show that the extension that will be installed belongs to the wix toolset net foundation
0
41,198
12,831,755,481
IssuesEvent
2020-07-07 06:14:06
rvvergara/todolist-api-igaku
https://api.github.com/repos/rvvergara/todolist-api-igaku
closed
CVE-2019-2391 (Medium) detected in bson-1.1.3.tgz
security vulnerability
## CVE-2019-2391 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bson-1.1.3.tgz</b></p></summary> <p>A bson parser for node.js and the browser</p> <p>Library home page: <a href="https://registry.npmjs.org/bson/-/bson-1.1.3.tgz">https://registry.npmjs.org/bson/-/bson-1.1.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/todolist-api-igaku/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/todolist-api-igaku/node_modules/bson/package.json</p> <p> Dependency Hierarchy: - mongoose-5.7.12.tgz (Root Library) - :x: **bson-1.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rvvergara/todolist-api-igaku/commit/e74ac424f4412547afcf733f031f27227a0f28e9">e74ac424f4412547afcf733f031f27227a0f28e9</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Incorrect parsing of certain JSON input may result in js-bson not correctly serializing BSON. This may cause unexpected application behaviour including data disclosure. <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2391>CVE-2019-2391</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mongodb/js-bson/pull/336/commits">https://github.com/mongodb/js-bson/pull/336/commits</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: v1.1.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-2391 (Medium) detected in bson-1.1.3.tgz - ## CVE-2019-2391 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bson-1.1.3.tgz</b></p></summary> <p>A bson parser for node.js and the browser</p> <p>Library home page: <a href="https://registry.npmjs.org/bson/-/bson-1.1.3.tgz">https://registry.npmjs.org/bson/-/bson-1.1.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/todolist-api-igaku/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/todolist-api-igaku/node_modules/bson/package.json</p> <p> Dependency Hierarchy: - mongoose-5.7.12.tgz (Root Library) - :x: **bson-1.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rvvergara/todolist-api-igaku/commit/e74ac424f4412547afcf733f031f27227a0f28e9">e74ac424f4412547afcf733f031f27227a0f28e9</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Incorrect parsing of certain JSON input may result in js-bson not correctly serializing BSON. This may cause unexpected application behaviour including data disclosure. <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2391>CVE-2019-2391</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mongodb/js-bson/pull/336/commits">https://github.com/mongodb/js-bson/pull/336/commits</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: v1.1.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in bson tgz cve medium severity vulnerability vulnerable library bson tgz a bson parser for node js and the browser library home page a href path to dependency file tmp ws scm todolist api igaku package json path to vulnerable library tmp ws scm todolist api igaku node modules bson package json dependency hierarchy mongoose tgz root library x bson tgz vulnerable library found in head commit a href vulnerability details incorrect parsing of certain json input may result in js bson not correctly serializing bson this may cause unexpected application behaviour including data disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
45,217
2,921,044,640
IssuesEvent
2015-06-24 22:09:35
notsecure/uTox
https://api.github.com/repos/notsecure/uTox
closed
installation too much work on Windows
enhancement high_priority Windows
From getting some people on Windows to use µTox, I found that the biggest installation issue they run into is that they have to choose an installation path manually. It's also unclear when you select a folder whether it'll create a subfolder or it'll just unpack in the folder you selected. Probably would be best if the default path was set to "C:\Tox" or something, and you can customize it if you want. Most people won't.
1.0
installation too much work on Windows - From getting some people on Windows to use µTox, I found that the biggest installation issue they run into is that they have to choose an installation path manually. It's also unclear when you select a folder whether it'll create a subfolder or it'll just unpack in the folder you selected. Probably would be best if the default path was set to "C:\Tox" or something, and you can customize it if you want. Most people won't.
priority
installation too much work on windows from getting some people on windows to use µtox i found that the biggest installation issue they run into is that they have to choose an installation path manually it s also unclear when you select a folder whether it ll create a subfolder or it ll just unpack in the folder you selected probably would be best if the default path was set to c tox or something and you can customize it if you want most people won t
1
508,664
14,704,252,422
IssuesEvent
2021-01-04 16:13:25
WilliBoelke/Cook-Shop
https://api.github.com/repos/WilliBoelke/Cook-Shop
closed
Steps Views
View feature priority 1
Views for Steps: - [ ] Add new Step - [ ] Update a Step - [ ] View Details - [ ] RecyclerView item
1.0
Steps Views - Views for Steps: - [ ] Add new Step - [ ] Update a Step - [ ] View Details - [ ] RecyclerView item
priority
steps views views for steps add new step update a step view details recyclerview item
1
37,113
2,815,222,665
IssuesEvent
2015-05-19 01:42:55
GoogleCloudPlatform/kubernetes
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
opened
Print object type as first column in `kubectl get`
area/usability priority/P2
When we originally implemented get, the first column was the type, i.e. "POD", not "NAME". One benefit is that the output (if saved or pasted or viewed later) gives a clue about the type. A downside is that the column width might be longer "SERVICEACCOUNT" this way. We should make a decision and be consistent (change Pod, or change everything else).
1.0
Print object type as first column in `kubectl get` - When we originally implemented get, the first column was the type, i.e. "POD", not "NAME". One benefit is that the output (if saved or pasted or viewed later) gives a clue about the type. A downside is that the column width might be longer "SERVICEACCOUNT" this way. We should make a decision and be consistent (change Pod, or change everything else).
priority
print object type as first column in kubectl get when we originally implemented get the first column was the type i e pod not name one benefit is that the output if saved or pasted or viewed later gives a clue about the type a downside is that the column width might be longer serviceaccount this way we should make a decision and be consistent change pod or change everything else
1
186,426
14,394,694,333
IssuesEvent
2020-12-03 01:54:27
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
flynn/flynn-discovery: Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go; 41 LoC
fresh small test
Found a possible issue in [flynn/flynn-discovery](https://www.github.com/flynn/flynn-discovery) at [Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85) <details> <summary>Click here to show the 41 line(s) of Go which triggered the analyzer.</summary> ```go for _, count := range []int{10000, 50, 432} { recs := make(chan *log.Record) done := make(chan int) spec := SpeculativeHandler(100, log.ChannelHandler(recs)) go func() { defer close(done) expectedCount := int(math.Min(float64(count), float64(100))) expectedIdx := count - expectedCount for r := range recs { if r.Ctx[1] != expectedIdx { t.Errorf("Bad ctx 'i', got %d expected %d", r.Ctx[1], expectedIdx) return } expectedIdx++ expectedCount-- if expectedCount == 0 { // got everything we expected break } } select { case <-recs: t.Errorf("got an extra record we shouldn't have!") default: } }() lg := log.New() lg.SetHandler(spec) for i := 0; i < count; i++ { lg.Debug("test speculative", "i", i) } go spec.Flush() // wait for the go routine to finish <-done } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable count used in defer or goroutine at line 52 </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 0fdc37b8399dd62aba3cfa194b0f810e945b6a0e
1.0
flynn/flynn-discovery: Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go; 41 LoC - Found a possible issue in [flynn/flynn-discovery](https://www.github.com/flynn/flynn-discovery) at [Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85) <details> <summary>Click here to show the 41 line(s) of Go which triggered the analyzer.</summary> ```go for _, count := range []int{10000, 50, 432} { recs := make(chan *log.Record) done := make(chan int) spec := SpeculativeHandler(100, log.ChannelHandler(recs)) go func() { defer close(done) expectedCount := int(math.Min(float64(count), float64(100))) expectedIdx := count - expectedCount for r := range recs { if r.Ctx[1] != expectedIdx { t.Errorf("Bad ctx 'i', got %d expected %d", r.Ctx[1], expectedIdx) return } expectedIdx++ expectedCount-- if expectedCount == 0 { // got everything we expected break } } select { case <-recs: t.Errorf("got an extra record we shouldn't have!") default: } }() lg := log.New() lg.SetHandler(spec) for i := 0; i < count; i++ { lg.Debug("test speculative", "i", i) } go spec.Flush() // wait for the go routine to finish <-done } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable count used in defer or goroutine at line 52 </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 0fdc37b8399dd62aba3cfa194b0f810e945b6a0e
non_priority
flynn flynn discovery godeps workspace src gopkg in inconshreveable ext ext test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for count range int recs make chan log record done make chan int spec speculativehandler log channelhandler recs go func defer close done expectedcount int math min count expectedidx count expectedcount for r range recs if r ctx expectedidx t errorf bad ctx i got d expected d r ctx expectedidx return expectedidx expectedcount if expectedcount got everything we expected break select case recs t errorf got an extra record we shouldn t have default lg log new lg sethandler spec for i i count i lg debug test speculative i i go spec flush wait for the go routine to finish done below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable count used in defer or goroutine at line leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
3,136
2,810,658,659
IssuesEvent
2015-05-17 03:03:02
HeavensGate/Eternal
https://api.github.com/repos/HeavensGate/Eternal
opened
Opt-in Verb
code Gate/writing
A verb which allows living players to indicate they are now willing to take a mid-round antag role. Preferably not spammable, since it'll bump the admins.
1.0
Opt-in Verb - A verb which allows living players to indicate they are now willing to take a mid-round antag role. Preferably not spammable, since it'll bump the admins.
non_priority
opt in verb a verb which allows living players to indicate they are now willing to take a mid round antag role preferably not spammable since it ll bump the admins
0
398,680
11,742,175,264
IssuesEvent
2020-03-11 23:51:59
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
Documents: 2011-2012 is not directly reachable
bug documents easy and fun priority: medium
In GitLab by gmulder on Jun 27, 2017, 01:21 ### One-sentence description The documents page should switch to the year 2011-2012 when it is selected from the drop-down box. ### Current behaviour The documents page does not switch to the year 2011-2012 when it is selected from the drop-down box. ### Expected behaviour The documents page switches to the year 2011-2012 when it is selected from the drop-down box. ### Steps to reproduce 1. Log in 2. Go to the `/documents/` page (e.g. Association -> Documents). 3. Navigate to "General meetings" 4. Click on `2011-2012` (a dropdown box appears) 5. Click on `2011-2012` in the dropdown box 6. Nothing happens. Alternatively, replace step 5 by clicking on another item (e.g. `2010-2011`) and then execute the original step 4 and 5 again. In that case, step 6 is replaced by the expected behaviour. Note: the period (i.e. `2011-2012`) is relative to the current period. When we switch to the next year, this period will be `2012-2013`.
1.0
Documents: 2011-2012 is not directly reachable - In GitLab by gmulder on Jun 27, 2017, 01:21 ### One-sentence description The documents page should switch to the year 2011-2012 when it is selected from the drop-down box. ### Current behaviour The documents page does not switch to the year 2011-2012 when it is selected from the drop-down box. ### Expected behaviour The documents page switches to the year 2011-2012 when it is selected from the drop-down box. ### Steps to reproduce 1. Log in 2. Go to the `/documents/` page (e.g. Association -> Documents). 3. Navigate to "General meetings" 4. Click on `2011-2012` (a dropdown box appears) 5. Click on `2011-2012` in the dropdown box 6. Nothing happens. Alternatively, replace step 5 by clicking on another item (e.g. `2010-2011`) and then execute the original step 4 and 5 again. In that case, step 6 is replaced by the expected behaviour. Note: the period (i.e. `2011-2012`) is relative to the current period. When we switch to the next year, this period will be `2012-2013`.
priority
documents is not directly reachable in gitlab by gmulder on jun one sentence description the documents page should switch to the year when it is selected from the drop down box current behaviour the documents page does not switch to the year when it is selected from the drop down box expected behaviour the documents page switches to the year when it is selected from the drop down box steps to reproduce log in go to the documents page e g association documents navigate to general meetings click on a dropdown box appears click on in the dropdown box nothing happens alternatively replace step by clicking on another item e g and then execute the original step and again in that case step is replaced by the expected behaviour note the period i e is relative to the current period when we switch to the next year this period will be
1
207,359
15,811,822,634
IssuesEvent
2021-04-05 03:48:55
lussierc/StockSwingPredictor
https://api.github.com/repos/lussierc/StockSwingPredictor
closed
Finalize Testing Suite
testing top-priority
As I near the completion of the project source code and the implementation of the prediction model, this code will need to be extensively tested to ensure it functions correctly. These test cases should help ensure a confidence in program correctness. To do this, test cases should be written for every part of the program. These test files should include: - `test_scraper.py`: ensure data is scraped and that it is in the correct format - `test_data_cleaner.py`: ensure data is properly converted - `test_UI.py`: test the CML and UI lightly to ensure they are working - `test_prediction.py`: ensures predictions are correctly generated by the program Now, I want to ensure all of my test cases I write are passing and I have a high success rate. However, in addition to ensuring what is being tested is working properly, I want to ensure that almost everything that can be tested is tested. This will ensure the tool as a whole functions well. With this, I want to make sure that my tool's code test coverage is high. Code coverage is a measure that analyzes how much source code of a program is executed (or tested) when a particular test suite runs. There is a library called [Coverage](https://coverage.readthedocs.io/en/coverage-5.5/) that will automatically measure the code coverage of your Python test suite, making it perfect for my project. With this, I will be using a version of the library that integrates with Pytest called [pytest-cov](https://pypi.org/project/pytest-cov/).
1.0
Finalize Testing Suite - As I near the completion of the project source code and the implementation of the prediction model, this code will need to be extensively tested to ensure it functions correctly. These test cases should help ensure a confidence in program correctness. To do this, test cases should be written for every part of the program. These test files should include: - `test_scraper.py`: ensure data is scraped and that it is in the correct format - `test_data_cleaner.py`: ensure data is properly converted - `test_UI.py`: test the CML and UI lightly to ensure they are working - `test_prediction.py`: ensures predictions are correctly generated by the program Now, I want to ensure all of my test cases I write are passing and I have a high success rate. However, in addition to ensuring what is being tested is working properly, I want to ensure that almost everything that can be tested is tested. This will ensure the tool as a whole functions well. With this, I want to make sure that my tool's code test coverage is high. Code coverage is a measure that analyzes how much source code of a program is executed (or tested) when a particular test suite runs. There is a library called [Coverage](https://coverage.readthedocs.io/en/coverage-5.5/) that will automatically measure the code coverage of your Python test suite, making it perfect for my project. With this, I will be using a version of the library that integrates with Pytest called [pytest-cov](https://pypi.org/project/pytest-cov/).
non_priority
finalize testing suite as i near the completion of the project source code and the implementation of the prediction model this code will need to be extensively tested to ensure it functions correctly these test cases should help ensure a confidence in program correctness to do this test cases should be written for every part of the program these test files should include test scraper py ensure data is scraped and that it is in the correct format test data cleaner py ensure data is properly converted test ui py test the cml and ui lightly to ensure they are working test prediction py ensures predictions are correctly generated by the program now i want to ensure all of my test cases i write are passing and i have a high success rate however in addition to ensuring what is being tested is working properly i want to ensure that almost everything that can be tested is tested this will ensure the tool as a whole functions well with this i want to make sure that my tool s code test coverage is high code coverage is a measure that analyzes how much source code of a program is executed or tested when a particular test suite runs there is a library called that will automatically measure the code coverage of your python test suite making it perfect for my project with this i will be using a version of the library that integrates with pytest called
0
6,205
9,107,428,823
IssuesEvent
2019-02-21 04:21:48
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Getting a parse error with no result in search results window
bug parse-tree-processing
Version 2.2.6672.28001 OS: Microsoft Windows NT 10.0.16299.0, x64 Host Product: Microsoft Office 2013 x86 Host Version: 15.0.5045.1000 Host Executable: MSACCESS.EXE This is persisting after fixing the error and getting a successful compile. [RubberduckLog.txt](https://github.com/rubberduck-vba/Rubberduck/files/2506927/RubberduckLog.txt)
1.0
Getting a parse error with no result in search results window - Version 2.2.6672.28001 OS: Microsoft Windows NT 10.0.16299.0, x64 Host Product: Microsoft Office 2013 x86 Host Version: 15.0.5045.1000 Host Executable: MSACCESS.EXE This is persisting after fixing the error and getting a successful compile. [RubberduckLog.txt](https://github.com/rubberduck-vba/Rubberduck/files/2506927/RubberduckLog.txt)
non_priority
getting a parse error with no result in search results window version os microsoft windows nt host product microsoft office host version host executable msaccess exe this is persisting after fixing the error and getting a successful compile
0
232,101
25,564,967,337
IssuesEvent
2022-11-30 13:40:42
jtimberlake/pacbot
https://api.github.com/repos/jtimberlake/pacbot
opened
CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz
security vulnerability
## CVE-2022-38900 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary> <p>A better decodeURIComponent</p> <p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p> <p>Path to dependency file: /webapp/package.json</p> <p>Path to vulnerable library: /webapp/node_modules/decode-uri-component/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - stylus-0.54.8.tgz - css-parse-2.0.0.tgz - css-2.2.4.tgz - source-map-resolve-0.5.3.tgz - :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/pacbot/commit/5a25dc15882324583860630e773af2e2647dd5fa">5a25dc15882324583860630e773af2e2647dd5fa</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS. <p>Publish Date: 2022-11-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz - ## CVE-2022-38900 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary> <p>A better decodeURIComponent</p> <p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p> <p>Path to dependency file: /webapp/package.json</p> <p>Path to vulnerable library: /webapp/node_modules/decode-uri-component/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - stylus-0.54.8.tgz - css-parse-2.0.0.tgz - css-2.2.4.tgz - source-map-resolve-0.5.3.tgz - :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jtimberlake/pacbot/commit/5a25dc15882324583860630e773af2e2647dd5fa">5a25dc15882324583860630e773af2e2647dd5fa</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS. <p>Publish Date: 2022-11-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_priority
cve high detected in decode uri component tgz cve high severity vulnerability vulnerable library decode uri component tgz a better decodeuricomponent library home page a href path to dependency file webapp package json path to vulnerable library webapp node modules decode uri component package json dependency hierarchy cli tgz root library stylus tgz css parse tgz css tgz source map resolve tgz x decode uri component tgz vulnerable library found in head commit a href found in base branch master vulnerability details decode uri component is vulnerable to improper input validation resulting in dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
253,798
21,705,656,287
IssuesEvent
2022-05-10 09:21:29
FutureNorthants/VirtualWorker
https://api.github.com/repos/FutureNorthants/VirtualWorker
closed
Missing 'Comment - Sent to Email' Field
enhancement North needs retesting
**Bug description** A clear and concise description of what the bug is, what is happening and what the impact is. NORTH - the comments field which tells an advisor where the case has been forwarded to is not appearing in a certain scenario. The scenario is that a case is created in CXM, the BOT identifies the wrong sovereign, the Hub manually change the sovereign, the customer sends an update, the BOT sends the new update to the relevant sovereign (last chosen one on CXM) however the Comments field is not present to prove where it has been sent. The impact is that staff are having to phone sovereigns to see if cases have been forwarded due to the fact that bug #74 has meant they have not been forwarding, so when the fix for #74 is in place we need to have visual proof of where the case has been sent or not to prevent manual checking. **Reference number(s)** EMN000535 in LIVE If you can - please attempt to replicate the error(s) in test and provide the TEST reference number. **Bug reproduction** Steps to reproduce the behaviour: 1. Complete a Contact Us form 2. From the Hub change the Sovereign in CXM 3. Send a customer update on the case 4. See in the case on CXM that the top Comments field, which shows where the case has been forwarded to, is not present. **Expected behaviour** 1. Customer to complete Contact us form 2. If BOT chooses wrong sovereign (A) the Hub wil reallocate to the correct sovereign (B) 3. Customer sends an update to the case 4. BOT sends the update to the latest sovereign (B) 5. The comments field to be displayed showing the trail for point 4 in CXM so that advisors can see where the update has been sent to. (this bit is missing at the moment) **Screenshots** There is one on unitary chat but i can't copy and paste from there and my snipping tool won't open either **Additional comments** This will impact us monitoring the Bug fix for #74
1.0
Missing 'Comment - Sent to Email' Field - **Bug description** A clear and concise description of what the bug is, what is happening and what the impact is. NORTH - the comments field which tells an advisor where the case has been forwarded to is not appearing in a certain scenario. The scenario is that a case is created in CXM, the BOT identifies the wrong sovereign, the Hub manually change the sovereign, the customer sends an update, the BOT sends the new update to the relevant sovereign (last chosen one on CXM) however the Comments field is not present to prove where it has been sent. The impact is that staff are having to phone sovereigns to see if cases have been forwarded due to the fact that bug #74 has meant they have not been forwarding, so when the fix for #74 is in place we need to have visual proof of where the case has been sent or not to prevent manual checking. **Reference number(s)** EMN000535 in LIVE If you can - please attempt to replicate the error(s) in test and provide the TEST reference number. **Bug reproduction** Steps to reproduce the behaviour: 1. Complete a Contact Us form 2. From the Hub change the Sovereign in CXM 3. Send a customer update on the case 4. See in the case on CXM that the top Comments field, which shows where the case has been forwarded to, is not present. **Expected behaviour** 1. Customer to complete Contact us form 2. If BOT chooses wrong sovereign (A) the Hub wil reallocate to the correct sovereign (B) 3. Customer sends an update to the case 4. BOT sends the update to the latest sovereign (B) 5. The comments field to be displayed showing the trail for point 4 in CXM so that advisors can see where the update has been sent to. (this bit is missing at the moment) **Screenshots** There is one on unitary chat but i can't copy and paste from there and my snipping tool won't open either **Additional comments** This will impact us monitoring the Bug fix for #74
non_priority
missing comment sent to email field bug description a clear and concise description of what the bug is what is happening and what the impact is north the comments field which tells an advisor where the case has been forwarded to is not appearing in a certain scenario the scenario is that a case is created in cxm the bot identifies the wrong sovereign the hub manually change the sovereign the customer sends an update the bot sends the new update to the relevant sovereign last chosen one on cxm however the comments field is not present to prove where it has been sent the impact is that staff are having to phone sovereigns to see if cases have been forwarded due to the fact that bug has meant they have not been forwarding so when the fix for is in place we need to have visual proof of where the case has been sent or not to prevent manual checking reference number s in live if you can please attempt to replicate the error s in test and provide the test reference number bug reproduction steps to reproduce the behaviour complete a contact us form from the hub change the sovereign in cxm send a customer update on the case see in the case on cxm that the top comments field which shows where the case has been forwarded to is not present expected behaviour customer to complete contact us form if bot chooses wrong sovereign a the hub wil reallocate to the correct sovereign b customer sends an update to the case bot sends the update to the latest sovereign b the comments field to be displayed showing the trail for point in cxm so that advisors can see where the update has been sent to this bit is missing at the moment screenshots there is one on unitary chat but i can t copy and paste from there and my snipping tool won t open either additional comments this will impact us monitoring the bug fix for
0
223,986
24,760,213,277
IssuesEvent
2022-10-21 22:40:09
TIBCOSoftware/js-workshops
https://api.github.com/repos/TIBCOSoftware/js-workshops
opened
CVE-2022-37598 (High) detected in uglify-js-3.6.0.tgz
security vulnerability
## CVE-2022-37598 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-3.6.0.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz</a></p> <p>Path to dependency file: /UrbanRoast-tutorial/resources/d3-funnel-master/package.json</p> <p>Path to vulnerable library: /UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/uglify-js/package.json</p> <p> Dependency Hierarchy: - gulp-uglify-3.0.2.tgz (Root Library) - :x: **uglify-js-3.6.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TIBCOSoftware/js-workshops/commits/8a027629e1d1d82ce0fb49a090c585960d0967cd">8a027629e1d1d82ce0fb49a090c585960d0967cd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js. <p>Publish Date: 2022-10-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37598>CVE-2022-37598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-20</p> <p>Fix Resolution: uglify-js - 3.13.10</p> </p> </details> <p></p>
True
CVE-2022-37598 (High) detected in uglify-js-3.6.0.tgz - ## CVE-2022-37598 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-3.6.0.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz</a></p> <p>Path to dependency file: /UrbanRoast-tutorial/resources/d3-funnel-master/package.json</p> <p>Path to vulnerable library: /UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/uglify-js/package.json</p> <p> Dependency Hierarchy: - gulp-uglify-3.0.2.tgz (Root Library) - :x: **uglify-js-3.6.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/TIBCOSoftware/js-workshops/commits/8a027629e1d1d82ce0fb49a090c585960d0967cd">8a027629e1d1d82ce0fb49a090c585960d0967cd</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js. <p>Publish Date: 2022-10-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37598>CVE-2022-37598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-20</p> <p>Fix Resolution: uglify-js - 3.13.10</p> </p> </details> <p></p>
non_priority
cve high detected in uglify js tgz cve high severity vulnerability vulnerable library uglify js tgz javascript parser mangler compressor and beautifier toolkit library home page a href path to dependency file urbanroast tutorial resources funnel master package json path to vulnerable library urbanroast tutorial resources funnel master node modules uglify js package json dependency hierarchy gulp uglify tgz root library x uglify js tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability in function defnode in ast js in mishoo uglifyjs via the name variable in ast js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution uglify js
0
351,031
25,010,838,218
IssuesEvent
2022-11-03 15:08:41
AY2223S1-CS2103T-W16-1/tp
https://api.github.com/repos/AY2223S1-CS2103T-W16-1/tp
closed
[PE-D][Tester B] Finding by a date range can return other irrelevant search results
type.Bug documentation severity.Low good to fix
Using this sample data: ![image.png](https://raw.githubusercontent.com/NicholasTYD/ped/main/files/fa9e12e6-6952-4965-81a3-964f1b903b2f.png) After typing `find 01/01/23 - 07/01/23`: ![image.png](https://raw.githubusercontent.com/NicholasTYD/ped/main/files/55ed9356-53e2-4ad1-9104-1e243e4fa250.png) This might be because the first search result has a “-“ as a request under room status, which the command detects. However, novice users might be confused by this. <!--session: 1666944239122-5fa2ca7e-3be6-428b-8e43-2cc245ff74fe--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FunctionalityBug` `severity.Low` original: NicholasTYD/ped#7
1.0
[PE-D][Tester B] Finding by a date range can return other irrelevant search results - Using this sample data: ![image.png](https://raw.githubusercontent.com/NicholasTYD/ped/main/files/fa9e12e6-6952-4965-81a3-964f1b903b2f.png) After typing `find 01/01/23 - 07/01/23`: ![image.png](https://raw.githubusercontent.com/NicholasTYD/ped/main/files/55ed9356-53e2-4ad1-9104-1e243e4fa250.png) This might be because the first search result has a “-“ as a request under room status, which the command detects. However, novice users might be confused by this. <!--session: 1666944239122-5fa2ca7e-3be6-428b-8e43-2cc245ff74fe--> <!--Version: Web v3.4.4--> ------------- Labels: `type.FunctionalityBug` `severity.Low` original: NicholasTYD/ped#7
non_priority
finding by a date range can return other irrelevant search results using this sample data after typing find this might be because the first search result has a “ “ as a request under room status which the command detects however novice users might be confused by this labels type functionalitybug severity low original nicholastyd ped
0
179,278
14,700,231,926
IssuesEvent
2021-01-04 09:53:31
superwerker/superwerker
https://api.github.com/repos/superwerker/superwerker
opened
Prepare for AWS Quick Starts
documentation
## Tasks - [ ] Logo (Quick Start Owner _This Quick Start was developed by AWS solutions architects._ e.g.) - [ ] High Level Briefing - [ ] Documentation (_Deployment Guide_) - Website ( https://aws-quickstart.github.io/quickstart-dotnetcore-cicd/ ) - PDF ( https://aws-quickstart.s3.amazonaws.com/quickstart-trek10-serverless-enterprise-cicd/doc/serverless-cicd-for-the-enterprise-on-the-aws-cloud.pdf ) - [ ] Architecture Diagramm ## Topics - What you'll build - How to deploy - Cost and licenses - Resources ## Examples - https://aws.amazon.com/quickstart/architecture/serverless-cicd-for-enterprise/ - https://aws.amazon.com/quickstart/architecture/eks-snyk/ - https://aws.amazon.com/quickstart/architecture/terraform-modules-on-aws/ - https://aws.amazon.com/quickstart/architecture/ibm-maximo/
1.0
Prepare for AWS Quick Starts - ## Tasks - [ ] Logo (Quick Start Owner _This Quick Start was developed by AWS solutions architects._ e.g.) - [ ] High Level Briefing - [ ] Documentation (_Deployment Guide_) - Website ( https://aws-quickstart.github.io/quickstart-dotnetcore-cicd/ ) - PDF ( https://aws-quickstart.s3.amazonaws.com/quickstart-trek10-serverless-enterprise-cicd/doc/serverless-cicd-for-the-enterprise-on-the-aws-cloud.pdf ) - [ ] Architecture Diagramm ## Topics - What you'll build - How to deploy - Cost and licenses - Resources ## Examples - https://aws.amazon.com/quickstart/architecture/serverless-cicd-for-enterprise/ - https://aws.amazon.com/quickstart/architecture/eks-snyk/ - https://aws.amazon.com/quickstart/architecture/terraform-modules-on-aws/ - https://aws.amazon.com/quickstart/architecture/ibm-maximo/
non_priority
prepare for aws quick starts tasks logo quick start owner this quick start was developed by aws solutions architects e g high level briefing documentation deployment guide website pdf architecture diagramm topics what you ll build how to deploy cost and licenses resources examples
0
72,924
8,790,572,890
IssuesEvent
2018-12-21 09:32:19
nextcloud/server
https://api.github.com/repos/nextcloud/server
closed
App icon disappears for overflowing settings page names
3. to review bug design stale
### Steps to reproduce 1. Create app with a rather long settings page name 2. Apply app-icon to app 3. App icon will not be shown in e.g. /index.php/settings/user ### Expected behaviour Settings page name gets shortened, but app icon stays visible. ### Actual behaviour Settings page name is shortened and app icon disappears. ### Nextcloud version 13.0.0RC1 ![bildschirmfoto 2018-01-18 um 14 58 46](https://user-images.githubusercontent.com/4127167/35101735-d97c7868-fc60-11e7-8e4e-c8878194e7fc.png) This is an issue that probably needs refactoring of the HTML structure. We have this ``` <a href="/nextcloud/index.php/settings/admin/my_app"> <img alt="" src="/nextcloud/apps/my_app/img/app-dark.svg"> <span>A very long app settings page name that is too long</span> </a> ``` And because the a-tag gets a padding at the position where the img is placed, it will be hidden, once the span-text overflows, because of overflow: hidden. At this moment I have no simple idea how to solve this, maybe someone else has.
1.0
App icon disappears for overflowing settings page names - ### Steps to reproduce 1. Create app with a rather long settings page name 2. Apply app-icon to app 3. App icon will not be shown in e.g. /index.php/settings/user ### Expected behaviour Settings page name gets shortened, but app icon stays visible. ### Actual behaviour Settings page name is shortened and app icon disappears. ### Nextcloud version 13.0.0RC1 ![bildschirmfoto 2018-01-18 um 14 58 46](https://user-images.githubusercontent.com/4127167/35101735-d97c7868-fc60-11e7-8e4e-c8878194e7fc.png) This is an issue that probably needs refactoring of the HTML structure. We have this ``` <a href="/nextcloud/index.php/settings/admin/my_app"> <img alt="" src="/nextcloud/apps/my_app/img/app-dark.svg"> <span>A very long app settings page name that is too long</span> </a> ``` And because the a-tag gets a padding at the position where the img is placed, it will be hidden, once the span-text overflows, because of overflow: hidden. At this moment I have no simple idea how to solve this, maybe someone else has.
non_priority
app icon disappears for overflowing settings page names steps to reproduce create app with a rather long settings page name apply app icon to app app icon will not be shown in e g index php settings user expected behaviour settings page name gets shortened but app icon stays visible actual behaviour settings page name is shortened and app icon disappears nextcloud version this is an issue that probably needs refactoring of the html structure we have this a very long app settings page name that is too long and because the a tag gets a padding at the position where the img is placed it will be hidden once the span text overflows because of overflow hidden at this moment i have no simple idea how to solve this maybe someone else has
0
311,966
23,411,977,491
IssuesEvent
2022-08-12 18:36:10
bcbi/PreprocessMD.jl
https://api.github.com/repos/bcbi/PreprocessMD.jl
closed
Examples should display function inputs
documentation good first issue
The docstring examples for each function display the output DataFrame. The input DataFrames should also be displayed so that the user can see how the output relates to the input, especially for `pivot()`.
1.0
Examples should display function inputs - The docstring examples for each function display the output DataFrame. The input DataFrames should also be displayed so that the user can see how the output relates to the input, especially for `pivot()`.
non_priority
examples should display function inputs the docstring examples for each function display the output dataframe the input dataframes should also be displayed so that the user can see how the output relates to the input especially for pivot
0
784,904
27,588,995,226
IssuesEvent
2023-03-08 22:20:11
AdrKacz/super-duper-guacamole
https://api.github.com/repos/AdrKacz/super-duper-guacamole
opened
❓ Comment profiter du cercles d'amis des utilisateurs
question priority:P1
# What do you need to discuss? Chaque utilisateurs à des amis, qui pourraient être des clients potentiels, ce qui rendrait à terme l'application viral. # Why is that important? Si on pouvait faire circuler l'application entre les cercles d'amis, cela augmenterait le nombre de téléchargement. Hors, comme on travaille un peu sur un réseau social, plus il y a de monde plus c'est fun. # What are the element we will need to discuss it? Quand il est question de _referal_ chaque app à sa stratégie, par exemple Revolut te paie 60€ par utilisateur apporté et Snapshat accéde a tes contacts pour te faciliter la communication avec eux.
1.0
❓ Comment profiter du cercles d'amis des utilisateurs - # What do you need to discuss? Chaque utilisateurs à des amis, qui pourraient être des clients potentiels, ce qui rendrait à terme l'application viral. # Why is that important? Si on pouvait faire circuler l'application entre les cercles d'amis, cela augmenterait le nombre de téléchargement. Hors, comme on travaille un peu sur un réseau social, plus il y a de monde plus c'est fun. # What are the element we will need to discuss it? Quand il est question de _referal_ chaque app à sa stratégie, par exemple Revolut te paie 60€ par utilisateur apporté et Snapshat accéde a tes contacts pour te faciliter la communication avec eux.
priority
❓ comment profiter du cercles d amis des utilisateurs what do you need to discuss chaque utilisateurs à des amis qui pourraient être des clients potentiels ce qui rendrait à terme l application viral why is that important si on pouvait faire circuler l application entre les cercles d amis cela augmenterait le nombre de téléchargement hors comme on travaille un peu sur un réseau social plus il y a de monde plus c est fun what are the element we will need to discuss it quand il est question de referal chaque app à sa stratégie par exemple revolut te paie € par utilisateur apporté et snapshat accéde a tes contacts pour te faciliter la communication avec eux
1
16,976
3,587,160,365
IssuesEvent
2016-01-30 04:04:50
tgstation/-tg-station
https://api.github.com/repos/tgstation/-tg-station
closed
You can teleport via mulebots
Bug In Game Exploit Tested / Reproduced
rev: 0debb6cb68829429f4d53f0e52e30dcbc0abcbf6 H2Repr0duce420noblazetelep0rt: 0. Turn off 'return home after delivery' on mulebot 1. Have your mulebot deliver you somewhere 2. Try to climb back onto the mule 3. Get blocked by flaps, you don't end up on the mulebot 4. Send mulebot home/somewhere else 5. Wait 6. Get teleported to the mule's latest drop point
1.0
You can teleport via mulebots - rev: 0debb6cb68829429f4d53f0e52e30dcbc0abcbf6 H2Repr0duce420noblazetelep0rt: 0. Turn off 'return home after delivery' on mulebot 1. Have your mulebot deliver you somewhere 2. Try to climb back onto the mule 3. Get blocked by flaps, you don't end up on the mulebot 4. Send mulebot home/somewhere else 5. Wait 6. Get teleported to the mule's latest drop point
non_priority
you can teleport via mulebots rev turn off return home after delivery on mulebot have your mulebot deliver you somewhere try to climb back onto the mule get blocked by flaps you don t end up on the mulebot send mulebot home somewhere else wait get teleported to the mule s latest drop point
0