Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
410,102 | 11,983,209,292 | IssuesEvent | 2020-04-07 14:07:21 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | closed | `ShowMoreCard` button used in developer reply has the wrong background color | component: add-on ratings contrib: welcome priority: p3 | ### Describe the problem and steps to reproduce it:
1. Go to https://addons-dev.allizom.org/en-US/firefox/addon/search_by_image/reviews/?src=recommended_fallback
2. Observe the "developer response" in the middle of the page
### What happened?
The `ShowMoreCard` displays a "read more" button with a white background:
<img width="1280" alt="Screen Shot 2020-03-31 at 14 38 17" src="https://user-images.githubusercontent.com/217628/78027415-9ceaae80-735d-11ea-8527-d8c26ea60816.png">
### What did you expect to happen?
The button background color should be blue-ish. | 1.0 | `ShowMoreCard` button used in developer reply has the wrong background color - ### Describe the problem and steps to reproduce it:
1. Go to https://addons-dev.allizom.org/en-US/firefox/addon/search_by_image/reviews/?src=recommended_fallback
2. Observe the "developer response" in the middle of the page
### What happened?
The `ShowMoreCard` displays a "read more" button with a white background:
<img width="1280" alt="Screen Shot 2020-03-31 at 14 38 17" src="https://user-images.githubusercontent.com/217628/78027415-9ceaae80-735d-11ea-8527-d8c26ea60816.png">
### What did you expect to happen?
The button background color should be blue-ish. | priority | showmorecard button used in developer reply has the wrong background color describe the problem and steps to reproduce it go to observe the developer response in the middle of the page what happened the showmorecard displays a read more button with a white background img width alt screen shot at src what did you expect to happen the button background color should be blue ish | 1 |
133,234 | 18,285,818,279 | IssuesEvent | 2021-10-05 10:08:20 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [Security Solution][Timeline]On Clicking + Add field user focus moved from the Field name to the value | bug triage_needed Team: SecuritySolution | **Describe the bug**
Timeline: On Clicking + Add field user focus moved from the Field name to the value
**Build Details**
`Version: 7.15.0`
**Steps**
- Login to kibana
- Create timeline
- Click on + Add field in the timeline drag area
- click on drop down and don't select any value from this
- Observed that user focus is moved to the value field. This makes it a pain to manually add fields because you have to stop and wait for it to do the focus change before you can start typing, otherwise it will move when you are in the middle of typing the field name
**Actual**
Timeline: On Clicking + Add field user focus moved from the Field name to the value
**Snap Shoot**

**Screen-Cast**
https://elastic.slack.com/archives/C012TJZ72J1/p1633426485272700
**logs**
- N/A | True | [Security Solution][Timeline]On Clicking + Add field user focus moved from the Field name to the value - **Describe the bug**
Timeline: On Clicking + Add field user focus moved from the Field name to the value
**Build Details**
`Version: 7.15.0`
**Steps**
- Login to kibana
- Create timeline
- Click on + Add field in the timeline drag area
- click on drop down and don't select any value from this
- Observed that user focus is moved to the value field. This makes it a pain to manually add fields because you have to stop and wait for it to do the focus change before you can start typing, otherwise it will move when you are in the middle of typing the field name
**Actual**
Timeline: On Clicking + Add field user focus moved from the Field name to the value
**Snap Shoot**

**Screen-Cast**
https://elastic.slack.com/archives/C012TJZ72J1/p1633426485272700
**logs**
- N/A | non_priority | on clicking add field user focus moved from the field name to the value describe the bug timeline on clicking add field user focus moved from the field name to the value build details version steps login to kibana create timeline click on add field in the timeline drag area click on drop down and don t select any value from this observed that user focus is moved to the value field this makes it a pain to manually add fields because you have to stop and wait for it to do the focus change before you can start typing otherwise it will move when you are in the middle of typing the field name actual timeline on clicking add field user focus moved from the field name to the value snap shoot screen cast logs n a | 0 |
562,345 | 16,657,647,556 | IssuesEvent | 2021-06-05 20:31:04 | dodona-edu/dodona | https://api.github.com/repos/dodona-edu/dodona | closed | HTML5 video tag does not support skipping forward/backward | bug low priority student | Video fragments embedded using the HTML5 tag `video` cannot be skipped forward nor backward.
Example: [exercise description](https://dodona.ugent.be/nl/courses/172/series/3251/activities/357352739/)
OS: Mac OS X
Browser: Google Chrome
Observed by: @toonijn @pverscha | 1.0 | HTML5 video tag does not support skipping forward/backward - Video fragments embedded using the HTML5 tag `video` cannot be skipped forward nor backward.
Example: [exercise description](https://dodona.ugent.be/nl/courses/172/series/3251/activities/357352739/)
OS: Mac OS X
Browser: Google Chrome
Observed by: @toonijn @pverscha | priority | video tag does not support skipping forward backward video fragments embedded using the tag video cannot be skipped forward nor backward example os mac os x browser google chrome observed by toonijn pverscha | 1 |
52,057 | 3,020,503,464 | IssuesEvent | 2015-07-31 08:25:44 | 52North/SOS | https://api.github.com/repos/52North/SOS | closed | Client problem with feature capabilities in DescribeSensor response with SensorML 1.0.1 encoded procedure | bug enhancement medium priority | The SweText of the feature capabilities of a DescribeSensor response procedure encoded in SensorML 1.0.1 does not contain the definition which is used by the Sensor Web Client/REST-API proxy to identify featureOfInterest.
Problem: The merging process replaces the field name. | 1.0 | Client problem with feature capabilities in DescribeSensor response with SensorML 1.0.1 encoded procedure - The SweText of the feature capabilities of a DescribeSensor response procedure encoded in SensorML 1.0.1 does not contain the definition which is used by the Sensor Web Client/REST-API proxy to identify featureOfInterest.
Problem: The merging process replaces the field name. | priority | client problem with feature capabilities in describesensor response with sensorml encoded procedure the swetext of the feature capabilities of a describesensor response procedure encoded in sensorml does not contain the definition which is used by the sensor web client rest api proxy to identify featureofinterest problem the merging process replaces the field name | 1 |
773,206 | 27,149,702,309 | IssuesEvent | 2023-02-16 23:31:46 | internetarchive/openlibrary | https://api.github.com/repos/internetarchive/openlibrary | opened | Migrate to Vue 3 | Type: Feature Request Priority: 3 1-off tasks Type: Epic Lead: @jimchamp | ### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
Open Library maintains a small number of web components that were created with Vue 2.
Vue 2 will reach end of life on 31 December 2023, so we'll have to migrate to Vue 3 in the near future.
This also give us the opportunity to try using Vite for our component builds. We currently use Vue CLI, and this is one of the slower steps of our build process. More importantly, Vue CLI is now in maintenance mode, and eventually will be no longer supported.
This epic can be considered complete when the following tasks are finished:
- [ ] Upgrade to Vue 3
- [ ] Test new build process with Vite
- [ ] Migrate all of the following components:
- [x] AuthorIdentifiers
- [x] HelloWorld
- [ ] LibraryExplorer
- [ ] MergeUI
- [x] ObservationForm
- [ ] Update our Vue [documentation](https://github.com/internetarchive/openlibrary/wiki/Using-Vue)
- [ ] Add link to this page to wiki sidebar
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
<!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? -->
### Additional context
<!-- Add any other context or screenshots about the feature request here. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
| 1.0 | Migrate to Vue 3 - ### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
Open Library maintains a small number of web components that were created with Vue 2.
Vue 2 will reach end of life on 31 December 2023, so we'll have to migrate to Vue 3 in the near future.
This also give us the opportunity to try using Vite for our component builds. We currently use Vue CLI, and this is one of the slower steps of our build process. More importantly, Vue CLI is now in maintenance mode, and eventually will be no longer supported.
This epic can be considered complete when the following tasks are finished:
- [ ] Upgrade to Vue 3
- [ ] Test new build process with Vite
- [ ] Migrate all of the following components:
- [x] AuthorIdentifiers
- [x] HelloWorld
- [ ] LibraryExplorer
- [ ] MergeUI
- [x] ObservationForm
- [ ] Update our Vue [documentation](https://github.com/internetarchive/openlibrary/wiki/Using-Vue)
- [ ] Add link to this page to wiki sidebar
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
<!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? -->
### Additional context
<!-- Add any other context or screenshots about the feature request here. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
| priority | migrate to vue describe the problem that you d like solved open library maintains a small number of web components that were created with vue vue will reach end of life on december so we ll have to migrate to vue in the near future this also give us the opportunity to try using vite for our component builds we currently use vue cli and this is one of the slower steps of our build process more importantly vue cli is now in maintenance mode and eventually will be no longer supported this epic can be considered complete when the following tasks are finished upgrade to vue test new build process with vite migrate all of the following components authoridentifiers helloworld libraryexplorer mergeui observationform update our vue add link to this page to wiki sidebar proposal constraints additional context stakeholders | 1 |
112,988 | 9,607,837,253 | IssuesEvent | 2019-05-11 22:53:01 | dimitri/pgloader | https://api.github.com/repos/dimitri/pgloader | closed | AWS RDS to AWS VM | How To... Question Needs more testing / information | Thanks for contributing to [pgloader](https://pgloader.io) by reporting an
issue! Reporting an issue is the only way we can solve problems, fix bugs,
and improve both the software and its user experience in general.
The best bug reports follow those 3 simple steps:
1. show what you did,
2. show the result you got,
3. explain how the result is not what you expected.
In the case of pgloader, here's the information I will need to read in your
bug report. Having all of this is a big help, and often means the bug you
reported can be fixed very efficiently as soon as I get to it.
Please provide the following information:
<!-- delete text above this line -->
- [ ] pgloader --version
```
<fill pgloader version here>
```
- [ ] did you test a fresh compile from the source tree?
Compiling pgloader from sources is documented in the
[README](https://github.com/dimitri/pgloader#build-from-sources), it's
easy to do, and if patches are to be made to fix your bug, you're going
to have to build from sources to get the fix anyway…
- [ ] did you search for other similar issues?
- [ ] how can I reproduce the bug?
Incude a self-contained pgloader command file.
If you're loading from a database, consider attaching a database dump to
your issue. For MySQL, use `mysqldump`. For SQLite, just send over your
source file, that's easy. Maybe be the one with your production data, of
course, the one with just the sample of data that allows me to reproduce
your bug.
When using a proprietary database system as a source, consider creating
a sample database on some Cloud service or somewhere you can then give
me access to, and see my email address on my GitHub profile to send me
the credentials. Still open a public issue for tracking and as
documentation for other users.
```
--
-- EDIT THIS FILE TO MATCH YOUR BUG REPORT
--
LOAD CSV
FROM INLINE with encoding 'ascii'
INTO postgresql:///pgloader
TARGET TABLE jordane
WITH truncate,
fields terminated by '|',
fields not enclosed,
fields escaped by backslash-quote
SET work_mem to '128MB',
standard_conforming_strings to 'on'
BEFORE LOAD DO
$$ drop table if exists jordane; $$,
$$ CREATE TABLE jordane
(
"NOM" character(20),
"PRENOM" character(20)
)
$$;
BORDET|Jordane
BORDET|Audrey
LASTNAME|"opening quote
BONNIER|testprenombe~aucouptroplong
JOURDAIN|héhé¶
```
- [ ] pgloader output you obtain
```
PASTE HERE THE OUTPUT OF THE PGLOADER COMMAND
```
- [ ] data that is being loaded, if relevant
```
PASTE HERE THE DATA THAT HAS BEEN LOADED
```
- [ ] How the data is different from what you expected, if relevant
| 1.0 | AWS RDS to AWS VM - Thanks for contributing to [pgloader](https://pgloader.io) by reporting an
issue! Reporting an issue is the only way we can solve problems, fix bugs,
and improve both the software and its user experience in general.
The best bug reports follow those 3 simple steps:
1. show what you did,
2. show the result you got,
3. explain how the result is not what you expected.
In the case of pgloader, here's the information I will need to read in your
bug report. Having all of this is a big help, and often means the bug you
reported can be fixed very efficiently as soon as I get to it.
Please provide the following information:
<!-- delete text above this line -->
- [ ] pgloader --version
```
<fill pgloader version here>
```
- [ ] did you test a fresh compile from the source tree?
Compiling pgloader from sources is documented in the
[README](https://github.com/dimitri/pgloader#build-from-sources), it's
easy to do, and if patches are to be made to fix your bug, you're going
to have to build from sources to get the fix anyway…
- [ ] did you search for other similar issues?
- [ ] how can I reproduce the bug?
Incude a self-contained pgloader command file.
If you're loading from a database, consider attaching a database dump to
your issue. For MySQL, use `mysqldump`. For SQLite, just send over your
source file, that's easy. Maybe be the one with your production data, of
course, the one with just the sample of data that allows me to reproduce
your bug.
When using a proprietary database system as a source, consider creating
a sample database on some Cloud service or somewhere you can then give
me access to, and see my email address on my GitHub profile to send me
the credentials. Still open a public issue for tracking and as
documentation for other users.
```
--
-- EDIT THIS FILE TO MATCH YOUR BUG REPORT
--
LOAD CSV
FROM INLINE with encoding 'ascii'
INTO postgresql:///pgloader
TARGET TABLE jordane
WITH truncate,
fields terminated by '|',
fields not enclosed,
fields escaped by backslash-quote
SET work_mem to '128MB',
standard_conforming_strings to 'on'
BEFORE LOAD DO
$$ drop table if exists jordane; $$,
$$ CREATE TABLE jordane
(
"NOM" character(20),
"PRENOM" character(20)
)
$$;
BORDET|Jordane
BORDET|Audrey
LASTNAME|"opening quote
BONNIER|testprenombe~aucouptroplong
JOURDAIN|héhé¶
```
- [ ] pgloader output you obtain
```
PASTE HERE THE OUTPUT OF THE PGLOADER COMMAND
```
- [ ] data that is being loaded, if relevant
```
PASTE HERE THE DATA THAT HAS BEEN LOADED
```
- [ ] How the data is different from what you expected, if relevant
| non_priority | aws rds to aws vm thanks for contributing to by reporting an issue reporting an issue is the only way we can solve problems fix bugs and improve both the software and its user experience in general the best bug reports follow those simple steps show what you did show the result you got explain how the result is not what you expected in the case of pgloader here s the information i will need to read in your bug report having all of this is a big help and often means the bug you reported can be fixed very efficiently as soon as i get to it please provide the following information pgloader version did you test a fresh compile from the source tree compiling pgloader from sources is documented in the it s easy to do and if patches are to be made to fix your bug you re going to have to build from sources to get the fix anyway… did you search for other similar issues how can i reproduce the bug incude a self contained pgloader command file if you re loading from a database consider attaching a database dump to your issue for mysql use mysqldump for sqlite just send over your source file that s easy maybe be the one with your production data of course the one with just the sample of data that allows me to reproduce your bug when using a proprietary database system as a source consider creating a sample database on some cloud service or somewhere you can then give me access to and see my email address on my github profile to send me the credentials still open a public issue for tracking and as documentation for other users edit this file to match your bug report load csv from inline with encoding ascii into postgresql pgloader target table jordane with truncate fields terminated by fields not enclosed fields escaped by backslash quote set work mem to standard conforming strings to on before load do drop table if exists jordane create table jordane nom character prenom character bordet jordane bordet audrey lastname opening quote bonnier testprenombe aucouptroplong jourdain héhé¶ pgloader output you obtain paste here the output of the pgloader command data that is being loaded if relevant paste here the data that has been loaded how the data is different from what you expected if relevant | 0 |
751,168 | 26,231,736,536 | IssuesEvent | 2023-01-05 01:14:06 | canonical/cn.ubuntu.com | https://api.github.com/repos/canonical/cn.ubuntu.com | closed | Empty links on Ubuntu Core特点概览 | Priority: High | I found 4 links not working on https://cn.ubuntu.com/internet-of-things/core/features
The current links and correct links are:
- OTA 更新
- https://cn.ubuntu.com/internet-of-things/features/ota-updates >> https://cn.ubuntu.com/internet-of-things/core/features/ota-updates
- 安全启动
- https://cn.ubuntu.com/internet-of-things/features/secure-boot >> https://cn.ubuntu.com/internet-of-things/core/features/secure-boot
- 全盘加密
- https://cn.ubuntu.com/internet-of-things/features/full-disk-encryption >> https://cn.ubuntu.com/internet-of-things/core/features/full-disk-encryption
- 恢复模式
- https://cn.ubuntu.com/internet-of-things/features/recovery >> https://cn.ubuntu.com/internet-of-things/core/features/recovery
thanks


| 1.0 | Empty links on Ubuntu Core特点概览 - I found 4 links not working on https://cn.ubuntu.com/internet-of-things/core/features
The current links and correct links are:
- OTA 更新
- https://cn.ubuntu.com/internet-of-things/features/ota-updates >> https://cn.ubuntu.com/internet-of-things/core/features/ota-updates
- 安全启动
- https://cn.ubuntu.com/internet-of-things/features/secure-boot >> https://cn.ubuntu.com/internet-of-things/core/features/secure-boot
- 全盘加密
- https://cn.ubuntu.com/internet-of-things/features/full-disk-encryption >> https://cn.ubuntu.com/internet-of-things/core/features/full-disk-encryption
- 恢复模式
- https://cn.ubuntu.com/internet-of-things/features/recovery >> https://cn.ubuntu.com/internet-of-things/core/features/recovery
thanks


| priority | empty links on ubuntu core特点概览 i found links not working on the current links and correct links are ota 更新 安全启动 全盘加密 恢复模式 thanks | 1 |
589,827 | 17,761,599,021 | IssuesEvent | 2021-08-29 20:00:53 | nic547/TauStellwerk | https://api.github.com/repos/nic547/TauStellwerk | closed | Replace "Random User $number" default usernames | priority:medium type:maintenance size:S | Differentiating users based on a number is kinda difficult.
Doing something like adjective + animal name or a similar approach would probably be better. | 1.0 | Replace "Random User $number" default usernames - Differentiating users based on a number is kinda difficult.
Doing something like adjective + animal name or a similar approach would probably be better. | priority | replace random user number default usernames differentiating users based on a number is kinda difficult doing something like adjective animal name or a similar approach would probably be better | 1 |
571,702 | 17,023,349,137 | IssuesEvent | 2021-07-03 01:33:46 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Public Libraries should show in Mapnik | Component: mapnik Priority: minor Resolution: fixed Type: enhancement | **[Submitted to the original trac issue database at 10.53pm, Tuesday, 13th January 2009]**
Libraries (specifically a node with amenity=library) show on osmarender but not mapnik.
Examples:
Forbes Library is the public library of Northampton, MA, USA:
http://openstreetmap.org/?lat=42.31708&lon=-72.63611&zoom=17&layers=B000FTFT
There are many libraries here (Boston, MA, USA): http://openstreetmap.org/?lat=42.33889&lon=-71.09373&zoom=17&layers=0B00FTF
The Reed Free Library in Surry, NH, USA even has the ref parameter set: http://openstreetmap.org/?lat=43.01949&lon=-72.32138&zoom=17&layers=0B00FTFT | 1.0 | Public Libraries should show in Mapnik - **[Submitted to the original trac issue database at 10.53pm, Tuesday, 13th January 2009]**
Libraries (specifically a node with amenity=library) show on osmarender but not mapnik.
Examples:
Forbes Library is the public library of Northampton, MA, USA:
http://openstreetmap.org/?lat=42.31708&lon=-72.63611&zoom=17&layers=B000FTFT
There are many libraries here (Boston, MA, USA): http://openstreetmap.org/?lat=42.33889&lon=-71.09373&zoom=17&layers=0B00FTF
The Reed Free Library in Surry, NH, USA even has the ref parameter set: http://openstreetmap.org/?lat=43.01949&lon=-72.32138&zoom=17&layers=0B00FTFT | priority | public libraries should show in mapnik libraries specifically a node with amenity library show on osmarender but not mapnik examples forbes library is the public library of northampton ma usa there are many libraries here boston ma usa the reed free library in surry nh usa even has the ref parameter set | 1 |
26,505 | 13,039,424,841 | IssuesEvent | 2020-07-28 16:43:45 | aivivn/d2l-vn | https://api.github.com/repos/aivivn/d2l-vn | closed | Revise "hybridize_vn" - Phần 3 | chapter: computational-performance status: phase 2 | Phần này được dịch bởi: **davidnvq**
Nếu bạn đã dịch phần này, vui lòng bỏ qua việc revise. | True | Revise "hybridize_vn" - Phần 3 - Phần này được dịch bởi: **davidnvq**
Nếu bạn đã dịch phần này, vui lòng bỏ qua việc revise. | non_priority | revise hybridize vn phần phần này được dịch bởi davidnvq nếu bạn đã dịch phần này vui lòng bỏ qua việc revise | 0 |
163,377 | 25,803,200,607 | IssuesEvent | 2022-12-11 06:39:31 | jcklpe/research-ops-hackpack | https://api.github.com/repos/jcklpe/research-ops-hackpack | opened | Create new research project: {Project} | design product management research | ### Description
blah blah
### Acceptance Criteria
- [ ] Create a research [Epic](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign%2Cproduct+management%2Cepic&template=epic.yaml&title=%7BProject%2FTrial%7D+Epic).
- [ ] Create a research project folder.
- [ ] Create [project brief issue](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign&template=project-brief.yaml&title=Write+new+project+brief%3A+%7BProject%7D+).
- [ ] All tickets have been created and assigned
| 1.0 | Create new research project: {Project} - ### Description
blah blah
### Acceptance Criteria
- [ ] Create a research [Epic](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign%2Cproduct+management%2Cepic&template=epic.yaml&title=%7BProject%2FTrial%7D+Epic).
- [ ] Create a research project folder.
- [ ] Create [project brief issue](https://github.com/jcklpe/research-ops-hackpack/issues/new?assignees=octocat&labels=research%2Cdesign&template=project-brief.yaml&title=Write+new+project+brief%3A+%7BProject%7D+).
- [ ] All tickets have been created and assigned
| non_priority | create new research project project description blah blah acceptance criteria create a research create a research project folder create all tickets have been created and assigned | 0 |
124,829 | 16,668,906,745 | IssuesEvent | 2021-06-07 08:28:12 | blockframes/blockframes | https://api.github.com/repos/blockframes/blockframes | closed | Warning State on Avails Criteria Search Forms | Design - UI | Objective here is to add an intermediary state between hint and error, which would be warning.
First use case for that warning state is for the avails form: as soon as users fill in one of the avails fields, it should be clear that they have to fill in all the other ones as well before applying their search.
Research is needed to know what is possible with Angular forms | 1.0 | Warning State on Avails Criteria Search Forms - Objective here is to add an intermediary state between hint and error, which would be warning.
First use case for that warning state is for the avails form: as soon as users fill in one of the avails fields, it should be clear that they have to fill in all the other ones as well before applying their search.
Research is needed to know what is possible with Angular forms | non_priority | warning state on avails criteria search forms objective here is to add an intermediary state between hint and error which would be warning first use case for that warning state is for the avails form as soon as users fill in one of the avails fields it should be clear that they have to fill in all the other ones as well before applying their search research is needed to know what is possible with angular forms | 0 |
492,490 | 14,214,384,107 | IssuesEvent | 2020-11-17 05:06:04 | bounswe/bounswe2020group3 | https://api.github.com/repos/bounswe/bounswe2020group3 | closed | [Backend] Create Dockerfile | Backend Priority: Critical Type: Enhancement | * **Project: BACKEND**
* **This is a: FEATURE REQUEST**
* **Description of the issue**
There should be a dockerfile for CI/CD configurations
* **Deadline for resolution:**
ASAP
| 1.0 | [Backend] Create Dockerfile - * **Project: BACKEND**
* **This is a: FEATURE REQUEST**
* **Description of the issue**
There should be a dockerfile for CI/CD configurations
* **Deadline for resolution:**
ASAP
| priority | create dockerfile project backend this is a feature request description of the issue there should be a dockerfile for ci cd configurations deadline for resolution asap | 1 |
59,982 | 3,117,659,987 | IssuesEvent | 2015-09-04 03:51:09 | framingeinstein/issues-test | https://api.github.com/repos/framingeinstein/issues-test | opened | SPK-613: Promo Code Help | priority:normal priority:normal priority:normal priority:normal priority:normal priority:normal resolution:in-progress resolution:in-progress resolution:in-progress | Hi Toby,
See SRP-90:
We entered a new ticket with Magento to request help on a promo code.
1 SKU that should be eligible is the only SKU not allowing a promo code to be used.
Magento reviewed this issue on our front end and believes the custom theme is creating the promo code issue.
Here is their feedback:
Today I took a look at your site and I found that you have the special set up and running live in your production environment so I tested this out. I was able to reproduce what you have described and realized something:
The system accepted the coupon, but didn't allow me the customer to have a 0.00$ choice when I got to shipping.
This means the coupon process did work but your interface did not show the option on the front end. Now, the frontend has been customized and this is where the problem is. Your developers will need to look at the code for their onepage checkout page. As a test I recommend you or they disable your custom theme and test to see if it works using the native Magento theme.. I believe strongly it will work just fine. This is the strongest indicator the custom theme coding is the reason why the 0.00$ ground choice does not exist for the one sku you mentioned.
If you revert to the Magento theme and it doesn't correct the behavior then yes make the backup and add the links to the backup files to this ticket and I will download this and look further into this, but if it does work using the magento theme then your developer will need to solve this custom problem. | 6.0 | SPK-613: Promo Code Help - Hi Toby,
See SRP-90:
We entered a new ticket with Magento to request help on a promo code.
1 SKU that should be eligible is the only SKU not allowing a promo code to be used.
Magento reviewed this issue on our front end and believes the custom theme is creating the promo code issue.
Here is their feedback:
Today I took a look at your site and I found that you have the special set up and running live in your production environment so I tested this out. I was able to reproduce what you have described and realized something:
The system accepted the coupon, but didn't allow me the customer to have a 0.00$ choice when I got to shipping.
This means the coupon process did work but your interface did not show the option on the front end. Now, the frontend has been customized and this is where the problem is. Your developers will need to look at the code for their onepage checkout page. As a test I recommend you or they disable your custom theme and test to see if it works using the native Magento theme.. I believe strongly it will work just fine. This is the strongest indicator the custom theme coding is the reason why the 0.00$ ground choice does not exist for the one sku you mentioned.
If you revert to the Magento theme and it doesn't correct the behavior then yes make the backup and add the links to the backup files to this ticket and I will download this and look further into this, but if it does work using the magento theme then your developer will need to solve this custom problem. | priority | spk promo code help hi toby see srp we entered a new ticket with magento to request help on a promo code sku that should be eligible is the only sku not allowing a promo code to be used magento reviewed this issue on our front end and believes the custom theme is creating the promo code issue here is their feedback today i took a look at your site and i found that you have the special set up and running live in your production environment so i tested this out i was able to reproduce what you have described and realized something the system accepted the coupon but didn t allow me the customer to have a choice when i got to shipping this means the coupon process did work but your interface did not show the option on the front end now the frontend has been customized and this is where the problem is your developers will need to look at the code for their onepage checkout page as a test i recommend you or they disable your custom theme and test to see if it works using the native magento theme i believe strongly it will work just fine this is the strongest indicator the custom theme coding is the reason why the ground choice does not exist for the one sku you mentioned if you revert to the magento theme and it doesn t correct the behavior then yes make the backup and add the links to the backup files to this ticket and i will download this and look further into this but if it does work using the magento theme then your developer will need to solve this custom problem | 1 |
769,778 | 27,018,820,732 | IssuesEvent | 2023-02-10 22:25:12 | googleapis/nodejs-automl | https://api.github.com/repos/googleapis/nodejs-automl | closed | Automl Vision Object Detection Deploy Model Test: should deploy a model with a specified node count failed | type: bug priority: p1 api: automl flakybot: issue flakybot: flaky | Note: #688 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 97411a2bb514b9921bb3932543a2d895c452d5c6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/407b55a7-9544-44ff-be2c-1f1106f810da), [Sponge](http://sponge2/407b55a7-9544-44ff-be2c-1f1106f810da)
status: failed
<details><summary>Test output</summary><br><pre>expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/
AssertionError: expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/
at Context.<anonymous> (test/vision_object_detection_deploy_model_node_count.test.js:39:12)</pre></details> | 1.0 | Automl Vision Object Detection Deploy Model Test: should deploy a model with a specified node count failed - Note: #688 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 97411a2bb514b9921bb3932543a2d895c452d5c6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/407b55a7-9544-44ff-be2c-1f1106f810da), [Sponge](http://sponge2/407b55a7-9544-44ff-be2c-1f1106f810da)
status: failed
<details><summary>Test output</summary><br><pre>expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/
AssertionError: expected '16 UNAUTHENTICATED: Request had inval…' to match /NOT_FOUND/
at Context.<anonymous> (test/vision_object_detection_deploy_model_node_count.test.js:39:12)</pre></details> | priority | automl vision object detection deploy model test should deploy a model with a specified node count failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output expected unauthenticated request had inval… to match not found assertionerror expected unauthenticated request had inval… to match not found at context test vision object detection deploy model node count test js | 1 |
304,756 | 9,335,313,785 | IssuesEvent | 2019-03-28 18:15:52 | robotframework/SeleniumLibrary | https://api.github.com/repos/robotframework/SeleniumLibrary | closed | Set Browser Implicit Wait weird behavior | bug priority: medium | * Using the keyword Set Browser Implicit Wait will set the implicit wait but it can not be read with Get Selenium Implicit Wait, which returns only the value set with Set Selenium Implicit Wait or when importing the library.
* When you open a new tab (with js for exemple) and go back to the main tab the implicit wait is gone but it will read the last value that have been set with Set Selenium Implicit Wait.
## Steps to reproduce the issue
```
*** Test Cases ***
# Open the browser and stuff...
Log implicit wait before
Set Browser Implicit Wait 3s
Log implicit wait after set browser
Set Selenium Implicit Wait 2s
Log implicit wait after set 2s
Execute Javascript window.open("about:blank")
Select Window NEW
Select Window MAIN
# this will output 'after tab switch = 2 seconds' but the real value will be 0.0
Log implicit wait after tab switch
*** Keyword ***
Log implicit wait
[Arguments] ${prefix}
${implicit_wait} = Get Selenium Implicit Wait
Log ${prefix} = ${implicit_wait} console=True
```
## Environment
Browser: Google Chrome 73.0.3683.75
Browser driver: chromedriver=2.44.609538
Operating System: Windows 10 x64
Libraries
- Robot Framework: 3.1
- Selenium: 3.141.0
- SeleniumLibrary: 3.3.1
- Interpreter: Python 3.7.0b4 on win32 | 1.0 | Set Browser Implicit Wait weird behavior - * Using the keyword Set Browser Implicit Wait will set the implicit wait but it can not be read with Get Selenium Implicit Wait, which returns only the value set with Set Selenium Implicit Wait or when importing the library.
* When you open a new tab (with js for exemple) and go back to the main tab the implicit wait is gone but it will read the last value that have been set with Set Selenium Implicit Wait.
## Steps to reproduce the issue
```
*** Test Cases ***
# Open the browser and stuff...
Log implicit wait before
Set Browser Implicit Wait 3s
Log implicit wait after set browser
Set Selenium Implicit Wait 2s
Log implicit wait after set 2s
Execute Javascript window.open("about:blank")
Select Window NEW
Select Window MAIN
# this will output 'after tab switch = 2 seconds' but the real value will be 0.0
Log implicit wait after tab switch
*** Keyword ***
Log implicit wait
[Arguments] ${prefix}
${implicit_wait} = Get Selenium Implicit Wait
Log ${prefix} = ${implicit_wait} console=True
```
## Environment
Browser: Google Chrome 73.0.3683.75
Browser driver: chromedriver=2.44.609538
Operating System: Windows 10 x64
Libraries
- Robot Framework: 3.1
- Selenium: 3.141.0
- SeleniumLibrary: 3.3.1
- Interpreter: Python 3.7.0b4 on win32 | priority | set browser implicit wait weird behavior using the keyword set browser implicit wait will set the implicit wait but it can not be read with get selenium implicit wait which returns only the value set with set selenium implicit wait or when importing the library when you open a new tab with js for exemple and go back to the main tab the implicit wait is gone but it will read the last value that have been set with set selenium implicit wait steps to reproduce the issue test cases open the browser and stuff log implicit wait before set browser implicit wait log implicit wait after set browser set selenium implicit wait log implicit wait after set execute javascript window open about blank select window new select window main this will output after tab switch seconds but the real value will be log implicit wait after tab switch keyword log implicit wait prefix implicit wait get selenium implicit wait log prefix implicit wait console true environment browser google chrome browser driver chromedriver operating system windows libraries robot framework selenium seleniumlibrary interpreter python on | 1 |
282,822 | 30,889,436,543 | IssuesEvent | 2023-08-04 02:43:18 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | reopened | CVE-2023-28772 (Medium) detected in linux-stable-rtv4.1.33 | Mend: dependency security vulnerability | ## CVE-2023-28772 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.13.3. lib/seq_buf.c has a seq_buf_putmem_hex buffer overflow.
<p>Publish Date: 2023-03-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28772>CVE-2023-28772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-28772">https://www.linuxkernelcves.com/cves/CVE-2023-28772</a></p>
<p>Release Date: 2023-03-23</p>
<p>Fix Resolution: v4.4.276,v4.9.276,v4.14.240,v4.19.198,v5.4.133,v5.10.51,v5.12.18,v5.13.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-28772 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2023-28772 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/lib/seq_buf.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.13.3. lib/seq_buf.c has a seq_buf_putmem_hex buffer overflow.
<p>Publish Date: 2023-03-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28772>CVE-2023-28772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-28772">https://www.linuxkernelcves.com/cves/CVE-2023-28772</a></p>
<p>Release Date: 2023-03-23</p>
<p>Fix Resolution: v4.4.276,v4.9.276,v4.14.240,v4.19.198,v5.4.133,v5.10.51,v5.12.18,v5.13.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files lib seq buf c lib seq buf c vulnerability details an issue was discovered in the linux kernel before lib seq buf c has a seq buf putmem hex buffer overflow publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
277,830 | 8,633,238,693 | IssuesEvent | 2018-11-22 13:17:39 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Eco Forum Unsecure Login. | High Priority | The official eco forum (http://ecoforum.strangeloopgames.com) uses insecure http for logging into the forum, Anyone who would be logging into the forum would be sending their username and password combination in cleartext over the internet allowing anyone sniffing traffic to steal an admin password for the forum. You can get a FREE SSL cert through lets encrypt. | 1.0 | Eco Forum Unsecure Login. - The official eco forum (http://ecoforum.strangeloopgames.com) uses insecure http for logging into the forum, Anyone who would be logging into the forum would be sending their username and password combination in cleartext over the internet allowing anyone sniffing traffic to steal an admin password for the forum. You can get a FREE SSL cert through lets encrypt. | priority | eco forum unsecure login the official eco forum uses insecure http for logging into the forum anyone who would be logging into the forum would be sending their username and password combination in cleartext over the internet allowing anyone sniffing traffic to steal an admin password for the forum you can get a free ssl cert through lets encrypt | 1 |
239,072 | 26,201,215,106 | IssuesEvent | 2023-01-03 17:38:40 | MValle21/circonus-unified-agent | https://api.github.com/repos/MValle21/circonus-unified-agent | opened | CVE-2020-26160 (High) detected in github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible | security vulnerability | ## CVE-2020-26160 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible</b></p></summary>
<p>ARCHIVE - Golang implementation of JSON Web Tokens (JWT). This project is now maintained at:</p>
<p>Library home page: <a href="https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip">https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jwt-go before 4.0.0-preview1 allows attackers to bypass intended access restrictions in situations with []string{} for m["aud"] (which is allowed by the specification). Because the type assertion fails, "" is the value of aud. This is a security problem if the JWT token is presented to a service that lacks its own audience check.
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-26160>CVE-2020-26160</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-26160">https://nvd.nist.gov/vuln/detail/CVE-2020-26160</a></p>
<p>Release Date: 2020-09-30</p>
<p>Fix Resolution: v4.0.0-preview1</p>
</p>
</details>
<p></p>
| True | CVE-2020-26160 (High) detected in github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible - ## CVE-2020-26160 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible</b></p></summary>
<p>ARCHIVE - Golang implementation of JSON Web Tokens (JWT). This project is now maintained at:</p>
<p>Library home page: <a href="https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip">https://proxy.golang.org/github.com/dgrijalva/jwt-go/@v/v3.2.1-0.20200107013213-dc14462fd587+incompatible.zip</a></p>
<p>
Dependency Hierarchy:
- :x: **github.com/dgrijalva/jwt-go-v3.2.1-0.20200107013213-dc14462fd587+incompatible** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jwt-go before 4.0.0-preview1 allows attackers to bypass intended access restrictions in situations with []string{} for m["aud"] (which is allowed by the specification). Because the type assertion fails, "" is the value of aud. This is a security problem if the JWT token is presented to a service that lacks its own audience check.
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-26160>CVE-2020-26160</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-26160">https://nvd.nist.gov/vuln/detail/CVE-2020-26160</a></p>
<p>Release Date: 2020-09-30</p>
<p>Fix Resolution: v4.0.0-preview1</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in github com dgrijalva jwt go incompatible cve high severity vulnerability vulnerable library github com dgrijalva jwt go incompatible archive golang implementation of json web tokens jwt this project is now maintained at library home page a href dependency hierarchy x github com dgrijalva jwt go incompatible vulnerable library found in base branch master vulnerability details jwt go before allows attackers to bypass intended access restrictions in situations with string for m which is allowed by the specification because the type assertion fails is the value of aud this is a security problem if the jwt token is presented to a service that lacks its own audience check publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
175,540 | 21,313,848,145 | IssuesEvent | 2022-04-16 01:08:57 | Nivaskumark/kernel_v4.1.15 | https://api.github.com/repos/Nivaskumark/kernel_v4.1.15 | opened | CVE-2019-3459 (Medium) detected in linuxlinux-4.6 | security vulnerability | ## CVE-2019-3459 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A heap address information leak while using L2CAP_GET_CONF_OPT was discovered in the Linux kernel before 5.1-rc1.
<p>Publish Date: 2019-04-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3459>CVE-2019-3459</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459</a></p>
<p>Release Date: 2019-04-11</p>
<p>Fix Resolution: v5.1-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-3459 (Medium) detected in linuxlinux-4.6 - ## CVE-2019-3459 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A heap address information leak while using L2CAP_GET_CONF_OPT was discovered in the Linux kernel before 5.1-rc1.
<p>Publish Date: 2019-04-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-3459>CVE-2019-3459</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-3459</a></p>
<p>Release Date: 2019-04-11</p>
<p>Fix Resolution: v5.1-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files net bluetooth core c net bluetooth core c vulnerability details a heap address information leak while using get conf opt was discovered in the linux kernel before publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
350,772 | 31,932,305,166 | IssuesEvent | 2023-09-19 08:15:59 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix jax_lax_operators.test_jax_expand_dims | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
| 1.0 | Fix jax_lax_operators.test_jax_expand_dims - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6223268002/job/16888850206"><img src=https://img.shields.io/badge/-failure-red></a>
| non_priority | fix jax lax operators test jax expand dims numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src | 0 |
474,379 | 13,658,034,303 | IssuesEvent | 2020-09-28 06:56:08 | AY2021S1-CS2103T-T11-3/tp | https://api.github.com/repos/AY2021S1-CS2103T-T11-3/tp | closed | Add feature: add finance records | priority.High type.Story | As a forgetful business owner, I can save my transaction history, so that I can track my financials easily.
As a small business owner, I can view a summary of my finances, so that I can plan the next steps of my business.
| 1.0 | Add feature: add finance records - As a forgetful business owner, I can save my transaction history, so that I can track my financials easily.
As a small business owner, I can view a summary of my finances, so that I can plan the next steps of my business.
| priority | add feature add finance records as a forgetful business owner i can save my transaction history so that i can track my financials easily as a small business owner i can view a summary of my finances so that i can plan the next steps of my business | 1 |
101,684 | 31,394,752,004 | IssuesEvent | 2023-08-26 19:48:06 | expo/expo | https://api.github.com/repos/expo/expo | opened | Android 13 (Media Library Permissions) | needs validation Development Builds | ### Summary
When invoking [`await MediaLibrary.requestPermissionsAsync()`](https://docs.expo.dev/versions/latest/sdk/media-library/#medialibraryrequestpermissionsasyncwriteonly) on Android 13, two permission prompts appear asking for music/audio followed by photos/videos. According to the android developer docs in Android 13 or higher they have introduced more [granular permissions](https://developer.android.com/about/versions/13/behavior-changes-13#granular-media-permissions). I'm interested in only presenting the user with a single prompt (as I don't need to ask for music/audio) which appears would be [`READ_MEDIA_IMAGES`](https://developer.android.com/reference/android/Manifest.permission#READ_MEDIA_IMAGES), but adding that permission doesn't seem to do the trick. I also don't see that permission in the list of [available permissions](https://docs.expo.dev/versions/latest/config/app/#permissions). Lastly it looks like this was [addressed](https://github.com/expo/expo/pull/20907) in the past, but perhaps I'm missing something.
**Screenshots**
<img src="https://github.com/expo/expo/assets/7604441/290f8bd9-f380-4384-968f-67a66c889b22" width="200"/>
<img src="https://github.com/expo/expo/assets/7604441/8599599c-80ff-469e-9956-aeafe903ed6f" width="200"/>
<img src="https://github.com/expo/expo/assets/7604441/9ae1af49-4999-4eea-97ee-c2967711e786" width="200"/>
### Managed or bare workflow?
managed
### What platform(s) does this occur on?
Android
### Package versions
```json
"expo": "~49.0.8",
"expo-dev-client": "~2.4.8",
"expo-media-library": "~15.4.1",
"expo-status-bar": "~1.6.0",
"expo-updates": "~0.18.12",
"react": "18.2.0",
"react-native": "0.72.4"
```
### Environment
```
expo-env-info 1.0.5 environment info:
System:
OS: macOS 13.5
Shell: 5.9 - /bin/zsh
Binaries:
Node: 16.16.0 - /usr/local/bin/node
Yarn: 1.22.19 - ~/.yarn/bin/yarn
npm: 9.8.1 - /usr/local/bin/npm
Managers:
CocoaPods: 1.11.3 - /opt/homebrew/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 22.4, iOS 16.4, macOS 13.3, tvOS 16.4, watchOS 9.4
IDEs:
Android Studio: 2022.3 AI-223.8836.35.2231.10406996
Xcode: 14.3.1/14E300c - /usr/bin/xcodebuild
npmPackages:
expo: ~49.0.8 => 49.0.8
react: 18.2.0 => 18.2.0
react-native: 0.72.4 => 0.72.4
npmGlobalPackages:
eas-cli: 4.1.2
expo-cli: 6.3.10
Expo Workflow: managed
```
### Reproducible demo
https://github.com/iM-GeeKy/media-permissions
### Stacktrace (if a crash is involved)
_No response_ | 1.0 | Android 13 (Media Library Permissions) - ### Summary
When invoking [`await MediaLibrary.requestPermissionsAsync()`](https://docs.expo.dev/versions/latest/sdk/media-library/#medialibraryrequestpermissionsasyncwriteonly) on Android 13, two permission prompts appear asking for music/audio followed by photos/videos. According to the android developer docs in Android 13 or higher they have introduced more [granular permissions](https://developer.android.com/about/versions/13/behavior-changes-13#granular-media-permissions). I'm interested in only presenting the user with a single prompt (as I don't need to ask for music/audio) which appears would be [`READ_MEDIA_IMAGES`](https://developer.android.com/reference/android/Manifest.permission#READ_MEDIA_IMAGES), but adding that permission doesn't seem to do the trick. I also don't see that permission in the list of [available permissions](https://docs.expo.dev/versions/latest/config/app/#permissions). Lastly it looks like this was [addressed](https://github.com/expo/expo/pull/20907) in the past, but perhaps I'm missing something.
**Screenshots**
<img src="https://github.com/expo/expo/assets/7604441/290f8bd9-f380-4384-968f-67a66c889b22" width="200"/>
<img src="https://github.com/expo/expo/assets/7604441/8599599c-80ff-469e-9956-aeafe903ed6f" width="200"/>
<img src="https://github.com/expo/expo/assets/7604441/9ae1af49-4999-4eea-97ee-c2967711e786" width="200"/>
### Managed or bare workflow?
managed
### What platform(s) does this occur on?
Android
### Package versions
```json
"expo": "~49.0.8",
"expo-dev-client": "~2.4.8",
"expo-media-library": "~15.4.1",
"expo-status-bar": "~1.6.0",
"expo-updates": "~0.18.12",
"react": "18.2.0",
"react-native": "0.72.4"
```
### Environment
```
expo-env-info 1.0.5 environment info:
System:
OS: macOS 13.5
Shell: 5.9 - /bin/zsh
Binaries:
Node: 16.16.0 - /usr/local/bin/node
Yarn: 1.22.19 - ~/.yarn/bin/yarn
npm: 9.8.1 - /usr/local/bin/npm
Managers:
CocoaPods: 1.11.3 - /opt/homebrew/bin/pod
SDKs:
iOS SDK:
Platforms: DriverKit 22.4, iOS 16.4, macOS 13.3, tvOS 16.4, watchOS 9.4
IDEs:
Android Studio: 2022.3 AI-223.8836.35.2231.10406996
Xcode: 14.3.1/14E300c - /usr/bin/xcodebuild
npmPackages:
expo: ~49.0.8 => 49.0.8
react: 18.2.0 => 18.2.0
react-native: 0.72.4 => 0.72.4
npmGlobalPackages:
eas-cli: 4.1.2
expo-cli: 6.3.10
Expo Workflow: managed
```
### Reproducible demo
https://github.com/iM-GeeKy/media-permissions
### Stacktrace (if a crash is involved)
_No response_ | non_priority | android media library permissions summary when invoking on android two permission prompts appear asking for music audio followed by photos videos according to the android developer docs in android or higher they have introduced more i m interested in only presenting the user with a single prompt as i don t need to ask for music audio which appears would be but adding that permission doesn t seem to do the trick i also don t see that permission in the list of lastly it looks like this was in the past but perhaps i m missing something screenshots managed or bare workflow managed what platform s does this occur on android package versions json expo expo dev client expo media library expo status bar expo updates react react native environment expo env info environment info system os macos shell bin zsh binaries node usr local bin node yarn yarn bin yarn npm usr local bin npm managers cocoapods opt homebrew bin pod sdks ios sdk platforms driverkit ios macos tvos watchos ides android studio ai xcode usr bin xcodebuild npmpackages expo react react native npmglobalpackages eas cli expo cli expo workflow managed reproducible demo stacktrace if a crash is involved no response | 0 |
133,685 | 10,855,606,209 | IssuesEvent | 2019-11-13 18:46:14 | rancher/k3s | https://api.github.com/repos/rancher/k3s | closed | Upgrade to v0.10.0-rc2 to use default kubelet directory duplicates pod entry in old as well as in new directories. | [zube]: To Test | **Version:**
Upgrade from v0.9.1 to v0.10.0-rc2
**Describe the bug**
1. Upgrade to v0.10.0-rc2 has all pod entries from /var/lib/rancher/k3s/agent/kubelet/pods/ copied onto /var/lib/kubelet except helm and the pods are restarted.
2. Delete a pod removes in the new default dir and not from the old dir.
3. Helm pod is present only in old dir and not copied to default dir as job is in Completed state.
**To Reproduce**
1. Install v0.9.1
`curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.9.1 sh -s - server`
Create a deployment.
Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/
Upgrade to rc2
`curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.10.0-rc2 sh -s - server`
Verify the pods status
`kubectl get pods -A`
Verify the entries in /var/lib/kubelet
Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/
Scale up the pods
Verify the new entries in /var/lib/kubelet
As expected the new pods are not in /var/lib/rancher/k3s/agent/kubelet/pods/
Scale down the replicas to 0.
Pod entries are removed from new default dir and continues to appear in old dir.
5. Helm is not copied to new dir
`ls /var/lib/kubelet/pods/container/` |grep helm
**Expected behavior**
Pods created after upgrade should be in new default dir
Pods created before upgrade continue to remain in old dir
All pods are in Running state
Deleting a pod removes it from respective dir
**Actual behavior**
Pods created after upgrade should be in new default dir
Pods created before upgrade continue to remain in old dir as well as in new dir
All pods are in running state.
Deleting a pod removes it from new dir only
*** Additional info ***
```
ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods
total 28
drwxr-x--- 5 root root 4096 Oct 21 18:19 ccd0be75-173e-4df4-aa1e-57979ca63a93
drwxr-x--- 5 root root 4096 Oct 21 18:19 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6
drwxr-x--- 5 root root 4096 Oct 21 18:19 946d4695-d059-470c-a074-4089bece0944
drwxr-x--- 5 root root 4096 Oct 21 18:19 0456318c-2e28-47f4-95e2-ccdc7b712b8b
drwxr-x--- 5 root root 4096 Oct 21 18:19 7e056156-0b88-442d-a841-1df67def1ae6
drwxr-x--- 5 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf
drwxr-x--- 5 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb
ls -ltr /var/lib/kubelet/pods
total 24
drwxr-x--- 4 root root 4096 Oct 21 18:20 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6
drwxr-x--- 4 root root 4096 Oct 21 18:20 0456318c-2e28-47f4-95e2-ccdc7b712b8b
drwxr-x--- 4 root root 4096 Oct 21 18:20 7e056156-0b88-442d-a841-1df67def1ae6
drwxr-x--- 4 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb
drwxr-x--- 4 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf
drwxr-x--- 5 root root 4096 Oct 21 18:21 ccd0be75-173e-4df4-aa1e-57979ca63a93
ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods/946d4695-d059-470c-a074-4089bece0944/containers/helm/
total 0
ls -ltr /var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944
ls: cannot access '/var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944': No such file or directory
``` | 1.0 | Upgrade to v0.10.0-rc2 to use default kubelet directory duplicates pod entry in old as well as in new directories. - **Version:**
Upgrade from v0.9.1 to v0.10.0-rc2
**Describe the bug**
1. Upgrade to v0.10.0-rc2 has all pod entries from /var/lib/rancher/k3s/agent/kubelet/pods/ copied onto /var/lib/kubelet except helm and the pods are restarted.
2. Delete a pod removes in the new default dir and not from the old dir.
3. Helm pod is present only in old dir and not copied to default dir as job is in Completed state.
**To Reproduce**
1. Install v0.9.1
`curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.9.1 sh -s - server`
Create a deployment.
Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/
Upgrade to rc2
`curl -sfL https://get.k3s.io | INSTALL_K3S_VERSION=v0.10.0-rc2 sh -s - server`
Verify the pods status
`kubectl get pods -A`
Verify the entries in /var/lib/kubelet
Verify the entries in /var/lib/rancher/k3s/agent/kubelet/pods/
Scale up the pods
Verify the new entries in /var/lib/kubelet
As expected the new pods are not in /var/lib/rancher/k3s/agent/kubelet/pods/
Scale down the replicas to 0.
Pod entries are removed from new default dir and continues to appear in old dir.
5. Helm is not copied to new dir
`ls /var/lib/kubelet/pods/container/` |grep helm
**Expected behavior**
Pods created after upgrade should be in new default dir
Pods created before upgrade continue to remain in old dir
All pods are in Running state
Deleting a pod removes it from respective dir
**Actual behavior**
Pods created after upgrade should be in new default dir
Pods created before upgrade continue to remain in old dir as well as in new dir
All pods are in running state.
Deleting a pod removes it from new dir only
*** Additional info ***
```
ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods
total 28
drwxr-x--- 5 root root 4096 Oct 21 18:19 ccd0be75-173e-4df4-aa1e-57979ca63a93
drwxr-x--- 5 root root 4096 Oct 21 18:19 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6
drwxr-x--- 5 root root 4096 Oct 21 18:19 946d4695-d059-470c-a074-4089bece0944
drwxr-x--- 5 root root 4096 Oct 21 18:19 0456318c-2e28-47f4-95e2-ccdc7b712b8b
drwxr-x--- 5 root root 4096 Oct 21 18:19 7e056156-0b88-442d-a841-1df67def1ae6
drwxr-x--- 5 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf
drwxr-x--- 5 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb
ls -ltr /var/lib/kubelet/pods
total 24
drwxr-x--- 4 root root 4096 Oct 21 18:20 a20c6b3f-2950-4ab9-89bb-6cf08e22c1f6
drwxr-x--- 4 root root 4096 Oct 21 18:20 0456318c-2e28-47f4-95e2-ccdc7b712b8b
drwxr-x--- 4 root root 4096 Oct 21 18:20 7e056156-0b88-442d-a841-1df67def1ae6
drwxr-x--- 4 root root 4096 Oct 21 18:20 58c1bb4a-7d58-4f12-9fa6-5780a50492eb
drwxr-x--- 4 root root 4096 Oct 21 18:20 75ab116e-7977-45bb-b0f2-1bb3e5a64bcf
drwxr-x--- 5 root root 4096 Oct 21 18:21 ccd0be75-173e-4df4-aa1e-57979ca63a93
ls -ltr /var/lib/rancher/k3s/agent/kubelet/pods/946d4695-d059-470c-a074-4089bece0944/containers/helm/
total 0
ls -ltr /var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944
ls: cannot access '/var/lib/kubelet/pods/946d4695-d059-470c-a074-4089bece0944': No such file or directory
``` | non_priority | upgrade to to use default kubelet directory duplicates pod entry in old as well as in new directories version upgrade from to describe the bug upgrade to has all pod entries from var lib rancher agent kubelet pods copied onto var lib kubelet except helm and the pods are restarted delete a pod removes in the new default dir and not from the old dir helm pod is present only in old dir and not copied to default dir as job is in completed state to reproduce install curl sfl install version sh s server create a deployment verify the entries in var lib rancher agent kubelet pods upgrade to curl sfl install version sh s server verify the pods status kubectl get pods a verify the entries in var lib kubelet verify the entries in var lib rancher agent kubelet pods scale up the pods verify the new entries in var lib kubelet as expected the new pods are not in var lib rancher agent kubelet pods scale down the replicas to pod entries are removed from new default dir and continues to appear in old dir helm is not copied to new dir ls var lib kubelet pods container grep helm expected behavior pods created after upgrade should be in new default dir pods created before upgrade continue to remain in old dir all pods are in running state deleting a pod removes it from respective dir actual behavior pods created after upgrade should be in new default dir pods created before upgrade continue to remain in old dir as well as in new dir all pods are in running state deleting a pod removes it from new dir only additional info ls ltr var lib rancher agent kubelet pods total drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct ls ltr var lib kubelet pods total drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct drwxr x root root oct ls ltr var lib rancher agent kubelet pods containers helm total ls ltr var lib kubelet pods ls cannot access var lib kubelet pods no such file or directory | 0 |
50,997 | 13,188,026,176 | IssuesEvent | 2020-08-13 05:20:36 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | [g4-tankresponse] memory leaks (Trac #1796) | Migrated from Trac combo simulation defect | potential memory leaks found by static analysis:
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1796">https://code.icecube.wisc.edu/ticket/1796</a>, reported by kjmeagher and owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-08-10T22:54:14",
"description": "potential memory leaks found by static analysis: \nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "invalid",
"_ts": "1470869654585407",
"component": "combo simulation",
"summary": "[g4-tankresponse] memory leaks",
"priority": "normal",
"keywords": "",
"time": "2016-07-27T07:57:17",
"milestone": "Long-Term Future",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [g4-tankresponse] memory leaks (Trac #1796) - potential memory leaks found by static analysis:
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath
http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1796">https://code.icecube.wisc.edu/ticket/1796</a>, reported by kjmeagher and owned by jgonzalez</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-08-10T22:54:14",
"description": "potential memory leaks found by static analysis: \nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-5e3769.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-314610.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-3268e5.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-45c21e.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-6d8586.html#EndPath\nhttp://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-f55f2c.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "invalid",
"_ts": "1470869654585407",
"component": "combo simulation",
"summary": "[g4-tankresponse] memory leaks",
"priority": "normal",
"keywords": "",
"time": "2016-07-27T07:57:17",
"milestone": "Long-Term Future",
"owner": "jgonzalez",
"type": "defect"
}
```
</p>
</details>
| non_priority | memory leaks trac potential memory leaks found by static analysis migrated from json status closed changetime description potential memory leaks found by static analysis n reporter kjmeagher cc resolution invalid ts component combo simulation summary memory leaks priority normal keywords time milestone long term future owner jgonzalez type defect | 0 |
43,509 | 2,889,826,327 | IssuesEvent | 2015-06-13 20:00:42 | damonkohler/android-scripting | https://api.github.com/repos/damonkohler/android-scripting | closed | Display additional info before installing interpretter | auto-migrated LowHangingFruit Priority-Medium Type-Enhancement | ```
Interpretter APKs should display version and total download size.
```
Original issue reported on code.google.com by `MeanEYE.rcf` on 7 Jul 2010 at 10:51 | 1.0 | Display additional info before installing interpretter - ```
Interpretter APKs should display version and total download size.
```
Original issue reported on code.google.com by `MeanEYE.rcf` on 7 Jul 2010 at 10:51 | priority | display additional info before installing interpretter interpretter apks should display version and total download size original issue reported on code google com by meaneye rcf on jul at | 1 |
187,839 | 6,761,563,228 | IssuesEvent | 2017-10-25 02:35:19 | opencollective/opencollective | https://api.github.com/repos/opencollective/opencollective | closed | Donor can't make a donation: logged in, system prompts her to login again | high priority | Thank you for taking the time to report an issue 🙏
The easier it is for us to reproduce it, the faster we can solve it.
So please try to be as complete as possible when filing your issue.
***
From user:
> The donation form beginning is a bit confusing for her: She’s already logged in, but placing her email in the form prompts her to log in again. Is this normal?
URL: https://opencollective.com/buttercup/donate
Logged in as: https://opencollective.com/ida-mitchell
Error message if any:
Expected result: being able to make a donation
Browser:
Bonus point if you can add a screenshot :-)
<img width="1328" alt="pastedgraphic-3" src="https://user-images.githubusercontent.com/3671070/31959680-5bdf7e36-b8c3-11e7-94c3-5194243b1169.png">
Thank you and have a wonderful day/evening!
(and sorry for inconvenience! We'll do our best to solve this ASAP)
| 1.0 | Donor can't make a donation: logged in, system prompts her to login again - Thank you for taking the time to report an issue 🙏
The easier it is for us to reproduce it, the faster we can solve it.
So please try to be as complete as possible when filing your issue.
***
From user:
> The donation form beginning is a bit confusing for her: She’s already logged in, but placing her email in the form prompts her to log in again. Is this normal?
URL: https://opencollective.com/buttercup/donate
Logged in as: https://opencollective.com/ida-mitchell
Error message if any:
Expected result: being able to make a donation
Browser:
Bonus point if you can add a screenshot :-)
<img width="1328" alt="pastedgraphic-3" src="https://user-images.githubusercontent.com/3671070/31959680-5bdf7e36-b8c3-11e7-94c3-5194243b1169.png">
Thank you and have a wonderful day/evening!
(and sorry for inconvenience! We'll do our best to solve this ASAP)
| priority | donor can t make a donation logged in system prompts her to login again thank you for taking the time to report an issue 🙏 the easier it is for us to reproduce it the faster we can solve it so please try to be as complete as possible when filing your issue from user the donation form beginning is a bit confusing for her she’s already logged in but placing her email in the form prompts her to log in again is this normal url logged in as error message if any expected result being able to make a donation browser bonus point if you can add a screenshot img width alt pastedgraphic src thank you and have a wonderful day evening and sorry for inconvenience we ll do our best to solve this asap | 1 |
3,124 | 8,972,202,604 | IssuesEvent | 2019-01-29 17:39:59 | firecracker-microvm/firecracker | https://api.github.com/repos/firecracker-microvm/firecracker | closed | Add Try_Read Functionality to EventFd Implementation | Contribute: Good First Issue Priority: Low Program: Architecture Quality: Improvement | We are currently using EventFds via the implementation from the sys_util crate, inherited from crosvm. One potentially significant downside is that it only exposes a blocking `read()` API.
There are places in our code where we use `read()` knowing that an event should definitely be available, because the logic is programmed that way, but if bugs are introduces, or something goes terribly wrong otherwise, the code will simply hang on the `read()` (not even panic).
Maybe it would make sense to add some sort of `try_read()` functionality, and do checked reads in this context. There will be some additional overhead involved, but it should not be significant. | 1.0 | Add Try_Read Functionality to EventFd Implementation - We are currently using EventFds via the implementation from the sys_util crate, inherited from crosvm. One potentially significant downside is that it only exposes a blocking `read()` API.
There are places in our code where we use `read()` knowing that an event should definitely be available, because the logic is programmed that way, but if bugs are introduces, or something goes terribly wrong otherwise, the code will simply hang on the `read()` (not even panic).
Maybe it would make sense to add some sort of `try_read()` functionality, and do checked reads in this context. There will be some additional overhead involved, but it should not be significant. | non_priority | add try read functionality to eventfd implementation we are currently using eventfds via the implementation from the sys util crate inherited from crosvm one potentially significant downside is that it only exposes a blocking read api there are places in our code where we use read knowing that an event should definitely be available because the logic is programmed that way but if bugs are introduces or something goes terribly wrong otherwise the code will simply hang on the read not even panic maybe it would make sense to add some sort of try read functionality and do checked reads in this context there will be some additional overhead involved but it should not be significant | 0 |
19,773 | 5,932,300,454 | IssuesEvent | 2017-05-24 08:59:48 | HGustavs/LenaSYS | https://api.github.com/repos/HGustavs/LenaSYS | closed | missleading text in code viewer dialog | All CodeViewer Needs Fixing | **In the picture below, should it really say "delete" and not "cancel"?**

| 1.0 | missleading text in code viewer dialog - **In the picture below, should it really say "delete" and not "cancel"?**

| non_priority | missleading text in code viewer dialog in the picture below should it really say delete and not cancel | 0 |
57,893 | 16,131,877,731 | IssuesEvent | 2021-04-29 06:39:32 | cython/cython | https://api.github.com/repos/cython/cython | closed | Default arguments in methods are not preserved for introspection | Python Semantics defect | Tried on the latest master:
```
# test.pyx
def run(a, b=1):
return a + b
cdef class A:
def run(self, a, b=1):
return a + b
# app.py
from test import A, run
import inspect
a = A()
print(inspect.signature(run)) # ok - prints (a, b=1)
print(inspect.signature(a.run)) # not ok - prints (a, b)
print(inspect.signature(A.run)) # not ok - prints (self, a, b)
```
| 1.0 | Default arguments in methods are not preserved for introspection - Tried on the latest master:
```
# test.pyx
def run(a, b=1):
return a + b
cdef class A:
def run(self, a, b=1):
return a + b
# app.py
from test import A, run
import inspect
a = A()
print(inspect.signature(run)) # ok - prints (a, b=1)
print(inspect.signature(a.run)) # not ok - prints (a, b)
print(inspect.signature(A.run)) # not ok - prints (self, a, b)
```
| non_priority | default arguments in methods are not preserved for introspection tried on the latest master test pyx def run a b return a b cdef class a def run self a b return a b app py from test import a run import inspect a a print inspect signature run ok prints a b print inspect signature a run not ok prints a b print inspect signature a run not ok prints self a b | 0 |
71,401 | 3,356,465,895 | IssuesEvent | 2015-11-18 20:37:16 | psouza4/mediacentermaster | https://api.github.com/repos/psouza4/mediacentermaster | closed | Create a file blacklist for parser | Affects-General-Usability Component-Functionality Feature-Downloads Priority-Medium Type-FeatureRequest | Create a user editable configuration file (JSON, XML,etc) to be processed for a blacklist of files to be deleted and ignored while parsing movies and tv shows. The blacklist will be a string value, and depending on code implementation, may support .Net regular expressions. | 1.0 | Create a file blacklist for parser - Create a user editable configuration file (JSON, XML,etc) to be processed for a blacklist of files to be deleted and ignored while parsing movies and tv shows. The blacklist will be a string value, and depending on code implementation, may support .Net regular expressions. | priority | create a file blacklist for parser create a user editable configuration file json xml etc to be processed for a blacklist of files to be deleted and ignored while parsing movies and tv shows the blacklist will be a string value and depending on code implementation may support net regular expressions | 1 |
121,229 | 17,648,388,093 | IssuesEvent | 2021-08-20 09:37:23 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [Security Solution]Rule name is clickable and navigated to rule page contradicting TGrid Form AC | bug Team: SecuritySolution | **Describe the bug**
Rule name is clickable and navigated to rule page contradicting TGrid Form AC
**Build Details**
```
Version:7.15.0
Commit:d791226d9385122f33f4a5ca38fa5369012fbec3
Build:43636
```
**Browsers**
all
**Precondition**
- Install the Endpoint Security on the kibana
**Steps to Reproduce**
1. Generate few Alerts.
2. Observed that Rule name value is clickable and let user to its rule page which is contradicting the updated TGrid form of the alert list.

**Actual Result**
Rule name is clickable and navigated to rule page contradicting TGrid Form AC
**Expected Result**
Rule name should not be clickable
**Screen-Shoot**

**logs**
N/A | True | [Security Solution]Rule name is clickable and navigated to rule page contradicting TGrid Form AC - **Describe the bug**
Rule name is clickable and navigated to rule page contradicting TGrid Form AC
**Build Details**
```
Version:7.15.0
Commit:d791226d9385122f33f4a5ca38fa5369012fbec3
Build:43636
```
**Browsers**
all
**Precondition**
- Install the Endpoint Security on the kibana
**Steps to Reproduce**
1. Generate few Alerts.
2. Observed that Rule name value is clickable and let user to its rule page which is contradicting the updated TGrid form of the alert list.

**Actual Result**
Rule name is clickable and navigated to rule page contradicting TGrid Form AC
**Expected Result**
Rule name should not be clickable
**Screen-Shoot**

**logs**
N/A | non_priority | rule name is clickable and navigated to rule page contradicting tgrid form ac describe the bug rule name is clickable and navigated to rule page contradicting tgrid form ac build details version commit build browsers all precondition install the endpoint security on the kibana steps to reproduce generate few alerts observed that rule name value is clickable and let user to its rule page which is contradicting the updated tgrid form of the alert list actual result rule name is clickable and navigated to rule page contradicting tgrid form ac expected result rule name should not be clickable screen shoot logs n a | 0 |
64,443 | 26,735,622,277 | IssuesEvent | 2023-01-30 09:15:07 | flipperdevices/flipperzero-firmware | https://api.github.com/repos/flipperdevices/flipperzero-firmware | closed | Show SD card CID in SD card info | feature request core+services | ### Describe the enhancement you're suggesting.
The card info shows filesystem type and size but not info on the physical hardware.
When there are troubles with some cards it would be nice to be able to see which card type is used.
### Anything else?
_No response_ | 1.0 | Show SD card CID in SD card info - ### Describe the enhancement you're suggesting.
The card info shows filesystem type and size but not info on the physical hardware.
When there are troubles with some cards it would be nice to be able to see which card type is used.
### Anything else?
_No response_ | non_priority | show sd card cid in sd card info describe the enhancement you re suggesting the card info shows filesystem type and size but not info on the physical hardware when there are troubles with some cards it would be nice to be able to see which card type is used anything else no response | 0 |
119,065 | 15,395,579,778 | IssuesEvent | 2021-03-03 19:25:27 | emory-libraries/blacklight-catalog | https://api.github.com/repos/emory-libraries/blacklight-catalog | opened | Review current bootstrap CSS and recommend changes needed for current theming and branding | UI Design View (Display and Navigation) | As the product owner, I would like to begin exploring what is needed to begin theming and branding the Blacklight application. The current bootstrap used for Emory Digital Collections is based off the pattern library for Emory Libraries website. Based on initial design of wireframes for header, footer, homepage, facets, and single item display, I would like to understand from a UI design perspective if any modifications are needed to the Bootstrap CSS. This will allow us to begin development of needed components of the header and footer in the coming sprints.
Acceptance criteria:
- [ ] Review the [Bootstrap Mapping for Emory Digital Collections](https://docs.google.com/spreadsheets/d/1-sbatCXlB7ysNHPrj3vT2mBy4nufx1ovlG7UwX2sdo8/edit?usp=sharing)
- [ ] Determine if any new components are needed at this point for the Blacklight Catalog based on current designs for header, footer, facets, home page and single item view page.
- [ ] Document and review recommendations with product owner. | 1.0 | Review current bootstrap CSS and recommend changes needed for current theming and branding - As the product owner, I would like to begin exploring what is needed to begin theming and branding the Blacklight application. The current bootstrap used for Emory Digital Collections is based off the pattern library for Emory Libraries website. Based on initial design of wireframes for header, footer, homepage, facets, and single item display, I would like to understand from a UI design perspective if any modifications are needed to the Bootstrap CSS. This will allow us to begin development of needed components of the header and footer in the coming sprints.
Acceptance criteria:
- [ ] Review the [Bootstrap Mapping for Emory Digital Collections](https://docs.google.com/spreadsheets/d/1-sbatCXlB7ysNHPrj3vT2mBy4nufx1ovlG7UwX2sdo8/edit?usp=sharing)
- [ ] Determine if any new components are needed at this point for the Blacklight Catalog based on current designs for header, footer, facets, home page and single item view page.
- [ ] Document and review recommendations with product owner. | non_priority | review current bootstrap css and recommend changes needed for current theming and branding as the product owner i would like to begin exploring what is needed to begin theming and branding the blacklight application the current bootstrap used for emory digital collections is based off the pattern library for emory libraries website based on initial design of wireframes for header footer homepage facets and single item display i would like to understand from a ui design perspective if any modifications are needed to the bootstrap css this will allow us to begin development of needed components of the header and footer in the coming sprints acceptance criteria review the determine if any new components are needed at this point for the blacklight catalog based on current designs for header footer facets home page and single item view page document and review recommendations with product owner | 0 |
228,517 | 7,552,437,773 | IssuesEvent | 2018-04-19 00:17:30 | leo-project/leofs | https://api.github.com/repos/leo-project/leofs | closed | [eleveldb] Make log files less fragmented | Improve Priority-MIDDLE survey v1.4 | As reported on https://github.com/leo-project/leofs/issues/940 by @vstax, eleveldb have a fragmentation problem for its .log files. This can be fixed by calling fallocate() or posix_fallocate() when creating .log file. and also we have to check how this fix would affect the performance of synced writes to .log file for safe. | 1.0 | [eleveldb] Make log files less fragmented - As reported on https://github.com/leo-project/leofs/issues/940 by @vstax, eleveldb have a fragmentation problem for its .log files. This can be fixed by calling fallocate() or posix_fallocate() when creating .log file. and also we have to check how this fix would affect the performance of synced writes to .log file for safe. | priority | make log files less fragmented as reported on by vstax eleveldb have a fragmentation problem for its log files this can be fixed by calling fallocate or posix fallocate when creating log file and also we have to check how this fix would affect the performance of synced writes to log file for safe | 1 |
301,839 | 22,777,115,619 | IssuesEvent | 2022-07-08 15:27:48 | orchest/orchest | https://api.github.com/repos/orchest/orchest | closed | Improve documentation of installation targets | improvement documentation | **Describe the problem this improvement solves**
At the moment, we are documenting two different installation methods:
- minikube
- `kubectl`
and, in addition, our convenience script.
However, it would be good to add more clear separation between all these methods, and add other interesting targets such as EKS (see #930, #885).
**Describe the solution you'd like**
A very clear and easy to navigate installation page in which users can pick the method that best suits them. | 1.0 | Improve documentation of installation targets - **Describe the problem this improvement solves**
At the moment, we are documenting two different installation methods:
- minikube
- `kubectl`
and, in addition, our convenience script.
However, it would be good to add more clear separation between all these methods, and add other interesting targets such as EKS (see #930, #885).
**Describe the solution you'd like**
A very clear and easy to navigate installation page in which users can pick the method that best suits them. | non_priority | improve documentation of installation targets describe the problem this improvement solves at the moment we are documenting two different installation methods minikube kubectl and in addition our convenience script however it would be good to add more clear separation between all these methods and add other interesting targets such as eks see describe the solution you d like a very clear and easy to navigate installation page in which users can pick the method that best suits them | 0 |
240,645 | 18,363,637,398 | IssuesEvent | 2021-10-09 17:14:49 | KylieScharf/flask_portfolio | https://api.github.com/repos/KylieScharf/flask_portfolio | opened | edit journal | documentation | Final Journals
- [x] kylie and khushi
[journal](https://docs.google.com/document/d/1eyTgQhMv7jFi28SIGlwOs3k95a9KklfJFCS0O9pXCzA/edit?usp=sharing)
- [x] Kevin and Hamza and Daniel
[journal](https://docs.google.com/document/d/1fy-J_PVrvafykD-OTxMm-Pe0L_UKrYUTC4-DP4vGXEw/edit?usp=sharing)
Add 3.5 and 3.6 notes to journal
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Add test corrections to journal
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Add TPT notes to journal
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Add TT notes to journal (if any)
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Finish all journal entries
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel | 1.0 | edit journal - Final Journals
- [x] kylie and khushi
[journal](https://docs.google.com/document/d/1eyTgQhMv7jFi28SIGlwOs3k95a9KklfJFCS0O9pXCzA/edit?usp=sharing)
- [x] Kevin and Hamza and Daniel
[journal](https://docs.google.com/document/d/1fy-J_PVrvafykD-OTxMm-Pe0L_UKrYUTC4-DP4vGXEw/edit?usp=sharing)
Add 3.5 and 3.6 notes to journal
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Add test corrections to journal
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Add TPT notes to journal
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Add TT notes to journal (if any)
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel
Finish all journal entries
- [x] kylie
- [x] khushi
- [x] kevin
- [x] hamza
- [x] daniel | non_priority | edit journal final journals kylie and khushi kevin and hamza and daniel add and notes to journal kylie khushi kevin hamza daniel add test corrections to journal kylie khushi kevin hamza daniel add tpt notes to journal kylie khushi kevin hamza daniel add tt notes to journal if any kylie khushi kevin hamza daniel finish all journal entries kylie khushi kevin hamza daniel | 0 |
151,617 | 23,848,813,285 | IssuesEvent | 2022-09-06 16:00:04 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | No comments pane | *as-designed | Type: <b>Bug</b>
I cannot find the [comments pane](https://code.visualstudio.com/updates/v1_71#_comments) anywhere. It doesn't matter which setting I use for it to appear.

VS Code version: Code 1.71.0 (784b0177c56c607789f9638da7b6bf3230d47a8c, 2022-09-01T07:25:10.472Z)
OS version: Linux x64 5.15.60-1-MANJARO
Modes:
Sandboxed: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i7-4790 CPU @ 3.60GHz (8 x 3870)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: disabled_software<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off|
|Load (avg)|0, 0, 1|
|Memory (System)|15.52GB (11.27GB free)|
|Process Argv|/a/bitbean/smg-auto/dealsmgr/api-server|
|Screen Reader|no|
|VM|0%|
|DESKTOP_SESSION|xfce|
|XDG_CURRENT_DESKTOP|XFCE|
|XDG_SESSION_DESKTOP|xfce|
|XDG_SESSION_TYPE|x11|
</details><details><summary>Extensions (32)</summary>
Extension|Author (truncated)|Version
---|---|---
gitdiffer|aag|0.0.3
Bookmarks|ale|13.3.1
better-toml|bun|0.3.2
dart-code|Dar|3.48.1
flutter|Dar|3.48.0
vscode-eslint|dba|2.2.6
vscode-deno|den|3.13.1
githistory|don|0.6.19
permute-lines|ear|1.1.0
EditorConfig|Edi|0.16.4
prettier-vscode|esb|9.8.0
vscode-solution-explorer|fer|0.7.1
go|gol|0.35.2
css-to-jss|inf|1.0.7
plantuml|jeb|2.17.4
code-eol|jef|1.0.12
docthis|joe|0.7.1
docomment|k--|0.1.31
csharpfixformat|Leo|0.0.84
csharp|ms-|1.25.0
python|ms-|2022.14.0
vscode-pylance|ms-|2022.8.50
jupyter|ms-|2022.8.1002431955
jupyter-keymap|ms-|1.0.0
jupyter-renderers|ms-|1.0.9
remote-containers|ms-|0.251.0
remote-ssh|ms-|0.84.0
remote-ssh-edit|ms-|0.80.0
remote-wsl|ms-|0.66.3
vscode-remote-extensionpack|ms-|0.21.0
gitlink|qez|1.2.4
vscode-yaml|red|1.10.1
</details>
<!-- generated by issue reporter --> | 1.0 | No comments pane - Type: <b>Bug</b>
I cannot find the [comments pane](https://code.visualstudio.com/updates/v1_71#_comments) anywhere. It doesn't matter which setting I use for it to appear.

VS Code version: Code 1.71.0 (784b0177c56c607789f9638da7b6bf3230d47a8c, 2022-09-01T07:25:10.472Z)
OS version: Linux x64 5.15.60-1-MANJARO
Modes:
Sandboxed: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i7-4790 CPU @ 3.60GHz (8 x 3870)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: disabled_software<br>video_encode: disabled_software<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off|
|Load (avg)|0, 0, 1|
|Memory (System)|15.52GB (11.27GB free)|
|Process Argv|/a/bitbean/smg-auto/dealsmgr/api-server|
|Screen Reader|no|
|VM|0%|
|DESKTOP_SESSION|xfce|
|XDG_CURRENT_DESKTOP|XFCE|
|XDG_SESSION_DESKTOP|xfce|
|XDG_SESSION_TYPE|x11|
</details><details><summary>Extensions (32)</summary>
Extension|Author (truncated)|Version
---|---|---
gitdiffer|aag|0.0.3
Bookmarks|ale|13.3.1
better-toml|bun|0.3.2
dart-code|Dar|3.48.1
flutter|Dar|3.48.0
vscode-eslint|dba|2.2.6
vscode-deno|den|3.13.1
githistory|don|0.6.19
permute-lines|ear|1.1.0
EditorConfig|Edi|0.16.4
prettier-vscode|esb|9.8.0
vscode-solution-explorer|fer|0.7.1
go|gol|0.35.2
css-to-jss|inf|1.0.7
plantuml|jeb|2.17.4
code-eol|jef|1.0.12
docthis|joe|0.7.1
docomment|k--|0.1.31
csharpfixformat|Leo|0.0.84
csharp|ms-|1.25.0
python|ms-|2022.14.0
vscode-pylance|ms-|2022.8.50
jupyter|ms-|2022.8.1002431955
jupyter-keymap|ms-|1.0.0
jupyter-renderers|ms-|1.0.9
remote-containers|ms-|0.251.0
remote-ssh|ms-|0.84.0
remote-ssh-edit|ms-|0.80.0
remote-wsl|ms-|0.66.3
vscode-remote-extensionpack|ms-|0.21.0
gitlink|qez|1.2.4
vscode-yaml|red|1.10.1
</details>
<!-- generated by issue reporter --> | non_priority | no comments pane type bug i cannot find the anywhere it doesn t matter which setting i use for it to appear vs code version code os version linux manjaro modes sandboxed no system info item value cpus intel r core tm cpu x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled multiple raster threads enabled on opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode disabled software video encode disabled software vulkan disabled off webgl enabled enabled webgpu disabled off load avg memory system free process argv a bitbean smg auto dealsmgr api server screen reader no vm desktop session xfce xdg current desktop xfce xdg session desktop xfce xdg session type extensions extension author truncated version gitdiffer aag bookmarks ale better toml bun dart code dar flutter dar vscode eslint dba vscode deno den githistory don permute lines ear editorconfig edi prettier vscode esb vscode solution explorer fer go gol css to jss inf plantuml jeb code eol jef docthis joe docomment k csharpfixformat leo csharp ms python ms vscode pylance ms jupyter ms jupyter keymap ms jupyter renderers ms remote containers ms remote ssh ms remote ssh edit ms remote wsl ms vscode remote extensionpack ms gitlink qez vscode yaml red | 0 |
204,081 | 15,398,716,972 | IssuesEvent | 2021-03-04 00:37:16 | nucleus-security/Test-repo | https://api.github.com/repos/nucleus-security/Test-repo | opened | Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [Medium] - CentOS Security Update for kernel(CESA-2017:2863) | Testing | Source: QUALYS
Finding Description: CentOS has released security update for kernel to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6
Impact: This vulnerability could be exploited to gain complete access to sensitive information. Malicious users could also use this vulnerability to change all the contents or configuration on the system. Additionally this vulnerability can also be used to cause a complete denial of service and could render the resource completely unavailable.</p>
Target(s): Asset name: 45.55.254.143
IP: 45.55.254.143
Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">centos 6</a> for updates and patch information.
<p>Patch:<br />
Following are links for downloading patches to fix the vulnerabilities:
</p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">CESA-2017:2863: centos 6</a></p>
References:
ID:256316
CVE:CVE-2017-7541
Category:CentOS
PCI Flagged:1
Vendor References:CESA-2017:2863 centos 6
Bugtraq IDs:99955
Severity: Medium
Date Discovered: 2020-01-07 14:35:48
Nucleus Notification Rules Triggered: GitHub Rule
Project Name: Ticketing Rules now apply to all vulnerabilities
| 1.0 | Nucleus - Project: Ticketing Rules now apply to all vulnerabilities - [Medium] - CentOS Security Update for kernel(CESA-2017:2863) - Source: QUALYS
Finding Description: CentOS has released security update for kernel to fix the vulnerabilities.<p>Affected Products:<br /><br />centos 6
Impact: This vulnerability could be exploited to gain complete access to sensitive information. Malicious users could also use this vulnerability to change all the contents or configuration on the system. Additionally this vulnerability can also be used to cause a complete denial of service and could render the resource completely unavailable.</p>
Target(s): Asset name: 45.55.254.143
IP: 45.55.254.143
Solution: To resolve this issue, upgrade to the latest packages which contain a patch. Refer to CentOS advisory <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">centos 6</a> for updates and patch information.
<p>Patch:<br />
Following are links for downloading patches to fix the vulnerabilities:
</p><p> <a href="https://lists.centos.org/pipermail/centos-announce/2017-october/022564.html">CESA-2017:2863: centos 6</a></p>
References:
ID:256316
CVE:CVE-2017-7541
Category:CentOS
PCI Flagged:1
Vendor References:CESA-2017:2863 centos 6
Bugtraq IDs:99955
Severity: Medium
Date Discovered: 2020-01-07 14:35:48
Nucleus Notification Rules Triggered: GitHub Rule
Project Name: Ticketing Rules now apply to all vulnerabilities
| non_priority | nucleus project ticketing rules now apply to all vulnerabilities centos security update for kernel cesa source qualys finding description centos has released security update for kernel to fix the vulnerabilities affected products centos impact this vulnerability could be exploited to gain complete access to sensitive information malicious users could also use this vulnerability to change all the contents or configuration on the system additionally this vulnerability can also be used to cause a complete denial of service and could render the resource completely unavailable target s asset name ip solution to resolve this issue upgrade to the latest packages which contain a patch refer to centos advisory for updates and patch information patch following are links for downloading patches to fix the vulnerabilities references id cve cve category centos pci flagged vendor references cesa centos bugtraq ids severity medium date discovered nucleus notification rules triggered github rule project name ticketing rules now apply to all vulnerabilities | 0 |
148,354 | 23,343,120,061 | IssuesEvent | 2022-08-09 15:30:29 | Atri-Labs/atrilabs-engine | https://api.github.com/repos/Atri-Labs/atrilabs-engine | closed | Change UI design of actions bar | design | 1. Send data
2. Send file - send own or different component
3. Internal navigation
4. External navigation | 1.0 | Change UI design of actions bar - 1. Send data
2. Send file - send own or different component
3. Internal navigation
4. External navigation | non_priority | change ui design of actions bar send data send file send own or different component internal navigation external navigation | 0 |
96,668 | 20,053,529,803 | IssuesEvent | 2022-02-03 09:36:23 | creativecommons/project_creativecommons.org | https://api.github.com/repos/creativecommons/project_creativecommons.org | opened | Determine whether we still (intend to) use Engaging Networks | 🟩 priority: low ✨ goal: improvement 💻 aspect: code 🚦 status: awaiting triage | There is a Gravityforms Engaging Networks addon defined as a dependency of our staging website. However, the legacy website does not contain the same plugin.
## Task
- [ ] determine whether we still use or intend to use the Engaging Networks service
- [ ] decide on whether or not to include the Gravityforms Engaging Networks addon based on our intentions | 1.0 | Determine whether we still (intend to) use Engaging Networks - There is a Gravityforms Engaging Networks addon defined as a dependency of our staging website. However, the legacy website does not contain the same plugin.
## Task
- [ ] determine whether we still use or intend to use the Engaging Networks service
- [ ] decide on whether or not to include the Gravityforms Engaging Networks addon based on our intentions | non_priority | determine whether we still intend to use engaging networks there is a gravityforms engaging networks addon defined as a dependency of our staging website however the legacy website does not contain the same plugin task determine whether we still use or intend to use the engaging networks service decide on whether or not to include the gravityforms engaging networks addon based on our intentions | 0 |
157,626 | 13,697,764,116 | IssuesEvent | 2020-10-01 04:01:24 | SE-Group4/PhotoGallery | https://api.github.com/repos/SE-Group4/PhotoGallery | opened | [1 mark] Create Backlog for All Three Sprints | documentation | [1 mark] Backlog for all three sprints.
Please identify team member(s) to whom each backlog item has been assigned to. Identify the Sprint I backlog items that you were not been able to complete by the end of the Sprint I.
A backlog is basically a prioritized list of concrete tasks assigned to individual members of the team to help implement the requested features. Each task shouldn’t generally take more than 4-5 hours to complete per one or two team members. | 1.0 | [1 mark] Create Backlog for All Three Sprints - [1 mark] Backlog for all three sprints.
Please identify team member(s) to whom each backlog item has been assigned to. Identify the Sprint I backlog items that you were not been able to complete by the end of the Sprint I.
A backlog is basically a prioritized list of concrete tasks assigned to individual members of the team to help implement the requested features. Each task shouldn’t generally take more than 4-5 hours to complete per one or two team members. | non_priority | create backlog for all three sprints backlog for all three sprints please identify team member s to whom each backlog item has been assigned to identify the sprint i backlog items that you were not been able to complete by the end of the sprint i a backlog is basically a prioritized list of concrete tasks assigned to individual members of the team to help implement the requested features each task shouldn’t generally take more than hours to complete per one or two team members | 0 |
343,460 | 10,330,819,606 | IssuesEvent | 2019-09-02 15:40:35 | bbc/simorgh | https://api.github.com/repos/bbc/simorgh | closed | Add cookie check to mPulse beacon | high priority simorgh-core-stream | **Is your feature request related to a problem? Please describe.**
If a user has disabled performance cookies, the mPulse beacon will still be shown, therefore not respecting a user's privacy settings.
**Describe the solution you'd like**
Check that the user has enabled cookies (exact perferences/type of cookie to check TBD) before rendering the mPulse beacon.
Check the Dynatrace Beacon in RN to see which cookie was required to be set before loading the beacon.
**Describe alternatives you've considered**
N/A
**Testing notes**
[Tester to complete]
Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc
**Additional context**
Add any other context or screenshots about the feature request here.
- [x] Initially labelled with ["Refinement needed"](https://github.com/bbc/simorgh/labels/Refinement%20Needed)
| 1.0 | Add cookie check to mPulse beacon - **Is your feature request related to a problem? Please describe.**
If a user has disabled performance cookies, the mPulse beacon will still be shown, therefore not respecting a user's privacy settings.
**Describe the solution you'd like**
Check that the user has enabled cookies (exact perferences/type of cookie to check TBD) before rendering the mPulse beacon.
Check the Dynatrace Beacon in RN to see which cookie was required to be set before loading the beacon.
**Describe alternatives you've considered**
N/A
**Testing notes**
[Tester to complete]
Dev insight: Will Cypress tests be required or are unit tests sufficient? Will there be any potential regression? etc
**Additional context**
Add any other context or screenshots about the feature request here.
- [x] Initially labelled with ["Refinement needed"](https://github.com/bbc/simorgh/labels/Refinement%20Needed)
| priority | add cookie check to mpulse beacon is your feature request related to a problem please describe if a user has disabled performance cookies the mpulse beacon will still be shown therefore not respecting a user s privacy settings describe the solution you d like check that the user has enabled cookies exact perferences type of cookie to check tbd before rendering the mpulse beacon check the dynatrace beacon in rn to see which cookie was required to be set before loading the beacon describe alternatives you ve considered n a testing notes dev insight will cypress tests be required or are unit tests sufficient will there be any potential regression etc additional context add any other context or screenshots about the feature request here initially labelled with | 1 |
324,570 | 9,905,480,169 | IssuesEvent | 2019-06-27 11:42:33 | kudobuilder/kudo | https://api.github.com/repos/kudobuilder/kudo | closed | CreateOrUpdate function fix | component/operator kind/bug priority/high | In the plan controller, the line:
```
result, err := controllerutil.CreateOrUpdate(context.TODO(), r.Client, obj, func(runtime.Object) error { return nil })
```
needs to be fixed. The last argument of this function is supposed to capture the modifications to the object pulled from the server. Need to replace it with something like this from instance_controller.go
```
did, err := controllerutil.CreateOrUpdate(context.TODO(), mgr.GetClient(), current, func(o runtime.Object) error {
t := true
o.(*maestrov1alpha1.PlanExecution).Spec.Suspend = &t
return nil
})
``` | 1.0 | CreateOrUpdate function fix - In the plan controller, the line:
```
result, err := controllerutil.CreateOrUpdate(context.TODO(), r.Client, obj, func(runtime.Object) error { return nil })
```
needs to be fixed. The last argument of this function is supposed to capture the modifications to the object pulled from the server. Need to replace it with something like this from instance_controller.go
```
did, err := controllerutil.CreateOrUpdate(context.TODO(), mgr.GetClient(), current, func(o runtime.Object) error {
t := true
o.(*maestrov1alpha1.PlanExecution).Spec.Suspend = &t
return nil
})
``` | priority | createorupdate function fix in the plan controller the line result err controllerutil createorupdate context todo r client obj func runtime object error return nil needs to be fixed the last argument of this function is supposed to capture the modifications to the object pulled from the server need to replace it with something like this from instance controller go did err controllerutil createorupdate context todo mgr getclient current func o runtime object error t true o planexecution spec suspend t return nil | 1 |
499,526 | 14,449,290,741 | IssuesEvent | 2020-12-08 07:51:35 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Incorrect stability check | platform:azure priority-3-normal reproduction needed type:bug | **What Renovate type, platform and version are you using?**
Self Hosted, Azure DevOps
**Describe the bug**
Renovate doesn't create PRs even if the update met stability days requirements.
**Relevant debug logs**
```
DEBUG: Updated 2 package files (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: No updated lock files in branch (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: 2 file(s) to commit (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: Committing files to branch renovate/microsoft-applicationinsights-packages (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
INFO: Branch created (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
"commitSha": "067fd1f"
DEBUG: Updating renovate/stability-days status check state to green (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: setBranchStatus(renovate/microsoft-applicationinsights-packages, renovate/stability-days, Updates have met stability days requirement, green, https://docs.renovatebot.com/) (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: Branch status pending (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
"commitSha": "067fd1f"
```
**To Reproduce**
1. Set stabilityDays: X
2. Set prCreation: "not-pending"
3. Add dependency that pass stabilityDays period.
4. Run renovate
**Additional context**
I think it's somehow related to this bug #4884
| 1.0 | Incorrect stability check - **What Renovate type, platform and version are you using?**
Self Hosted, Azure DevOps
**Describe the bug**
Renovate doesn't create PRs even if the update met stability days requirements.
**Relevant debug logs**
```
DEBUG: Updated 2 package files (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: No updated lock files in branch (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: 2 file(s) to commit (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: Committing files to branch renovate/microsoft-applicationinsights-packages (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
INFO: Branch created (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
"commitSha": "067fd1f"
DEBUG: Updating renovate/stability-days status check state to green (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: setBranchStatus(renovate/microsoft-applicationinsights-packages, renovate/stability-days, Updates have met stability days requirement, green, https://docs.renovatebot.com/) (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
DEBUG: Branch status pending (repository=MyCompany/ProjectX, branch=renovate/microsoft-applicationinsights-packages)
"commitSha": "067fd1f"
```
**To Reproduce**
1. Set stabilityDays: X
2. Set prCreation: "not-pending"
3. Add dependency that pass stabilityDays period.
4. Run renovate
**Additional context**
I think it's somehow related to this bug #4884
| priority | incorrect stability check what renovate type platform and version are you using self hosted azure devops describe the bug renovate doesn t create prs even if the update met stability days requirements relevant debug logs debug updated package files repository mycompany projectx branch renovate microsoft applicationinsights packages debug no updated lock files in branch repository mycompany projectx branch renovate microsoft applicationinsights packages debug file s to commit repository mycompany projectx branch renovate microsoft applicationinsights packages debug committing files to branch renovate microsoft applicationinsights packages repository mycompany projectx branch renovate microsoft applicationinsights packages info branch created repository mycompany projectx branch renovate microsoft applicationinsights packages commitsha debug updating renovate stability days status check state to green repository mycompany projectx branch renovate microsoft applicationinsights packages debug setbranchstatus renovate microsoft applicationinsights packages renovate stability days updates have met stability days requirement green repository mycompany projectx branch renovate microsoft applicationinsights packages debug branch status pending repository mycompany projectx branch renovate microsoft applicationinsights packages commitsha to reproduce set stabilitydays x set prcreation not pending add dependency that pass stabilitydays period run renovate additional context i think it s somehow related to this bug | 1 |
214,558 | 7,274,378,728 | IssuesEvent | 2018-02-21 09:48:24 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | clientauth.jwt jar version mentioned in the doc is different from the actual jar version created | Affected/5.5.0-Alpha Priority/High Type/Docs | clientauth.jwt jar version mentioned in the doc [1] is different from the actual jar version created.
[1] https://docs.wso2.com/display/IS550/Private+Key+JWT+Client+Authentication+for+OIDC
It says,
Place the target/org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.2-SNAPSHOT.jar in the <IS_HOME>/repository/component/dropins directory.
But the actual jar created is as,
org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.3-SNAPSHOT.jar | 1.0 | clientauth.jwt jar version mentioned in the doc is different from the actual jar version created - clientauth.jwt jar version mentioned in the doc [1] is different from the actual jar version created.
[1] https://docs.wso2.com/display/IS550/Private+Key+JWT+Client+Authentication+for+OIDC
It says,
Place the target/org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.2-SNAPSHOT.jar in the <IS_HOME>/repository/component/dropins directory.
But the actual jar created is as,
org.wso2.carbon.identity.oauth2.token.handler.clientauth.jwt-1.0.3-SNAPSHOT.jar | priority | clientauth jwt jar version mentioned in the doc is different from the actual jar version created clientauth jwt jar version mentioned in the doc is different from the actual jar version created it says place the target org carbon identity token handler clientauth jwt snapshot jar in the repository component dropins directory but the actual jar created is as org carbon identity token handler clientauth jwt snapshot jar | 1 |
372,812 | 11,028,587,565 | IssuesEvent | 2019-12-06 12:03:10 | coder3101/cp-editor2 | https://api.github.com/repos/coder3101/cp-editor2 | closed | Showing compiler warnings | enhancement high_priority linux macOs windows | **Is your feature request related to a problem? Please describe.**
I can't see the compiler warnings.
**Describe the solution you'd like**
Show compiler warnings after compiling.
Maybe add a setting of whether to show the warnings or not. (However, this can be done by adding `-w` in the compile command.)
**Describe alternatives you've considered**
N/A
**Additional context**
N/A
| 1.0 | Showing compiler warnings - **Is your feature request related to a problem? Please describe.**
I can't see the compiler warnings.
**Describe the solution you'd like**
Show compiler warnings after compiling.
Maybe add a setting of whether to show the warnings or not. (However, this can be done by adding `-w` in the compile command.)
**Describe alternatives you've considered**
N/A
**Additional context**
N/A
| priority | showing compiler warnings is your feature request related to a problem please describe i can t see the compiler warnings describe the solution you d like show compiler warnings after compiling maybe add a setting of whether to show the warnings or not however this can be done by adding w in the compile command describe alternatives you ve considered n a additional context n a | 1 |
303,193 | 22,958,859,177 | IssuesEvent | 2022-07-19 13:52:23 | fga-eps-mds/UnbFlow | https://api.github.com/repos/fga-eps-mds/UnbFlow | closed | Docs: Escrever historias de usuário | documentation | # Descrição
Criar as histórias de usuário com base nos epicos, features e no prototipo de média fidelidade
## Critério de aceitação
- [x] Escrever histórias de usuário que serão realizadas durante as issues | 1.0 | Docs: Escrever historias de usuário - # Descrição
Criar as histórias de usuário com base nos epicos, features e no prototipo de média fidelidade
## Critério de aceitação
- [x] Escrever histórias de usuário que serão realizadas durante as issues | non_priority | docs escrever historias de usuário descrição criar as histórias de usuário com base nos epicos features e no prototipo de média fidelidade critério de aceitação escrever histórias de usuário que serão realizadas durante as issues | 0 |
186,220 | 6,734,466,543 | IssuesEvent | 2017-10-18 18:08:56 | octobercms/october | https://api.github.com/repos/octobercms/october | closed | Using Artisan::add in init.php does not work | Priority: Medium Status: Review Needed Type: Unconfirmed Bug | https://octobercms.com/docs/console/development states that you should be able to write `Artisan::add(new Acme\Blog\Console\MyCommand);` in `init.php` and the command will be added
Instead I get an exception
```
[Symfony\Component\Debug\Exception\FatalErrorException]
Call to undefined method October\Rain\Foundation\Console\Kernel::add()
```
| 1.0 | Using Artisan::add in init.php does not work - https://octobercms.com/docs/console/development states that you should be able to write `Artisan::add(new Acme\Blog\Console\MyCommand);` in `init.php` and the command will be added
Instead I get an exception
```
[Symfony\Component\Debug\Exception\FatalErrorException]
Call to undefined method October\Rain\Foundation\Console\Kernel::add()
```
| priority | using artisan add in init php does not work states that you should be able to write artisan add new acme blog console mycommand in init php and the command will be added instead i get an exception call to undefined method october rain foundation console kernel add | 1 |
18,777 | 24,678,890,034 | IssuesEvent | 2022-10-18 19:25:58 | dtcenter/MET | https://api.github.com/repos/dtcenter/MET | opened | Investigate `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml | type: bug alert: NEED ACCOUNT KEY requestor: METplus Team MET: PreProcessing Tools (Point) priority: high | ## Describe the Problem ##
During review of #2294 for issue #2276, a problem was discovered in the output of the `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml. The output file created by this test (HourlyData_20220312.nc) contains values of Infinity (`Inf`). While the GHA run for that PR did increase the occurrence of Inf in the output, the problem existed prior to those code changes.
This issue is to investigate the source of the `Inf` values appearing in the output, and fix the code to avoid them.
### Expected Behavior ###
The output of ascii2nc should never contain a value of infinity. The code should be enhanced by adding more error checking to avoid them. Perhaps, they should be reported as bad data value (i.e. -9999) rather than `Inf`?
### Environment ###
Describe your runtime environment:
*1. Visible in the output of GHA and in the output of the MET nightly build on seneca.*
### To Reproduce ###
Describe the steps to reproduce the behavior:
*1. Log on to 'seneca'*
*2. Go to NB area:*
```
cd /d1/projects/MET/MET_regression/develop/NB20221018
```
*3. Dump to ascii:
```
Rscript MET-develop/scripts/Rscripts/pntnc2ascii.R MET-develop/test_output/ascii2nc/airnow/HourlyData_20220312.nc > HourlyData_20220312.txt
```
*4. See error in columns 6 and 9 of the output:*
```
grep Inf HourlyData_20220312.txt | wc -l
33
```
*Post relevant sample data following these instructions:*
*https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Organization** level **Project** for support of the current coordinated release
- [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next bugfix version
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Bugfix Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **main_\<Version>**.
Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>`
- [ ] Fix the bug and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **main_\<Version>**.
Pull request: `bugfix <Issue Number> main_<Version> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Organization** level software support **Project** for the current coordinated release
Select: **Milestone** as the next bugfix version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Complete the steps above to fix the bug on the **develop** branch.
Branch name: `bugfix_<Issue Number>_develop_<Description>`
Pull request: `bugfix <Issue Number> develop <Description>`
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Close this issue.
| 1.0 | Investigate `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml - ## Describe the Problem ##
During review of #2294 for issue #2276, a problem was discovered in the output of the `ascii2nc_airnow_hourly` test in unit_ascii2nc.xml. The output file created by this test (HourlyData_20220312.nc) contains values of Infinity (`Inf`). While the GHA run for that PR did increase the occurrence of Inf in the output, the problem existed prior to those code changes.
This issue is to investigate the source of the `Inf` values appearing in the output, and fix the code to avoid them.
### Expected Behavior ###
The output of ascii2nc should never contain a value of infinity. The code should be enhanced by adding more error checking to avoid them. Perhaps, they should be reported as bad data value (i.e. -9999) rather than `Inf`?
### Environment ###
Describe your runtime environment:
*1. Visible in the output of GHA and in the output of the MET nightly build on seneca.*
### To Reproduce ###
Describe the steps to reproduce the behavior:
*1. Log on to 'seneca'*
*2. Go to NB area:*
```
cd /d1/projects/MET/MET_regression/develop/NB20221018
```
*3. Dump to ascii:
```
Rscript MET-develop/scripts/Rscripts/pntnc2ascii.R MET-develop/test_output/ascii2nc/airnow/HourlyData_20220312.nc > HourlyData_20220312.txt
```
*4. See error in columns 6 and 9 of the output:*
```
grep Inf HourlyData_20220312.txt | wc -l
33
```
*Post relevant sample data following these instructions:*
*https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Organization** level **Project** for support of the current coordinated release
- [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next bugfix version
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Bugfix Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **main_\<Version>**.
Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>`
- [ ] Fix the bug and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **main_\<Version>**.
Pull request: `bugfix <Issue Number> main_<Version> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Organization** level software support **Project** for the current coordinated release
Select: **Milestone** as the next bugfix version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Complete the steps above to fix the bug on the **develop** branch.
Branch name: `bugfix_<Issue Number>_develop_<Description>`
Pull request: `bugfix <Issue Number> develop <Description>`
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Close this issue.
| non_priority | investigate airnow hourly test in unit xml describe the problem during review of for issue a problem was discovered in the output of the airnow hourly test in unit xml the output file created by this test hourlydata nc contains values of infinity inf while the gha run for that pr did increase the occurrence of inf in the output the problem existed prior to those code changes this issue is to investigate the source of the inf values appearing in the output and fix the code to avoid them expected behavior the output of should never contain a value of infinity the code should be enhanced by adding more error checking to avoid them perhaps they should be reported as bad data value i e rather than inf environment describe your runtime environment visible in the output of gha and in the output of the met nightly build on seneca to reproduce describe the steps to reproduce the behavior log on to seneca go to nb area cd projects met met regression develop dump to ascii rscript met develop scripts rscripts r met develop test output airnow hourlydata nc hourlydata txt see error in columns and of the output grep inf hourlydata txt wc l post relevant sample data following these instructions relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and linked issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version close this issue | 0 |
566,801 | 16,830,786,287 | IssuesEvent | 2021-06-18 04:14:50 | CertifaiAI/classifai | https://api.github.com/repos/CertifaiAI/classifai | opened | Backend language support | bug low priority | **Describe the bug**
As ClassifAI is supporting multiple languages, The title of UI prompted from backend is still in english.
**Expected behavior**
The language should be dynamic in backend as well
**Screenshots**

**Desktop (please complete the following information):**
- OS: Windows
- Browser chrome
- Version v2.0.0 alpha1
| 1.0 | Backend language support - **Describe the bug**
As ClassifAI is supporting multiple languages, The title of UI prompted from backend is still in english.
**Expected behavior**
The language should be dynamic in backend as well
**Screenshots**

**Desktop (please complete the following information):**
- OS: Windows
- Browser chrome
- Version v2.0.0 alpha1
| priority | backend language support describe the bug as classifai is supporting multiple languages the title of ui prompted from backend is still in english expected behavior the language should be dynamic in backend as well screenshots desktop please complete the following information os windows browser chrome version | 1 |
288,020 | 24,882,269,162 | IssuesEvent | 2022-10-28 03:01:02 | MPMG-DCC-UFMG/F01 | https://api.github.com/repos/MPMG-DCC-UFMG/F01 | closed | Teste de generalizacao para a tag Orçamento - Legislação - Lagamar | generalization test development template - Memory (66) tag - Orçamento subtag - Legislação | DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Lagamar. | 1.0 | Teste de generalizacao para a tag Orçamento - Legislação - Lagamar - DoD: Realizar o teste de Generalização do validador da tag Orçamento - Legislação para o Município de Lagamar. | non_priority | teste de generalizacao para a tag orçamento legislação lagamar dod realizar o teste de generalização do validador da tag orçamento legislação para o município de lagamar | 0 |
622,689 | 19,654,052,470 | IssuesEvent | 2022-01-10 10:34:47 | pgfmc/Core | https://api.github.com/repos/pgfmc/Core | closed | Fix when AFK deactivates | bug low priority | What needs to change:
- AFK toggle on teleport
- Don't AFK toggle on Y level descend | 1.0 | Fix when AFK deactivates - What needs to change:
- AFK toggle on teleport
- Don't AFK toggle on Y level descend | priority | fix when afk deactivates what needs to change afk toggle on teleport don t afk toggle on y level descend | 1 |
461,080 | 13,223,211,077 | IssuesEvent | 2020-08-17 16:49:19 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Apps can't be installed in latest master from Dashboard | [zube]: Next Up alpha-priority/0 kind/bug-qa | **What kind of request is this:**
**Steps to reproduce:**
- Install Rancher HA ( k8s `master-head (08/10/2020)` _502b1839e_)
- Navigate to the cluster Dashboard and Apps
**Result:**


**Other details that may be helpful:**
**Environment information**
- Rancher version: `master-head (08/10/2020)` _502b1839e_
- Installation option: HA k8s
| 1.0 | Apps can't be installed in latest master from Dashboard - **What kind of request is this:**
**Steps to reproduce:**
- Install Rancher HA ( k8s `master-head (08/10/2020)` _502b1839e_)
- Navigate to the cluster Dashboard and Apps
**Result:**


**Other details that may be helpful:**
**Environment information**
- Rancher version: `master-head (08/10/2020)` _502b1839e_
- Installation option: HA k8s
| priority | apps can t be installed in latest master from dashboard what kind of request is this steps to reproduce install rancher ha master head navigate to the cluster dashboard and apps result other details that may be helpful environment information rancher version master head installation option ha | 1 |
95,033 | 3,933,560,384 | IssuesEvent | 2016-04-25 19:33:40 | ghutchis/avogadro | https://api.github.com/repos/ghutchis/avogadro | closed | Avogadro crashes when exporting bitmap graphics on Mac. | auto-migrated high priority v_1.1.0 | Once correctly opened and visualized a .log or a .cube g09 file, I cannot export the visualized image using File->Export->Graphics... because the program suddenly crashes. Only the vectorial image can be exported, but the color are strongly altered.
Reported by: *anonymous | 1.0 | Avogadro crashes when exporting bitmap graphics on Mac. - Once correctly opened and visualized a .log or a .cube g09 file, I cannot export the visualized image using File->Export->Graphics... because the program suddenly crashes. Only the vectorial image can be exported, but the color are strongly altered.
Reported by: *anonymous | priority | avogadro crashes when exporting bitmap graphics on mac once correctly opened and visualized a log or a cube file i cannot export the visualized image using file export graphics because the program suddenly crashes only the vectorial image can be exported but the color are strongly altered reported by anonymous | 1 |
382,846 | 11,339,118,576 | IssuesEvent | 2020-01-23 00:43:55 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | record_function decorator in autograd profiler does not work with RPC | high priority module: rpc | ## 🐛 Bug
Nested `RecordFunction`s during profiling do not work with RPC calls, due to mismanagement of some internal variables in `record_function.cpp` caused by handling `RecordFunction` objects completing in different hreads.
## To Reproduce
```
with torch.autograd.profiler.profiler() as prof:
with torch.autograd.profiler.record_function("foo"):
rpc.rpc_async(...)
```
will result in an error:
```
> I0122 16:32:15.967217 594065 record_function.cpp:214] Exception in RecordFunction::end(): (thread_local_func_ == thi
s) || (thread_local_func_ == nullptr && threadId_ != 0) INTERNAL ASSERT FAILED at caffe2/torch/csrc/autograd/record_fu
nction.cpp:235, please report a bug to PyTorch. baze: must be top of stack. If you are calling RecordFunction::end in
aseparate thread, call RecordFunction::setThreadId() in the creatingthread. (end at caffe2/torch/csrc/autograd/record_
function.cpp:235)
```
This is because the handling of the thread local variables `thread_local_func` and `parent_` are not managed properly when `RecordFunction::end` is called from a separate thread.
Expected behavior:
The nested scopes should be profiled correctly as non-RPC code would be. | 1.0 | record_function decorator in autograd profiler does not work with RPC - ## 🐛 Bug
Nested `RecordFunction`s during profiling do not work with RPC calls, due to mismanagement of some internal variables in `record_function.cpp` caused by handling `RecordFunction` objects completing in different hreads.
## To Reproduce
```
with torch.autograd.profiler.profiler() as prof:
with torch.autograd.profiler.record_function("foo"):
rpc.rpc_async(...)
```
will result in an error:
```
> I0122 16:32:15.967217 594065 record_function.cpp:214] Exception in RecordFunction::end(): (thread_local_func_ == thi
s) || (thread_local_func_ == nullptr && threadId_ != 0) INTERNAL ASSERT FAILED at caffe2/torch/csrc/autograd/record_fu
nction.cpp:235, please report a bug to PyTorch. baze: must be top of stack. If you are calling RecordFunction::end in
aseparate thread, call RecordFunction::setThreadId() in the creatingthread. (end at caffe2/torch/csrc/autograd/record_
function.cpp:235)
```
This is because the handling of the thread local variables `thread_local_func` and `parent_` are not managed properly when `RecordFunction::end` is called from a separate thread.
Expected behavior:
The nested scopes should be profiled correctly as non-RPC code would be. | priority | record function decorator in autograd profiler does not work with rpc 🐛 bug nested recordfunction s during profiling do not work with rpc calls due to mismanagement of some internal variables in record function cpp caused by handling recordfunction objects completing in different hreads to reproduce with torch autograd profiler profiler as prof with torch autograd profiler record function foo rpc rpc async will result in an error record function cpp exception in recordfunction end thread local func thi s thread local func nullptr threadid internal assert failed at torch csrc autograd record fu nction cpp please report a bug to pytorch baze must be top of stack if you are calling recordfunction end in aseparate thread call recordfunction setthreadid in the creatingthread end at torch csrc autograd record function cpp this is because the handling of the thread local variables thread local func and parent are not managed properly when recordfunction end is called from a separate thread expected behavior the nested scopes should be profiled correctly as non rpc code would be | 1 |
15,368 | 3,461,414,456 | IssuesEvent | 2015-12-20 02:02:27 | ehmorris/Facebook-Mood | https://api.github.com/repos/ehmorris/Facebook-Mood | closed | Extension tests | extension-related testing | This is actually going to be kinda cool. Writing a testing module for javascript as a chrome extension. | 1.0 | Extension tests - This is actually going to be kinda cool. Writing a testing module for javascript as a chrome extension. | non_priority | extension tests this is actually going to be kinda cool writing a testing module for javascript as a chrome extension | 0 |
187,005 | 21,993,039,292 | IssuesEvent | 2022-05-26 01:22:44 | raindigi/GraphqlType-API-Registration | https://api.github.com/repos/raindigi/GraphqlType-API-Registration | opened | CVE-2022-24434 (High) detected in dicer-0.3.0.tgz | security vulnerability | ## CVE-2022-24434 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dicer-0.3.0.tgz</b></p></summary>
<p>A very fast streaming multipart parser for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz">https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz</a></p>
<p>Path to dependency file: /GraphqlType-API-Registration/package.json</p>
<p>Path to vulnerable library: /node_modules/dicer/package.json</p>
<p>
Dependency Hierarchy:
- apollo-server-express-2.3.3.tgz (Root Library)
- apollo-server-core-2.3.3.tgz
- graphql-upload-8.0.4.tgz
- busboy-0.3.0.tgz
- :x: **dicer-0.3.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/raindigi/GraphqlType-API-Registration/commit/ddbf6735cb305c07ce82a0faefa069bd89b7581f">ddbf6735cb305c07ce82a0faefa069bd89b7581f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects all versions of package dicer.
A malicious attacker can send a modified form to server, and crash the nodejs service. An attacker could sent the payload again and again so that the service continuously crashes.
<p>Publish Date: 2022-05-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24434>CVE-2022-24434</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-24434">https://nvd.nist.gov/vuln/detail/CVE-2022-24434</a></p>
<p>Release Date: 2022-05-20</p>
<p>Fix Resolution: no_fix</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-24434 (High) detected in dicer-0.3.0.tgz - ## CVE-2022-24434 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dicer-0.3.0.tgz</b></p></summary>
<p>A very fast streaming multipart parser for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz">https://registry.npmjs.org/dicer/-/dicer-0.3.0.tgz</a></p>
<p>Path to dependency file: /GraphqlType-API-Registration/package.json</p>
<p>Path to vulnerable library: /node_modules/dicer/package.json</p>
<p>
Dependency Hierarchy:
- apollo-server-express-2.3.3.tgz (Root Library)
- apollo-server-core-2.3.3.tgz
- graphql-upload-8.0.4.tgz
- busboy-0.3.0.tgz
- :x: **dicer-0.3.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/raindigi/GraphqlType-API-Registration/commit/ddbf6735cb305c07ce82a0faefa069bd89b7581f">ddbf6735cb305c07ce82a0faefa069bd89b7581f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects all versions of package dicer.
A malicious attacker can send a modified form to server, and crash the nodejs service. An attacker could sent the payload again and again so that the service continuously crashes.
<p>Publish Date: 2022-05-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24434>CVE-2022-24434</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-24434">https://nvd.nist.gov/vuln/detail/CVE-2022-24434</a></p>
<p>Release Date: 2022-05-20</p>
<p>Fix Resolution: no_fix</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in dicer tgz cve high severity vulnerability vulnerable library dicer tgz a very fast streaming multipart parser for node js library home page a href path to dependency file graphqltype api registration package json path to vulnerable library node modules dicer package json dependency hierarchy apollo server express tgz root library apollo server core tgz graphql upload tgz busboy tgz x dicer tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects all versions of package dicer a malicious attacker can send a modified form to server and crash the nodejs service an attacker could sent the payload again and again so that the service continuously crashes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution no fix step up your open source security game with whitesource | 0 |
50,519 | 13,187,554,369 | IssuesEvent | 2020-08-13 03:47:37 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | test ticket (Trac #869) | Migrated from Trac cmake defect |
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/869
, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-12T06:31:40",
"description": "",
"reporter": "nega",
"cc": "",
"resolution": "invalid",
"_ts": "1423722700498868",
"component": "cmake",
"summary": "test ticket",
"priority": "normal",
"keywords": "",
"time": "2015-02-11T23:11:24",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | test ticket (Trac #869) -
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/869
, reported by nega and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-02-12T06:31:40",
"description": "",
"reporter": "nega",
"cc": "",
"resolution": "invalid",
"_ts": "1423722700498868",
"component": "cmake",
"summary": "test ticket",
"priority": "normal",
"keywords": "",
"time": "2015-02-11T23:11:24",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| non_priority | test ticket trac migrated from reported by nega and owned by nega json status closed changetime description reporter nega cc resolution invalid ts component cmake summary test ticket priority normal keywords time milestone owner nega type defect | 0 |
810,198 | 30,229,778,490 | IssuesEvent | 2023-07-06 05:44:42 | glific/mobile | https://api.github.com/repos/glific/mobile | closed | Start a flow for a collection | Priority: High | **Start a flow for a collection**
Implement 'Start a flow' in Collection Chat Screen Option. When selected from options, a dialog box will appear, providing users with the ability to choose a flow and initiate it using the "Start" button or cancel the action with the "Cancel" button.
**Approach**
- Implement dialog box component according to design.
- Implement a query to fetch all the available flows from the backend.
- Implement logic to start the chosen flow when the "Start" button is pressed.
Figma link: https://www.figma.com/file/SbP0ru8kiB2l334ieX1ESe/Mobile-app?type=design&node-id=754-5118&mode=design&t=DfFgoqouNYlgv68A-0
Get all flows API: https://api.glific.com/#e0f1066e-471b-47de-b83d-0003387e9032
```sh
variables = {
filter: {
status: "published",
isActive: true,
},
opts: {
limit: null,
offset: 0,
order: "ASC",
}
}
```
Start a flow API: https://api.glific.com/#cb956d81-9ce5-4101-ad5f-2c65402c4171
```sh
variables = {
flowId: "2054",
groupId: "2",
}
```
| 1.0 | Start a flow for a collection - **Start a flow for a collection**
Implement 'Start a flow' in Collection Chat Screen Option. When selected from options, a dialog box will appear, providing users with the ability to choose a flow and initiate it using the "Start" button or cancel the action with the "Cancel" button.
**Approach**
- Implement dialog box component according to design.
- Implement a query to fetch all the available flows from the backend.
- Implement logic to start the chosen flow when the "Start" button is pressed.
Figma link: https://www.figma.com/file/SbP0ru8kiB2l334ieX1ESe/Mobile-app?type=design&node-id=754-5118&mode=design&t=DfFgoqouNYlgv68A-0
Get all flows API: https://api.glific.com/#e0f1066e-471b-47de-b83d-0003387e9032
```sh
variables = {
filter: {
status: "published",
isActive: true,
},
opts: {
limit: null,
offset: 0,
order: "ASC",
}
}
```
Start a flow API: https://api.glific.com/#cb956d81-9ce5-4101-ad5f-2c65402c4171
```sh
variables = {
flowId: "2054",
groupId: "2",
}
```
| priority | start a flow for a collection start a flow for a collection implement start a flow in collection chat screen option when selected from options a dialog box will appear providing users with the ability to choose a flow and initiate it using the start button or cancel the action with the cancel button approach implement dialog box component according to design implement a query to fetch all the available flows from the backend implement logic to start the chosen flow when the start button is pressed figma link get all flows api sh variables filter status published isactive true opts limit null offset order asc start a flow api sh variables flowid groupid | 1 |
665,016 | 22,296,070,020 | IssuesEvent | 2022-06-13 01:49:26 | ESIPFed/Geoweaver | https://api.github.com/repos/ESIPFed/Geoweaver | closed | Improve the weaver GUI | enhancement high-priority | Some existing problems:
- The words inside the circles are hardly readable
- The process names often go out of bounds
- The status indicators are not easy to understand the color change
- The toolbar is not self explainable
- The process status bars are not indicating anything
- ... | 1.0 | Improve the weaver GUI - Some existing problems:
- The words inside the circles are hardly readable
- The process names often go out of bounds
- The status indicators are not easy to understand the color change
- The toolbar is not self explainable
- The process status bars are not indicating anything
- ... | priority | improve the weaver gui some existing problems the words inside the circles are hardly readable the process names often go out of bounds the status indicators are not easy to understand the color change the toolbar is not self explainable the process status bars are not indicating anything | 1 |
273,400 | 8,530,235,069 | IssuesEvent | 2018-11-03 20:20:27 | projectcalico/calico | https://api.github.com/repos/projectcalico/calico | closed | Document need for `iptables -P FORWARD ACCEPT` for IPVS until K8s fixes are merged | area/docs/content content/needed priority/P3 | Currently, when using kube-proxy in IPVS mode, for NodePorts to work (when workloads are on a host other than the host initially being connected to) it is necessary to run `iptables -P FORWARD ACCEPT` on all nodes. The issue is raised upstream https://github.com/kubernetes/kubernetes/issues/59656 and the fix is up as this PR https://github.com/kubernetes/kubernetes/pull/62007.
## Expected Behavior
Should document the need for the above at least until the fix is in a K8s release.
## Current Behavior
Not documented.
/cc @song-jiang | 1.0 | Document need for `iptables -P FORWARD ACCEPT` for IPVS until K8s fixes are merged - Currently, when using kube-proxy in IPVS mode, for NodePorts to work (when workloads are on a host other than the host initially being connected to) it is necessary to run `iptables -P FORWARD ACCEPT` on all nodes. The issue is raised upstream https://github.com/kubernetes/kubernetes/issues/59656 and the fix is up as this PR https://github.com/kubernetes/kubernetes/pull/62007.
## Expected Behavior
Should document the need for the above at least until the fix is in a K8s release.
## Current Behavior
Not documented.
/cc @song-jiang | priority | document need for iptables p forward accept for ipvs until fixes are merged currently when using kube proxy in ipvs mode for nodeports to work when workloads are on a host other than the host initially being connected to it is necessary to run iptables p forward accept on all nodes the issue is raised upstream and the fix is up as this pr expected behavior should document the need for the above at least until the fix is in a release current behavior not documented cc song jiang | 1 |
702,123 | 24,120,640,480 | IssuesEvent | 2022-09-20 18:22:51 | googleapis/nodejs-analytics-data | https://api.github.com/repos/googleapis/nodejs-analytics-data | closed | Realtime report with multiple dimensions: should run realtime with multiple dimensions failed | type: bug priority: p1 api: analyticsdata flakybot: issue | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: ae646a54bc64abff0cf92625117ffb258e303e8b
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c), [Sponge](http://sponge2/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
Error: Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/runRealtimeReportWithMultipleDimensions.test.js:24:28)
at Context.<anonymous> (test/runRealtimeReportWithMultipleDimensions.test.js:31:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | 1.0 | Realtime report with multiple dimensions: should run realtime with multiple dimensions failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: ae646a54bc64abff0cf92625117ffb258e303e8b
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c), [Sponge](http://sponge2/2d9c997b-ceda-4ff3-895e-1cdaa46f3a8c)
status: failed
<details><summary>Test output</summary><br><pre>Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
Error: Command failed: node ./runRealtimeReportWithMultipleDimensions.js 222596558
16 UNAUTHENTICATED: Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.
at checkExecSyncError (child_process.js:635:11)
at Object.execSync (child_process.js:671:15)
at execSync (test/runRealtimeReportWithMultipleDimensions.test.js:24:28)
at Context.<anonymous> (test/runRealtimeReportWithMultipleDimensions.test.js:31:20)
at processImmediate (internal/timers.js:461:21)</pre></details> | priority | realtime report with multiple dimensions should run realtime with multiple dimensions failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output command failed node runrealtimereportwithmultipledimensions js unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see error command failed node runrealtimereportwithmultipledimensions js unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see at checkexecsyncerror child process js at object execsync child process js at execsync test runrealtimereportwithmultipledimensions test js at context test runrealtimereportwithmultipledimensions test js at processimmediate internal timers js | 1 |
121,118 | 15,857,163,455 | IssuesEvent | 2021-04-08 04:07:14 | asoffer/Icarus | https://api.github.com/repos/asoffer/Icarus | closed | Should array types allow implicit integer casts? | design decision | Currently array lengths need to be integral but will be implicitly cast to the underlying array-type's length implementation type. This is currently a `uint64_t` so any natively supported type in Icarus can convert to this losslessly (because negative lengths are prohibited). However, if Icarus starts supporting larger integer types this cast will be lossy and implicit.
This issue tracks simply making a decision here.
There is a nice consistency in always requiring the cast. There is also no reason to support arrays of length 2^64 or greater in any reasonable future. | 1.0 | Should array types allow implicit integer casts? - Currently array lengths need to be integral but will be implicitly cast to the underlying array-type's length implementation type. This is currently a `uint64_t` so any natively supported type in Icarus can convert to this losslessly (because negative lengths are prohibited). However, if Icarus starts supporting larger integer types this cast will be lossy and implicit.
This issue tracks simply making a decision here.
There is a nice consistency in always requiring the cast. There is also no reason to support arrays of length 2^64 or greater in any reasonable future. | non_priority | should array types allow implicit integer casts currently array lengths need to be integral but will be implicitly cast to the underlying array type s length implementation type this is currently a t so any natively supported type in icarus can convert to this losslessly because negative lengths are prohibited however if icarus starts supporting larger integer types this cast will be lossy and implicit this issue tracks simply making a decision here there is a nice consistency in always requiring the cast there is also no reason to support arrays of length or greater in any reasonable future | 0 |
140,427 | 5,408,859,972 | IssuesEvent | 2017-03-01 01:33:53 | google/roboto | https://api.github.com/repos/google/roboto | opened | Roboto Mono 'l' legibility | Priority-High | Requested change:
> Change the lowercase L in Roboto Mono to a reverse S shape, like Andale Mono or Monaco have; Roboto Mono has clearer distinction between lowercase L and 1 than Courier does, but the reverse-S type of L is clearer still.
| 1.0 | Roboto Mono 'l' legibility - Requested change:
> Change the lowercase L in Roboto Mono to a reverse S shape, like Andale Mono or Monaco have; Roboto Mono has clearer distinction between lowercase L and 1 than Courier does, but the reverse-S type of L is clearer still.
| priority | roboto mono l legibility requested change change the lowercase l in roboto mono to a reverse s shape like andale mono or monaco have roboto mono has clearer distinction between lowercase l and than courier does but the reverse s type of l is clearer still | 1 |
817,370 | 30,639,264,360 | IssuesEvent | 2023-07-24 20:24:25 | GoogleCloudPlatform/python-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples | closed | speech.snippets.transcribe_gcs_v2_test: test_transcribe_gcs_v2 failed | priority: p2 type: bug api: speech samples flakybot: issue flakybot: flaky | Note: #9423 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 16ab3cf93caddb092710ba507d537945128c884c
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/297c18a8-be85-43ab-92b9-655f63a7de99), [Sponge](http://sponge2/297c18a8-be85-43ab-92b9-655f63a7de99)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 1030, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking
raise _InactiveRpcError(state) # pytype: disable=not-instantiable
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "502:Bad Gateway"
debug_error_string = "UNKNOWN:Error received from peer {grpc_message:"502:Bad Gateway", grpc_status:14, created_time:"2023-06-17T13:39:31.187553795+00:00"}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 191, in retry_target
return target()
File "/workspace/speech/snippets/transcribe_gcs_v2_test.py", line 38, in test_transcribe_gcs_v2
response = transcribe_gcs_v2.transcribe_gcs_v2(
File "/workspace/speech/snippets/transcribe_gcs_v2.py", line 48, in transcribe_gcs_v2
operation = client.create_recognizer(request=request)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/speech_v2/services/speech/client.py", line 695, in create_recognizer
response = rpc(
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.ServiceUnavailable: 503 502:Bad Gateway
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 349, in retry_wrapped_func
return retry_target(
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 207, in retry_target
raise exceptions.RetryError(
google.api_core.exceptions.RetryError: Deadline of 120.0s exceeded while calling target function, last exception: 503 502:Bad Gateway</pre></details> | 1.0 | speech.snippets.transcribe_gcs_v2_test: test_transcribe_gcs_v2 failed - Note: #9423 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 16ab3cf93caddb092710ba507d537945128c884c
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/297c18a8-be85-43ab-92b9-655f63a7de99), [Sponge](http://sponge2/297c18a8-be85-43ab-92b9-655f63a7de99)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 1030, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking
raise _InactiveRpcError(state) # pytype: disable=not-instantiable
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.UNAVAILABLE
details = "502:Bad Gateway"
debug_error_string = "UNKNOWN:Error received from peer {grpc_message:"502:Bad Gateway", grpc_status:14, created_time:"2023-06-17T13:39:31.187553795+00:00"}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 191, in retry_target
return target()
File "/workspace/speech/snippets/transcribe_gcs_v2_test.py", line 38, in test_transcribe_gcs_v2
response = transcribe_gcs_v2.transcribe_gcs_v2(
File "/workspace/speech/snippets/transcribe_gcs_v2.py", line 48, in transcribe_gcs_v2
operation = client.create_recognizer(request=request)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/speech_v2/services/speech/client.py", line 695, in create_recognizer
response = rpc(
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable
raise exceptions.from_grpc_error(exc) from exc
google.api_core.exceptions.ServiceUnavailable: 503 502:Bad Gateway
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 349, in retry_wrapped_func
return retry_target(
File "/workspace/speech/snippets/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/retry.py", line 207, in retry_target
raise exceptions.RetryError(
google.api_core.exceptions.RetryError: Deadline of 120.0s exceeded while calling target function, last exception: 503 502:Bad Gateway</pre></details> | priority | speech snippets transcribe gcs test test transcribe gcs failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace speech snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace speech snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unavailable details bad gateway debug error string unknown error received from peer grpc message bad gateway grpc status created time the above exception was the direct cause of the following exception traceback most recent call last file workspace speech snippets nox py lib site packages google api core retry py line in retry target return target file workspace speech snippets transcribe gcs test py line in test transcribe gcs response transcribe gcs transcribe gcs file workspace speech snippets transcribe gcs py line in transcribe gcs operation client create recognizer request request file workspace speech snippets nox py lib site packages google cloud speech services speech client py line in create recognizer response rpc file workspace speech snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace speech snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions serviceunavailable bad gateway the above exception was the direct cause of the following exception traceback most recent call last file workspace speech snippets nox py lib site packages google api core retry py line in retry wrapped func return retry target file workspace speech snippets nox py lib site packages google api core retry py line in retry target raise exceptions retryerror google api core exceptions retryerror deadline of exceeded while calling target function last exception bad gateway | 1 |
257,248 | 8,134,897,438 | IssuesEvent | 2018-08-19 21:14:51 | python/mypy | https://api.github.com/repos/python/mypy | opened | Re-work how fine grained targets are processed | needs discussion priority-1-normal refactoring topic-fine-grained-incremental | Currently, fine grained targets are processed per updated module. This can lead to files being processed multiple times (and also a bit hard to reason, but this may be subjective). I propose to reorganise them to be processed in topologically sorted order. So the algorithm would be like this:
1. Process all edited files, calculate all fired triggers, chain them to find all invalid targets, check for blocking errors.
2. Group targets per module, order them per SCC, then per same heuristics we use to order modules within SCCs in coarse grained incremental. Within module targets are ordered by line number, this is unchanged.
3. Process targets in one module from the queue, calculate updated deps and fired triggers, update invalid targets queue (maintaining the sort order), check blockers
4. Continue step 3 until no modules left in the queue at this SCC
5. Flush error messages
6. Continue steps 3-5 until no SCCs left
This way it is much less likely that we will reprocess the same module twice. This will probably give an especially significant performance gain for cold runs, where many modules are updated w.r.t. to remote cache. Also IMO this algorithm is easier to reason about (and more similar to what happens in coarse-grained mode).
This idea appeared some time ago, but was postponed. Filing an issue to not forget about this. | 1.0 | Re-work how fine grained targets are processed - Currently, fine grained targets are processed per updated module. This can lead to files being processed multiple times (and also a bit hard to reason, but this may be subjective). I propose to reorganise them to be processed in topologically sorted order. So the algorithm would be like this:
1. Process all edited files, calculate all fired triggers, chain them to find all invalid targets, check for blocking errors.
2. Group targets per module, order them per SCC, then per same heuristics we use to order modules within SCCs in coarse grained incremental. Within module targets are ordered by line number, this is unchanged.
3. Process targets in one module from the queue, calculate updated deps and fired triggers, update invalid targets queue (maintaining the sort order), check blockers
4. Continue step 3 until no modules left in the queue at this SCC
5. Flush error messages
6. Continue steps 3-5 until no SCCs left
This way it is much less likely that we will reprocess the same module twice. This will probably give an especially significant performance gain for cold runs, where many modules are updated w.r.t. to remote cache. Also IMO this algorithm is easier to reason about (and more similar to what happens in coarse-grained mode).
This idea appeared some time ago, but was postponed. Filing an issue to not forget about this. | priority | re work how fine grained targets are processed currently fine grained targets are processed per updated module this can lead to files being processed multiple times and also a bit hard to reason but this may be subjective i propose to reorganise them to be processed in topologically sorted order so the algorithm would be like this process all edited files calculate all fired triggers chain them to find all invalid targets check for blocking errors group targets per module order them per scc then per same heuristics we use to order modules within sccs in coarse grained incremental within module targets are ordered by line number this is unchanged process targets in one module from the queue calculate updated deps and fired triggers update invalid targets queue maintaining the sort order check blockers continue step until no modules left in the queue at this scc flush error messages continue steps until no sccs left this way it is much less likely that we will reprocess the same module twice this will probably give an especially significant performance gain for cold runs where many modules are updated w r t to remote cache also imo this algorithm is easier to reason about and more similar to what happens in coarse grained mode this idea appeared some time ago but was postponed filing an issue to not forget about this | 1 |
56,680 | 14,078,482,024 | IssuesEvent | 2020-11-04 13:38:02 | themagicalmammal/android_kernel_samsung_s5neolte | https://api.github.com/repos/themagicalmammal/android_kernel_samsung_s5neolte | opened | CVE-2019-15926 (High) detected in linuxv3.10 | security vulnerability | ## CVE-2019-15926 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_s5neolte/commit/f978d7dbb980bbe5267a625da958c4226e1a8ae0">f978d7dbb980bbe5267a625da958c4226e1a8ae0</a></p>
<p>Found in base branch: <b>cosmic-experimental-1.6</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.2.3. Out of bounds access exists in the functions ath6kl_wmi_pstream_timeout_event_rx and ath6kl_wmi_cac_event_rx in the file drivers/net/wireless/ath/ath6kl/wmi.c.
<p>Publish Date: 2019-09-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15926>CVE-2019-15926</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3</a></p>
<p>Release Date: 2019-09-04</p>
<p>Fix Resolution: v5.3-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-15926 (High) detected in linuxv3.10 - ## CVE-2019-15926 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_s5neolte/commit/f978d7dbb980bbe5267a625da958c4226e1a8ae0">f978d7dbb980bbe5267a625da958c4226e1a8ae0</a></p>
<p>Found in base branch: <b>cosmic-experimental-1.6</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.2.3. Out of bounds access exists in the functions ath6kl_wmi_pstream_timeout_event_rx and ath6kl_wmi_cac_event_rx in the file drivers/net/wireless/ath/ath6kl/wmi.c.
<p>Publish Date: 2019-09-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-15926>CVE-2019-15926</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3">https://cdn.kernel.org/pub/linux/kernel/v5.x/ChangeLog-5.2.3</a></p>
<p>Release Date: 2019-09-04</p>
<p>Fix Resolution: v5.3-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch cosmic experimental vulnerable source files vulnerability details an issue was discovered in the linux kernel before out of bounds access exists in the functions wmi pstream timeout event rx and wmi cac event rx in the file drivers net wireless ath wmi c publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
70,221 | 9,382,912,243 | IssuesEvent | 2019-04-05 00:32:19 | certbot/certbot | https://api.github.com/repos/certbot/certbot | closed | Clarify documentation of configuration file | area: documentation | https://certbot.eff.org/docs/using.html#configuration-file says::
```
# All flags used by the client can be configured here. Run Certbot with
# "--help" to learn more about the available options.
```
This begs the question of whether there are configuration options that are not flags to the client, and it is not answered in this section. This should be clarified, either by discussing which options are configuration file only, or changing the above to something like:
```
# The possible configuration options are exactly the same as the flags
# used by the client. Run Certbot with "--help" to learn more about the
# available options.
```
| 1.0 | Clarify documentation of configuration file - https://certbot.eff.org/docs/using.html#configuration-file says::
```
# All flags used by the client can be configured here. Run Certbot with
# "--help" to learn more about the available options.
```
This begs the question of whether there are configuration options that are not flags to the client, and it is not answered in this section. This should be clarified, either by discussing which options are configuration file only, or changing the above to something like:
```
# The possible configuration options are exactly the same as the flags
# used by the client. Run Certbot with "--help" to learn more about the
# available options.
```
| non_priority | clarify documentation of configuration file says all flags used by the client can be configured here run certbot with help to learn more about the available options this begs the question of whether there are configuration options that are not flags to the client and it is not answered in this section this should be clarified either by discussing which options are configuration file only or changing the above to something like the possible configuration options are exactly the same as the flags used by the client run certbot with help to learn more about the available options | 0 |
104,038 | 4,194,219,805 | IssuesEvent | 2016-06-25 00:18:10 | UHMDCmd/DCmd | https://api.github.com/repos/UHMDCmd/DCmd | opened | new Service-Host Assignment does not show. | bug Medium Priority | When editing an application, creating a new service host assignment does not show in the grid once created or updated. Audit log shows that something had happen, but nothing is shown for it.
On the Host page side, if the application has no service and try to make a new service host assignment, nothing happens. Only works if an existing service is there. | 1.0 | new Service-Host Assignment does not show. - When editing an application, creating a new service host assignment does not show in the grid once created or updated. Audit log shows that something had happen, but nothing is shown for it.
On the Host page side, if the application has no service and try to make a new service host assignment, nothing happens. Only works if an existing service is there. | priority | new service host assignment does not show when editing an application creating a new service host assignment does not show in the grid once created or updated audit log shows that something had happen but nothing is shown for it on the host page side if the application has no service and try to make a new service host assignment nothing happens only works if an existing service is there | 1 |
129,388 | 18,092,941,060 | IssuesEvent | 2021-09-22 05:16:59 | AlexRogalskiy/typescript-tools | https://api.github.com/repos/AlexRogalskiy/typescript-tools | closed | CVE-2021-32796 (Medium) detected in xmldom-0.6.0.tgz | security vulnerability Status: Invalid | ## CVE-2021-32796 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmldom-0.6.0.tgz</b></p></summary>
<p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz</a></p>
<p>Path to dependency file: typescript-tools/package.json</p>
<p>Path to vulnerable library: /node_modules/xmldom/package.json</p>
<p>
Dependency Hierarchy:
- :x: **xmldom-0.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/typescript-tools/commit/afc1c33cea76f8d57316e4485c82961b29f04d08">afc1c33cea76f8d57316e4485c82961b29f04d08</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
xmldom is an open source pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module. xmldom versions 0.6.0 and older do not correctly escape special characters when serializing elements removed from their ancestor. This may lead to unexpected syntactic changes during XML processing in some downstream applications. This issue has been resolved in version 0.7.0. As a workaround downstream applications can validate the input and reject the maliciously crafted documents.
<p>Publish Date: 2021-07-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32796>CVE-2021-32796</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q">https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q</a></p>
<p>Release Date: 2021-07-27</p>
<p>Fix Resolution: xmldom - 0.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-32796 (Medium) detected in xmldom-0.6.0.tgz - ## CVE-2021-32796 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmldom-0.6.0.tgz</b></p></summary>
<p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz</a></p>
<p>Path to dependency file: typescript-tools/package.json</p>
<p>Path to vulnerable library: /node_modules/xmldom/package.json</p>
<p>
Dependency Hierarchy:
- :x: **xmldom-0.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/typescript-tools/commit/afc1c33cea76f8d57316e4485c82961b29f04d08">afc1c33cea76f8d57316e4485c82961b29f04d08</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
xmldom is an open source pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module. xmldom versions 0.6.0 and older do not correctly escape special characters when serializing elements removed from their ancestor. This may lead to unexpected syntactic changes during XML processing in some downstream applications. This issue has been resolved in version 0.7.0. As a workaround downstream applications can validate the input and reject the maliciously crafted documents.
<p>Publish Date: 2021-07-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32796>CVE-2021-32796</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q">https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q</a></p>
<p>Release Date: 2021-07-27</p>
<p>Fix Resolution: xmldom - 0.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in xmldom tgz cve medium severity vulnerability vulnerable library xmldom tgz a pure javascript standard based xml dom level core domparser and xmlserializer module library home page a href path to dependency file typescript tools package json path to vulnerable library node modules xmldom package json dependency hierarchy x xmldom tgz vulnerable library found in head commit a href found in base branch master vulnerability details xmldom is an open source pure javascript standard based xml dom level core domparser and xmlserializer module xmldom versions and older do not correctly escape special characters when serializing elements removed from their ancestor this may lead to unexpected syntactic changes during xml processing in some downstream applications this issue has been resolved in version as a workaround downstream applications can validate the input and reject the maliciously crafted documents publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xmldom step up your open source security game with whitesource | 0 |
526,540 | 15,295,240,829 | IssuesEvent | 2021-02-24 04:21:34 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Missing translations for onboarding modal on first launch | OS/Desktop QA/Yes l10n onboarding priority/P2 release-notes/exclude | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Missing translations for onboarding modal on first launch
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Set locale to France (FR)
2. Install 1.21.65 x64 beta build
3. Launch browser, trigger onboarding modal that is not translated
## Actual result:
<!--Please add screenshots if needed-->

## Expected result:
Translated strings when locale is not EN
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.21.62 Chromium: 88.0.4324.182 (Official Build) beta (64-bit)
-- | --
Revision | `73ee5087001dcef33047c4ed650471b225dd8caf-refs/branch-heads/4324@{#2202}`
OS | Linux
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @rebron @mkarolin | 1.0 | Missing translations for onboarding modal on first launch - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
Missing translations for onboarding modal on first launch
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Set locale to France (FR)
2. Install 1.21.65 x64 beta build
3. Launch browser, trigger onboarding modal that is not translated
## Actual result:
<!--Please add screenshots if needed-->

## Expected result:
Translated strings when locale is not EN
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easy
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.21.62 Chromium: 88.0.4324.182 (Official Build) beta (64-bit)
-- | --
Revision | `73ee5087001dcef33047c4ed650471b225dd8caf-refs/branch-heads/4324@{#2202}`
OS | Linux
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? NA
- Can you reproduce this issue with the beta channel? Yes
- Can you reproduce this issue with the nightly channel? Yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? NA
- Does the issue resolve itself when disabling Brave Rewards? NA
- Is the issue reproducible on the latest version of Chrome? NA
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
cc: @rebron @mkarolin | priority | missing translations for onboarding modal on first launch have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description missing translations for onboarding modal on first launch steps to reproduce set locale to france fr install beta build launch browser trigger onboarding modal that is not translated actual result expected result translated strings when locale is not en reproduces how often easy brave version brave version info brave chromium official build beta bit revision refs branch heads os linux version channel information can you reproduce this issue with the current release na can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc rebron mkarolin | 1 |
305,968 | 9,379,024,072 | IssuesEvent | 2019-04-04 14:07:48 | AnalyticalGraphicsInc/cesium | https://api.github.com/repos/AnalyticalGraphicsInc/cesium | closed | Investigate image quality degradation | category - terrain and imagery priority - high type - bug | Reported [on the forum](https://groups.google.com/forum/#!topic/cesium-dev/xWcXPYkh3c8)
Check out the below image. You'll notice the Cesium version is washed out and slightly blurry compared to the same image in Bing Maps (click it for full resolution). At some point in the rendering pipeline (texture filtering, AA, lighting, etc..) Cesium processing is reducing imagery quality.
I looked at the source files in the network debugger and confirmed that both engines are using the same exact imagery and the rendered Bing map version is identical to the source (since it's just doing img elements like traditional webmaps). This means that Cesium is definitely the culprit here.

| 1.0 | Investigate image quality degradation - Reported [on the forum](https://groups.google.com/forum/#!topic/cesium-dev/xWcXPYkh3c8)
Check out the below image. You'll notice the Cesium version is washed out and slightly blurry compared to the same image in Bing Maps (click it for full resolution). At some point in the rendering pipeline (texture filtering, AA, lighting, etc..) Cesium processing is reducing imagery quality.
I looked at the source files in the network debugger and confirmed that both engines are using the same exact imagery and the rendered Bing map version is identical to the source (since it's just doing img elements like traditional webmaps). This means that Cesium is definitely the culprit here.

| priority | investigate image quality degradation reported check out the below image you ll notice the cesium version is washed out and slightly blurry compared to the same image in bing maps click it for full resolution at some point in the rendering pipeline texture filtering aa lighting etc cesium processing is reducing imagery quality i looked at the source files in the network debugger and confirmed that both engines are using the same exact imagery and the rendered bing map version is identical to the source since it s just doing img elements like traditional webmaps this means that cesium is definitely the culprit here | 1 |
2,782 | 2,643,518,606 | IssuesEvent | 2015-03-12 11:34:29 | owncloud/core | https://api.github.com/repos/owncloud/core | closed | [oc update 8.0.2] Loading gif showing forever in update view | Design | When installing an update (from 8.0.0 to 8.0.2) and experiencing an error the loading-small.gif is displayed forever below the "3. Replace" section even though the update process terminated.
Also, should this really be a .gif animation? Suggesting a replacement with a more modern looking css or js loading symbol. (http://cssload.net/ and http://tobiasahlin.com/spinkit/ etc.) | 1.0 | [oc update 8.0.2] Loading gif showing forever in update view - When installing an update (from 8.0.0 to 8.0.2) and experiencing an error the loading-small.gif is displayed forever below the "3. Replace" section even though the update process terminated.
Also, should this really be a .gif animation? Suggesting a replacement with a more modern looking css or js loading symbol. (http://cssload.net/ and http://tobiasahlin.com/spinkit/ etc.) | non_priority | loading gif showing forever in update view when installing an update from to and experiencing an error the loading small gif is displayed forever below the replace section even though the update process terminated also should this really be a gif animation suggesting a replacement with a more modern looking css or js loading symbol and etc | 0 |
303,876 | 26,235,122,000 | IssuesEvent | 2023-01-05 06:14:08 | kubernetes-sigs/kwok | https://api.github.com/repos/kubernetes-sigs/kwok | closed | Failing `kwokctl snapshot` with runtime `nerdctl` | kind/failing-test | ### Which jobs are failing?
https://github.com/kubernetes-sigs/kwok/blob/88e598265a3c899bfb010088d8a6d85ebf429a57/.github/workflows/test.yaml#L137-L141
### Which tests are failing?
https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840
``` console
================================================================================
Testing kwokctl/kwokctl_nerdctl_snapshot...
github.com/docker/buildx 0.[9](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:10).1+azure-2 ed00243a0ce2a0aee75311b06e32d33b44729689
mkdir: cannot create directory ‘tmp’: File exists
unpacking docker.io/local/kwok:test (sha256:5019cea8f3dd71453e676622dd7825d12c2e7ae4272560aa295d7787dfbf0245)...
Loaded image: docker.io/local/kwok:testTest snapshot on nerdctl for 1.25.3 1.24.7 1.23.13 1.22.15 1.21.14 1.20.15
------------------------------
Testing snapshot on nerdctl for 1.25.3
{"time":"2022-11-28T14:50:53.869177459Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
{"time":"2022-11-28T14:50:54.550340799Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
time="2022-11-28T14:50:54Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-25-3"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.25.3"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.25.3"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.25.3"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-etcd"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver"
You can now use your cluster with:
{"time":"2022-11-28T14:50:59.168789831Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-25-3"}
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-25-3
Thanks for using kwok!
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
Download https://github.com/etcd-io/etcd/releases/download/v3.5.6/etcd-v3.5.6-linux-amd64.tar.gz
{"time":"2022-11-28T14:51:37.882622576Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-25-3/etcd-data kwok-snapshot-cluster-nerdctl-1-25-3-etcd:/: exit status 1\ntime=\"2022-11-28T14:51:37Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/14518/ns/net\\\": no such file or directory\" id=2af2754d12fd763d179bf1e3225b62a7ea3d5a86a37eabc1bb29944a04c8ce85\ntime=\"2022-11-28T14:51:37Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
Error: Empty snapshot restore failed
Expected:
Actual: NAMESPACE NAME
default fake-pod-78f5b8f676-gj6r9
NAME
fake-node
{"time":"2022-11-28T14:52:34.878035422Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
time="2022-11-28T14:52:34Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-etcd"
time="2022-11-28T14:52:35Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-25-3"
{"time":"2022-11-28T14:52:35.731964445Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
{"time":"2022-11-28T14:52:35.882827316Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-25-3"}
------------------------------
Testing snapshot on nerdctl for 1.24.7
{"time":"2022-11-28T14:52:35.899447477Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
{"time":"2022-11-28T14:52:37.003300129Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
time="2022-11-28T14:52:37Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-24-7"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.24.7"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.24.7"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.24.7"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-etcd"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver"
You can now use your cluster with:
{"time":"2022-11-28T14:52:41.660050875Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-24-7"}
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-24-7
Thanks for using kwok!
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
{"time":"2022-11-28T14:53:19.502931142Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-24-7/etcd-data kwok-snapshot-cluster-nerdctl-1-24-7-etcd:/: exit status 1\ntime=\"2022-11-28T14:53:19Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/18073/ns/net\\\": no such file or directory\" id=4fb042bd8586bcd039dac994d5262e857ed2631ddcc8509367aa5ac50201a179\ntime=\"2022-11-28T14:53:19Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
Error: Empty snapshot restore failed
Expected:
Actual: NAMESPACE NAME
default fake-pod-6f5fffcbc-hff5p
NAME
fake-node
{"time":"2022-11-28T14:54:16.753867524Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller"
time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler"
time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager"
time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver"
time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-etcd"
time="2022-11-28T14:54:17Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-24-7"
{"time":"2022-11-28T14:54:17.535599245Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
{"time":"2022-11-28T14:54:17.677782486Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-24-7"}
------------------------------
Testing snapshot on nerdctl for 1.23.13
{"time":"2022-11-28T14:54:17.692811822Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
{"time":"2022-11-28T14:54:18.561574115Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
time="2022-11-28T14:54:18Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-23-13"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.23.13"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.23.13"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.23.13"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-etcd"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller"
You can now use your cluster with:
{"time":"2022-11-28T14:54:22.92009441Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-23-13"}
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-23-13
Thanks for using kwok!
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
{"time":"2022-11-28T14:55:00.555011[10](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:11)1Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-23-13/etcd-data kwok-snapshot-cluster-nerdctl-1-23-13-etcd:/: exit status 1\ntime=\"2022-[11](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:12)-28T14:55:00Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/21546/ns/net\\\": no such file or directory\" id=37a024f06534b4c24888948cb02592bfbf348a86def52b0ef420c9b2278c5e95\ntime=\"2022-11-28T14:55:00Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
Error: Empty snapshot restore failed
Expected:
Actual: NAMESPACE NAME
default fake-pod-6bf4bdd9cc-5jqlf
NAME
fake-node
{"time":"2022-11-28T14:55:57.588653024Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller"
time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler"
time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager"
time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver"
time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-etcd"
time="2022-11-28T14:55:58Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-23-13"
{"time":"2022-11-28T14:55:58.389470291Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
{"time":"2022-11-28T14:55:58.526502308Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-23-13"}
------------------------------
Testing snapshot on nerdctl for 1.22.15
{"time":"2022-11-28T14:55:58.545860907Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"}
{"time":"2022-11-28T14:55:59.42[12](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:13)21125Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"}
time="2022-11-28T[14](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:15):55:59Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-22-[15](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:16)"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.22.15"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.22.15"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.22.15"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-controller-manager"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kwok-controller"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-scheduler"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-etcd"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-apiserver"
You can now use your cluster with:
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-22-15
Thanks for using kwok!
{"time":"2022-11-28T14:56:03.6658[16](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:17)766Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-22-15"}
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
{"time":"2022-11-28T14:56:41.2365[17](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:18)354Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.[19](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:20).3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-22-15/etcd-data kwok-snapshot-cluster-nerdctl-1-22-15-etcd:/: exit status 1\ntime=\"[20](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:21)[22](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:23)-11-28T14:56:40Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/25054/ns/net\\\": no such file or directory\" id=8ec[23](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:24)8fa41f2f3a7039bf6e120f2de73ad776cc8c775db43b3301157aa430f[26](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:27)\ntime=\"2022-11-[28](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:29)T14:56:[40](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:41)Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
```
### Since when has it been failing?
Always
### Reason for failure (if possible)
_No response_
### Anything else we need to know?
_No response_ | 1.0 | Failing `kwokctl snapshot` with runtime `nerdctl` - ### Which jobs are failing?
https://github.com/kubernetes-sigs/kwok/blob/88e598265a3c899bfb010088d8a6d85ebf429a57/.github/workflows/test.yaml#L137-L141
### Which tests are failing?
https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840
``` console
================================================================================
Testing kwokctl/kwokctl_nerdctl_snapshot...
github.com/docker/buildx 0.[9](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:10).1+azure-2 ed00243a0ce2a0aee75311b06e32d33b44729689
mkdir: cannot create directory ‘tmp’: File exists
unpacking docker.io/local/kwok:test (sha256:5019cea8f3dd71453e676622dd7825d12c2e7ae4272560aa295d7787dfbf0245)...
Loaded image: docker.io/local/kwok:testTest snapshot on nerdctl for 1.25.3 1.24.7 1.23.13 1.22.15 1.21.14 1.20.15
------------------------------
Testing snapshot on nerdctl for 1.25.3
{"time":"2022-11-28T14:50:53.869177459Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
{"time":"2022-11-28T14:50:54.550340799Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
time="2022-11-28T14:50:54Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-25-3"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.25.3"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.25.3"
time="2022-11-28T14:50:54Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.25.3"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-etcd"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager"
time="2022-11-28T14:50:54Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver"
You can now use your cluster with:
{"time":"2022-11-28T14:50:59.168789831Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-25-3"}
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-25-3
Thanks for using kwok!
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
Download https://github.com/etcd-io/etcd/releases/download/v3.5.6/etcd-v3.5.6-linux-amd64.tar.gz
{"time":"2022-11-28T14:51:37.882622576Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-25-3/etcd-data kwok-snapshot-cluster-nerdctl-1-25-3-etcd:/: exit status 1\ntime=\"2022-11-28T14:51:37Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/14518/ns/net\\\": no such file or directory\" id=2af2754d12fd763d179bf1e3225b62a7ea3d5a86a37eabc1bb29944a04c8ce85\ntime=\"2022-11-28T14:51:37Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
Error: Empty snapshot restore failed
Expected:
Actual: NAMESPACE NAME
default fake-pod-78f5b8f676-gj6r9
NAME
fake-node
{"time":"2022-11-28T14:52:34.878035422Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
time="2022-11-28T14:52:34Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-scheduler"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-controller-manager"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kwok-controller"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-kube-apiserver"
time="2022-11-28T14:52:35Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-25-3-etcd"
time="2022-11-28T14:52:35Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-25-3"
{"time":"2022-11-28T14:52:35.731964445Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-25-3"}
{"time":"2022-11-28T14:52:35.882827316Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-25-3"}
------------------------------
Testing snapshot on nerdctl for 1.24.7
{"time":"2022-11-28T14:52:35.899447477Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
{"time":"2022-11-28T14:52:37.003300129Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
time="2022-11-28T14:52:37Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-24-7"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.24.7"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.24.7"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.24.7"
time="2022-11-28T14:52:37Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-etcd"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller"
time="2022-11-28T14:52:37Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver"
You can now use your cluster with:
{"time":"2022-11-28T14:52:41.660050875Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-24-7"}
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-24-7
Thanks for using kwok!
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
{"time":"2022-11-28T14:53:19.502931142Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-24-7/etcd-data kwok-snapshot-cluster-nerdctl-1-24-7-etcd:/: exit status 1\ntime=\"2022-11-28T14:53:19Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/18073/ns/net\\\": no such file or directory\" id=4fb042bd8586bcd039dac994d5262e857ed2631ddcc8509367aa5ac50201a179\ntime=\"2022-11-28T14:53:19Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
Error: Empty snapshot restore failed
Expected:
Actual: NAMESPACE NAME
default fake-pod-6f5fffcbc-hff5p
NAME
fake-node
{"time":"2022-11-28T14:54:16.753867524Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kwok-controller"
time="2022-11-28T14:54:16Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-scheduler"
time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-controller-manager"
time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-kube-apiserver"
time="2022-11-28T14:54:17Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-24-7-etcd"
time="2022-11-28T14:54:17Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-24-7"
{"time":"2022-11-28T14:54:17.535599245Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-24-7"}
{"time":"2022-11-28T14:54:17.677782486Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-24-7"}
------------------------------
Testing snapshot on nerdctl for 1.23.13
{"time":"2022-11-28T14:54:17.692811822Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
{"time":"2022-11-28T14:54:18.561574115Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
time="2022-11-28T14:54:18Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-23-13"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.23.13"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.23.13"
time="2022-11-28T14:54:18Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.23.13"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-etcd"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver"
time="2022-11-28T14:54:18Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller"
You can now use your cluster with:
{"time":"2022-11-28T14:54:22.92009441Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-23-13"}
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-23-13
Thanks for using kwok!
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
{"time":"2022-11-28T14:55:00.555011[10](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:11)1Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.19.3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-23-13/etcd-data kwok-snapshot-cluster-nerdctl-1-23-13-etcd:/: exit status 1\ntime=\"2022-[11](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:12)-28T14:55:00Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/21546/ns/net\\\": no such file or directory\" id=37a024f06534b4c24888948cb02592bfbf348a86def52b0ef420c9b2278c5e95\ntime=\"2022-11-28T14:55:00Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
Error: Empty snapshot restore failed
Expected:
Actual: NAMESPACE NAME
default fake-pod-6bf4bdd9cc-5jqlf
NAME
fake-node
{"time":"2022-11-28T14:55:57.588653024Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Stopping cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kwok-controller"
time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-scheduler"
time="2022-11-28T14:55:57Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-controller-manager"
time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-kube-apiserver"
time="2022-11-28T14:55:58Z" level=info msg="Removing container kwok-snapshot-cluster-nerdctl-1-23-13-etcd"
time="2022-11-28T14:55:58Z" level=info msg="Removing network kwok-snapshot-cluster-nerdctl-1-23-13"
{"time":"2022-11-28T14:55:58.389470291Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Deleting cluster","cluster":"snapshot-cluster-nerdctl-1-23-13"}
{"time":"2022-11-28T14:55:58.526502308Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/delete/cluster/cluster.go:45","msg":"Cluster deleted","cluster":"snapshot-cluster-nerdctl-1-23-13"}
------------------------------
Testing snapshot on nerdctl for 1.22.15
{"time":"2022-11-28T14:55:58.545860907Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Creating cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"}
{"time":"2022-11-28T14:55:59.42[12](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:13)21125Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Starting cluster","cluster":"snapshot-cluster-nerdctl-1-22-15"}
time="2022-11-28T[14](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:15):55:59Z" level=info msg="Creating network kwok-snapshot-cluster-nerdctl-1-22-[15](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:16)"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_apiserver: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_controller_manager: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kube_scheduler: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: service kwok_controller: [Links]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=warning msg="Ignoring: volume: Bind: [CreateHostPath]"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/etcd:3.5.6-0"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-apiserver:v1.22.15"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-controller-manager:v1.22.15"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image registry.k8s.io/kube-scheduler:v1.22.15"
time="2022-11-28T14:55:59Z" level=info msg="Ensuring image local/kwok:test"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-controller-manager"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kwok-controller"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-scheduler"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-etcd"
time="2022-11-28T14:55:59Z" level=info msg="Creating container kwok-snapshot-cluster-nerdctl-1-22-15-kube-apiserver"
You can now use your cluster with:
kubectl config use-context kwok-snapshot-cluster-nerdctl-1-22-15
Thanks for using kwok!
{"time":"2022-11-28T14:56:03.6658[16](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:17)766Z","level":"INFO","source":"/home/runner/work/kwok/kwok/pkg/kwokctl/cmd/create/cluster/cluster.go:75","msg":"Cluster is ready","cluster":"snapshot-cluster-nerdctl-1-22-15"}
No resources found
No resources found
deployment.apps/fake-pod created
node/fake-node created
{"time":"2022-11-28T14:56:41.2365[17](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:18)354Z","level":"ERROR","source":"/opt/hostedtoolcache/go/1.[19](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:20).3/x64/src/runtime/proc.go:250","msg":"Execute exit","err":"nerdctl cp /home/runner/.kwok/clusters/snapshot-cluster-nerdctl-1-22-15/etcd-data kwok-snapshot-cluster-nerdctl-1-22-15-etcd:/: exit status 1\ntime=\"[20](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:21)[22](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:23)-11-28T14:56:40Z\" level=warning msg=\"failed to inspect NetNS\" error=\"failed to Statfs \\\"/proc/25054/ns/net\\\": no such file or directory\" id=8ec[23](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:24)8fa41f2f3a7039bf6e120f2de73ad776cc8c775db43b3301157aa430f[26](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:27)\ntime=\"2022-11-[28](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:29)T14:56:[40](https://github.com/kubernetes-sigs/kwok/actions/runs/3565851048/jobs/5991537840#step:13:41)Z\" level=fatal msg=\"expected container status running, got stopped\"\n"}
```
### Since when has it been failing?
Always
### Reason for failure (if possible)
_No response_
### Anything else we need to know?
_No response_ | non_priority | failing kwokctl snapshot with runtime nerdctl which jobs are failing which tests are failing console testing kwokctl kwokctl nerdctl snapshot github com docker buildx mkdir cannot create directory ‘tmp’ file exists unpacking docker io local kwok test loaded image docker io local kwok testtest snapshot on nerdctl for testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg starting cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image local kwok test time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg creating container kwok snapshot cluster nerdctl kwok controller time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl kube apiserver you can now use your cluster with time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg cluster is ready cluster snapshot cluster nerdctl kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok no resources found no resources found deployment apps fake pod created node fake node created download time level error source opt hostedtoolcache go src runtime proc go msg execute exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id ntime level fatal msg expected container status running got stopped n error empty snapshot restore failed expected actual namespace name default fake pod name fake node time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg stopping cluster cluster snapshot cluster nerdctl time level info msg removing container kwok snapshot cluster nerdctl kube scheduler time level info msg removing container kwok snapshot cluster nerdctl kube controller manager time level info msg removing container kwok snapshot cluster nerdctl kwok controller time level info msg removing container kwok snapshot cluster nerdctl kube apiserver time level info msg removing container kwok snapshot cluster nerdctl etcd time level info msg removing network kwok snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg deleting cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg cluster deleted cluster snapshot cluster nerdctl testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg starting cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg ensuring image local kwok test time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kwok controller time level info msg creating container kwok snapshot cluster nerdctl kube apiserver you can now use your cluster with time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg cluster is ready cluster snapshot cluster nerdctl kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok no resources found no resources found deployment apps fake pod created node fake node created time level error source opt hostedtoolcache go src runtime proc go msg execute exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id ntime level fatal msg expected container status running got stopped n error empty snapshot restore failed expected actual namespace name default fake pod name fake node time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg stopping cluster cluster snapshot cluster nerdctl time level info msg removing container kwok snapshot cluster nerdctl kwok controller time level info msg removing container kwok snapshot cluster nerdctl kube scheduler time level info msg removing container kwok snapshot cluster nerdctl kube controller manager time level info msg removing container kwok snapshot cluster nerdctl kube apiserver time level info msg removing container kwok snapshot cluster nerdctl etcd time level info msg removing network kwok snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg deleting cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg cluster deleted cluster snapshot cluster nerdctl testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg starting cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image local kwok test time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl kube apiserver time level info msg creating container kwok snapshot cluster nerdctl kwok controller you can now use your cluster with time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg cluster is ready cluster snapshot cluster nerdctl kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok no resources found no resources found deployment apps fake pod created node fake node created time exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id ntime level fatal msg expected container status running got stopped n error empty snapshot restore failed expected actual namespace name default fake pod name fake node time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg stopping cluster cluster snapshot cluster nerdctl time level info msg removing container kwok snapshot cluster nerdctl kwok controller time level info msg removing container kwok snapshot cluster nerdctl kube scheduler time level info msg removing container kwok snapshot cluster nerdctl kube controller manager time level info msg removing container kwok snapshot cluster nerdctl kube apiserver time level info msg removing container kwok snapshot cluster nerdctl etcd time level info msg removing network kwok snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg deleting cluster cluster snapshot cluster nerdctl time level info source home runner work kwok kwok pkg kwokctl cmd delete cluster cluster go msg cluster deleted cluster snapshot cluster nerdctl testing snapshot on nerdctl for time level info source home runner work kwok kwok pkg kwokctl cmd create cluster cluster go msg creating cluster cluster snapshot cluster nerdctl time cluster cluster snapshot cluster nerdctl time level info msg creating network kwok snapshot cluster nerdctl time level warning msg ignoring service kube apiserver time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube controller manager time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kube scheduler time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring service kwok controller time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level warning msg ignoring volume bind time level info msg ensuring image registry io etcd time level info msg ensuring image registry io kube apiserver time level info msg ensuring image registry io kube controller manager time level info msg ensuring image registry io kube scheduler time level info msg ensuring image local kwok test time level info msg creating container kwok snapshot cluster nerdctl kube controller manager time level info msg creating container kwok snapshot cluster nerdctl kwok controller time level info msg creating container kwok snapshot cluster nerdctl kube scheduler time level info msg creating container kwok snapshot cluster nerdctl etcd time level info msg creating container kwok snapshot cluster nerdctl kube apiserver you can now use your cluster with kubectl config use context kwok snapshot cluster nerdctl thanks for using kwok time is ready cluster snapshot cluster nerdctl no resources found no resources found deployment apps fake pod created node fake node created time exit err nerdctl cp home runner kwok clusters snapshot cluster nerdctl etcd data kwok snapshot cluster nerdctl etcd exit status ntime level warning msg failed to inspect netns error failed to statfs proc ns net no such file or directory id level fatal msg expected container status running got stopped n since when has it been failing always reason for failure if possible no response anything else we need to know no response | 0 |
163,259 | 6,194,884,665 | IssuesEvent | 2017-07-05 11:04:46 | zero-os/0-orchestrator | https://api.github.com/repos/zero-os/0-orchestrator | closed | Rename rocksdb-version parameter to storage-engine-version | priority_major state_verification type_bug | To specify the ardb flist version we are now using `rocksdb-version`, which is a bad name because it points to the storage engine used by ardb to store the key values, while it is actually a pointer to the flist we need to start for bringing up the key value stores.
`storage-engine-version` is a better name because it points more to what the flist actually delivers.
> **Important** Also adapt the documentation (setup.md) | 1.0 | Rename rocksdb-version parameter to storage-engine-version - To specify the ardb flist version we are now using `rocksdb-version`, which is a bad name because it points to the storage engine used by ardb to store the key values, while it is actually a pointer to the flist we need to start for bringing up the key value stores.
`storage-engine-version` is a better name because it points more to what the flist actually delivers.
> **Important** Also adapt the documentation (setup.md) | priority | rename rocksdb version parameter to storage engine version to specify the ardb flist version we are now using rocksdb version which is a bad name because it points to the storage engine used by ardb to store the key values while it is actually a pointer to the flist we need to start for bringing up the key value stores storage engine version is a better name because it points more to what the flist actually delivers important also adapt the documentation setup md | 1 |
369,184 | 10,889,408,872 | IssuesEvent | 2019-11-18 18:10:50 | SpaceStation0/star-engine | https://api.github.com/repos/SpaceStation0/star-engine | opened | Implementation of ECS + event system | A-ecs E-medium P-priority-high T-feature | The specific details will be fleshed out once the specification has been written, but for now, this is known for sure:
- The way the engine will process things will be a hybrid of a pure ECS and an event system
- Once the ECS finishes processing, an event queue variable will be extracted and sent to the event system
- The event system has a set of notifiers. Events will be send to notifiers, which can also emit events.
- Events emitted by notifiers will be collected, and sent out to their respective notifiers
- The process ends once there are no more events to process | 1.0 | Implementation of ECS + event system - The specific details will be fleshed out once the specification has been written, but for now, this is known for sure:
- The way the engine will process things will be a hybrid of a pure ECS and an event system
- Once the ECS finishes processing, an event queue variable will be extracted and sent to the event system
- The event system has a set of notifiers. Events will be send to notifiers, which can also emit events.
- Events emitted by notifiers will be collected, and sent out to their respective notifiers
- The process ends once there are no more events to process | priority | implementation of ecs event system the specific details will be fleshed out once the specification has been written but for now this is known for sure the way the engine will process things will be a hybrid of a pure ecs and an event system once the ecs finishes processing an event queue variable will be extracted and sent to the event system the event system has a set of notifiers events will be send to notifiers which can also emit events events emitted by notifiers will be collected and sent out to their respective notifiers the process ends once there are no more events to process | 1 |
341,394 | 30,584,770,052 | IssuesEvent | 2023-07-21 12:34:10 | proarc/proarc-client | https://api.github.com/repos/proarc/proarc-client | closed | Chyba založení STT vícedílné monografie | 1 chyba 6 k testování 7 návrh na zavření 6c otestováno: KNAV 6f otestováno: MKP priorita 1 | Při pokusu o založení STT vícedílné monografie se záznam z katalogu nepřevzal do formuláře. Testovací záznam: sysno = 000737900, katalog knav monografie.

| 3.0 | Chyba založení STT vícedílné monografie - Při pokusu o založení STT vícedílné monografie se záznam z katalogu nepřevzal do formuláře. Testovací záznam: sysno = 000737900, katalog knav monografie.

| non_priority | chyba založení stt vícedílné monografie při pokusu o založení stt vícedílné monografie se záznam z katalogu nepřevzal do formuláře testovací záznam sysno katalog knav monografie | 0 |
337,095 | 10,210,335,695 | IssuesEvent | 2019-08-14 14:36:10 | Signbank/Global-signbank | https://api.github.com/repos/Signbank/Global-signbank | opened | Order gloss typeahead results by length | ASL blocking bug high priority | When creating a new gloss, Julie wants to search for and select the lemma 'WE'. However, it is not in the list of 5 suggestions, because longer words like 'jewelry' and 'next-week' apparently go first, and the actual thing she is looking does not 'make the list'.
A solution could be to simply sort the list of matches by length; for long words you can simply continue typing if your match is not in the list, while this is not possible for short words like WE. | 1.0 | Order gloss typeahead results by length - When creating a new gloss, Julie wants to search for and select the lemma 'WE'. However, it is not in the list of 5 suggestions, because longer words like 'jewelry' and 'next-week' apparently go first, and the actual thing she is looking does not 'make the list'.
A solution could be to simply sort the list of matches by length; for long words you can simply continue typing if your match is not in the list, while this is not possible for short words like WE. | priority | order gloss typeahead results by length when creating a new gloss julie wants to search for and select the lemma we however it is not in the list of suggestions because longer words like jewelry and next week apparently go first and the actual thing she is looking does not make the list a solution could be to simply sort the list of matches by length for long words you can simply continue typing if your match is not in the list while this is not possible for short words like we | 1 |
415,918 | 28,058,792,075 | IssuesEvent | 2023-03-29 11:12:01 | CarsOk/castillo_del_rey | https://api.github.com/repos/CarsOk/castillo_del_rey | opened | Corregir el diagrama relacional | documentation | # HISTORIA DE USUARIO
## Yo como
Instructor SENA
## Quiero
Corregir el diagrama relacional
## Para
Tener la información correcta y no tener contratiempos al momento del desarrollo del software
# Criterios de aceptación
- [ ] Que tengan las relaciones
- [ ] Las llaves primarias y foráneas | 1.0 | Corregir el diagrama relacional - # HISTORIA DE USUARIO
## Yo como
Instructor SENA
## Quiero
Corregir el diagrama relacional
## Para
Tener la información correcta y no tener contratiempos al momento del desarrollo del software
# Criterios de aceptación
- [ ] Que tengan las relaciones
- [ ] Las llaves primarias y foráneas | non_priority | corregir el diagrama relacional historia de usuario yo como instructor sena quiero corregir el diagrama relacional para tener la información correcta y no tener contratiempos al momento del desarrollo del software criterios de aceptación que tengan las relaciones las llaves primarias y foráneas | 0 |
273,296 | 20,780,547,081 | IssuesEvent | 2022-03-16 14:25:15 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | opened | [Docs] Data Loading Page | area/documentation | ### Description
YugabyteDB data loading experience is completely different from the experience you have with single-server relational databases. For instance, application developers who use Postgres, MySQL or SQL Server still load data with multiple INSERTs which is considered an anti-pattern in YugabyteDB. Developers do this with Postgres and MySQL simply because the loading time is fast and many sample datasets that you download from the Internet use multiple INSERTs instead of COPY. Such developers will be puzzled to see how the loading time with multiple INSERTS takes 2x, 3x, 10x longer with YugabyteDB.
We need to educate and guide developers on how to approach the initial data loading in YugabyteDB. This is the step where the developer experience is poor.
**Data Loading page**
Create a separate **Data Loading** page. Place it right after or near the **Installation** and **Getting Started** sections. The logic here is that once the developers deploys YugabyteDB and succeeds with a getting-started-level app/tutorial she will move forward with some sample data that represents a subset of the data of a company's app or a data set downloaded from the Internet. Thus, the **Data Loading** page has to be visible in the navigation tree.
**Data Loading page content**
The content needs to explain various data loading optimization techniques and explain why some approaches don't work as expected (for instance the multiple INSERTs case).
Some techniques to mention:
1. Loading with INSERTS
* Replace multiple INSERTs to a single or a few INSERT statements
* Execute the INSERT statement(s) in a single transaction - use `BEGIN` insert `COMMIT` structure
* Disable triggers (but don't forget to turn them back) - `ALTER TABLE <table_name> disable TRIGGER ALL`
* etc. more techniques are discused here: https://docs.google.com/document/d/1jCLiHDEHiYpgVObILDC_2Ormr-Kx36YhkqHXUCVGO1Q/edit#heading=h.28rbkno9vzy9
It's important to explain why each step is necessary in YugabyteDB, how we're different and why this is OK. Remember that the performance is poor even with a single-node YugbyteDB cluster running locally if you compare to a local Postgres instance.
2. Loading with the COPY command
Suggest this as an alternate (preffered?) option to the #1 above. Provide steps and instructions. You can use Franck's blog for reference: https://dev.to/yugabyte/copy-progression-in-yugabytedb-4ghb
**YugabyteDB on-prem vs. Yugabyte Cloud - fighting networking latency**
We can deal with a developer who uses Yugabyte DB (or Platform) in her own environment or a developer who has started with Yugabyte Cloud.
For the on-prem scenarious, suggest loading the data from a location close to the YugabyteDB deployment.
For Yugabyte Cloud, well, it's more complicated because the developer needs to create an instance in a cloud region, where the database is running, and from that instance do the loading. For now, at least, we can provide some basic steps.
| 1.0 | [Docs] Data Loading Page - ### Description
YugabyteDB data loading experience is completely different from the experience you have with single-server relational databases. For instance, application developers who use Postgres, MySQL or SQL Server still load data with multiple INSERTs which is considered an anti-pattern in YugabyteDB. Developers do this with Postgres and MySQL simply because the loading time is fast and many sample datasets that you download from the Internet use multiple INSERTs instead of COPY. Such developers will be puzzled to see how the loading time with multiple INSERTS takes 2x, 3x, 10x longer with YugabyteDB.
We need to educate and guide developers on how to approach the initial data loading in YugabyteDB. This is the step where the developer experience is poor.
**Data Loading page**
Create a separate **Data Loading** page. Place it right after or near the **Installation** and **Getting Started** sections. The logic here is that once the developers deploys YugabyteDB and succeeds with a getting-started-level app/tutorial she will move forward with some sample data that represents a subset of the data of a company's app or a data set downloaded from the Internet. Thus, the **Data Loading** page has to be visible in the navigation tree.
**Data Loading page content**
The content needs to explain various data loading optimization techniques and explain why some approaches don't work as expected (for instance the multiple INSERTs case).
Some techniques to mention:
1. Loading with INSERTS
* Replace multiple INSERTs to a single or a few INSERT statements
* Execute the INSERT statement(s) in a single transaction - use `BEGIN` insert `COMMIT` structure
* Disable triggers (but don't forget to turn them back) - `ALTER TABLE <table_name> disable TRIGGER ALL`
* etc. more techniques are discused here: https://docs.google.com/document/d/1jCLiHDEHiYpgVObILDC_2Ormr-Kx36YhkqHXUCVGO1Q/edit#heading=h.28rbkno9vzy9
It's important to explain why each step is necessary in YugabyteDB, how we're different and why this is OK. Remember that the performance is poor even with a single-node YugbyteDB cluster running locally if you compare to a local Postgres instance.
2. Loading with the COPY command
Suggest this as an alternate (preffered?) option to the #1 above. Provide steps and instructions. You can use Franck's blog for reference: https://dev.to/yugabyte/copy-progression-in-yugabytedb-4ghb
**YugabyteDB on-prem vs. Yugabyte Cloud - fighting networking latency**
We can deal with a developer who uses Yugabyte DB (or Platform) in her own environment or a developer who has started with Yugabyte Cloud.
For the on-prem scenarious, suggest loading the data from a location close to the YugabyteDB deployment.
For Yugabyte Cloud, well, it's more complicated because the developer needs to create an instance in a cloud region, where the database is running, and from that instance do the loading. For now, at least, we can provide some basic steps.
| non_priority | data loading page description yugabytedb data loading experience is completely different from the experience you have with single server relational databases for instance application developers who use postgres mysql or sql server still load data with multiple inserts which is considered an anti pattern in yugabytedb developers do this with postgres and mysql simply because the loading time is fast and many sample datasets that you download from the internet use multiple inserts instead of copy such developers will be puzzled to see how the loading time with multiple inserts takes longer with yugabytedb we need to educate and guide developers on how to approach the initial data loading in yugabytedb this is the step where the developer experience is poor data loading page create a separate data loading page place it right after or near the installation and getting started sections the logic here is that once the developers deploys yugabytedb and succeeds with a getting started level app tutorial she will move forward with some sample data that represents a subset of the data of a company s app or a data set downloaded from the internet thus the data loading page has to be visible in the navigation tree data loading page content the content needs to explain various data loading optimization techniques and explain why some approaches don t work as expected for instance the multiple inserts case some techniques to mention loading with inserts replace multiple inserts to a single or a few insert statements execute the insert statement s in a single transaction use begin insert commit structure disable triggers but don t forget to turn them back alter table disable trigger all etc more techniques are discused here it s important to explain why each step is necessary in yugabytedb how we re different and why this is ok remember that the performance is poor even with a single node yugbytedb cluster running locally if you compare to a local postgres instance loading with the copy command suggest this as an alternate preffered option to the above provide steps and instructions you can use franck s blog for reference yugabytedb on prem vs yugabyte cloud fighting networking latency we can deal with a developer who uses yugabyte db or platform in her own environment or a developer who has started with yugabyte cloud for the on prem scenarious suggest loading the data from a location close to the yugabytedb deployment for yugabyte cloud well it s more complicated because the developer needs to create an instance in a cloud region where the database is running and from that instance do the loading for now at least we can provide some basic steps | 0 |
88,300 | 25,365,034,562 | IssuesEvent | 2022-11-21 05:10:07 | wixtoolset/issues | https://api.github.com/repos/wixtoolset/issues | closed | Sign the vsix packages | enhancement build process | If this issue is a feature request:
* Describe the scenario and benefits that the feature supports.
> the user feals more comfortable that he selected the right extension from the marketplace
* Describe how you're accomplishing the feature today (if possible).
> at the moment you see that it is not signed.
> 
* Describe what you'd like the new feature to do.
> show that the extension that will be installed belongs to the wix toolset (.NET Foundation)
| 1.0 | Sign the vsix packages - If this issue is a feature request:
* Describe the scenario and benefits that the feature supports.
> the user feals more comfortable that he selected the right extension from the marketplace
* Describe how you're accomplishing the feature today (if possible).
> at the moment you see that it is not signed.
> 
* Describe what you'd like the new feature to do.
> show that the extension that will be installed belongs to the wix toolset (.NET Foundation)
| non_priority | sign the vsix packages if this issue is a feature request describe the scenario and benefits that the feature supports the user feals more comfortable that he selected the right extension from the marketplace describe how you re accomplishing the feature today if possible at the moment you see that it is not signed describe what you d like the new feature to do show that the extension that will be installed belongs to the wix toolset net foundation | 0 |
41,198 | 12,831,755,481 | IssuesEvent | 2020-07-07 06:14:06 | rvvergara/todolist-api-igaku | https://api.github.com/repos/rvvergara/todolist-api-igaku | closed | CVE-2019-2391 (Medium) detected in bson-1.1.3.tgz | security vulnerability | ## CVE-2019-2391 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bson-1.1.3.tgz</b></p></summary>
<p>A bson parser for node.js and the browser</p>
<p>Library home page: <a href="https://registry.npmjs.org/bson/-/bson-1.1.3.tgz">https://registry.npmjs.org/bson/-/bson-1.1.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/todolist-api-igaku/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/todolist-api-igaku/node_modules/bson/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-5.7.12.tgz (Root Library)
- :x: **bson-1.1.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/todolist-api-igaku/commit/e74ac424f4412547afcf733f031f27227a0f28e9">e74ac424f4412547afcf733f031f27227a0f28e9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Incorrect parsing of certain JSON input may result in js-bson not correctly serializing BSON. This may cause unexpected application behaviour including data disclosure.
<p>Publish Date: 2020-03-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2391>CVE-2019-2391</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/mongodb/js-bson/pull/336/commits">https://github.com/mongodb/js-bson/pull/336/commits</a></p>
<p>Release Date: 2020-03-31</p>
<p>Fix Resolution: v1.1.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-2391 (Medium) detected in bson-1.1.3.tgz - ## CVE-2019-2391 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bson-1.1.3.tgz</b></p></summary>
<p>A bson parser for node.js and the browser</p>
<p>Library home page: <a href="https://registry.npmjs.org/bson/-/bson-1.1.3.tgz">https://registry.npmjs.org/bson/-/bson-1.1.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/todolist-api-igaku/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/todolist-api-igaku/node_modules/bson/package.json</p>
<p>
Dependency Hierarchy:
- mongoose-5.7.12.tgz (Root Library)
- :x: **bson-1.1.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/todolist-api-igaku/commit/e74ac424f4412547afcf733f031f27227a0f28e9">e74ac424f4412547afcf733f031f27227a0f28e9</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Incorrect parsing of certain JSON input may result in js-bson not correctly serializing BSON. This may cause unexpected application behaviour including data disclosure.
<p>Publish Date: 2020-03-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-2391>CVE-2019-2391</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/mongodb/js-bson/pull/336/commits">https://github.com/mongodb/js-bson/pull/336/commits</a></p>
<p>Release Date: 2020-03-31</p>
<p>Fix Resolution: v1.1.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in bson tgz cve medium severity vulnerability vulnerable library bson tgz a bson parser for node js and the browser library home page a href path to dependency file tmp ws scm todolist api igaku package json path to vulnerable library tmp ws scm todolist api igaku node modules bson package json dependency hierarchy mongoose tgz root library x bson tgz vulnerable library found in head commit a href vulnerability details incorrect parsing of certain json input may result in js bson not correctly serializing bson this may cause unexpected application behaviour including data disclosure publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
45,217 | 2,921,044,640 | IssuesEvent | 2015-06-24 22:09:35 | notsecure/uTox | https://api.github.com/repos/notsecure/uTox | closed | installation too much work on Windows | enhancement high_priority Windows | From getting some people on Windows to use µTox, I found that the biggest installation issue they run into is that they have to choose an installation path manually. It's also unclear when you select a folder whether it'll create a subfolder or it'll just unpack in the folder you selected.
Probably would be best if the default path was set to "C:\Tox" or something, and you can customize it if you want. Most people won't. | 1.0 | installation too much work on Windows - From getting some people on Windows to use µTox, I found that the biggest installation issue they run into is that they have to choose an installation path manually. It's also unclear when you select a folder whether it'll create a subfolder or it'll just unpack in the folder you selected.
Probably would be best if the default path was set to "C:\Tox" or something, and you can customize it if you want. Most people won't. | priority | installation too much work on windows from getting some people on windows to use µtox i found that the biggest installation issue they run into is that they have to choose an installation path manually it s also unclear when you select a folder whether it ll create a subfolder or it ll just unpack in the folder you selected probably would be best if the default path was set to c tox or something and you can customize it if you want most people won t | 1 |
508,664 | 14,704,252,422 | IssuesEvent | 2021-01-04 16:13:25 | WilliBoelke/Cook-Shop | https://api.github.com/repos/WilliBoelke/Cook-Shop | closed | Steps Views | View feature priority 1 | Views for Steps:
- [ ] Add new Step
- [ ] Update a Step
- [ ] View Details
- [ ] RecyclerView item | 1.0 | Steps Views - Views for Steps:
- [ ] Add new Step
- [ ] Update a Step
- [ ] View Details
- [ ] RecyclerView item | priority | steps views views for steps add new step update a step view details recyclerview item | 1 |
37,113 | 2,815,222,665 | IssuesEvent | 2015-05-19 01:42:55 | GoogleCloudPlatform/kubernetes | https://api.github.com/repos/GoogleCloudPlatform/kubernetes | opened | Print object type as first column in `kubectl get` | area/usability priority/P2 | When we originally implemented get, the first column was the type, i.e. "POD", not "NAME". One benefit is that the output (if saved or pasted or viewed later) gives a clue about the type. A downside is that the column width might be longer "SERVICEACCOUNT" this way.
We should make a decision and be consistent (change Pod, or change everything else). | 1.0 | Print object type as first column in `kubectl get` - When we originally implemented get, the first column was the type, i.e. "POD", not "NAME". One benefit is that the output (if saved or pasted or viewed later) gives a clue about the type. A downside is that the column width might be longer "SERVICEACCOUNT" this way.
We should make a decision and be consistent (change Pod, or change everything else). | priority | print object type as first column in kubectl get when we originally implemented get the first column was the type i e pod not name one benefit is that the output if saved or pasted or viewed later gives a clue about the type a downside is that the column width might be longer serviceaccount this way we should make a decision and be consistent change pod or change everything else | 1 |
186,426 | 14,394,694,333 | IssuesEvent | 2020-12-03 01:54:27 | github-vet/rangeclosure-findings | https://api.github.com/repos/github-vet/rangeclosure-findings | closed | flynn/flynn-discovery: Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go; 41 LoC | fresh small test |
Found a possible issue in [flynn/flynn-discovery](https://www.github.com/flynn/flynn-discovery) at [Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
<details>
<summary>Click here to show the 41 line(s) of Go which triggered the analyzer.</summary>
```go
for _, count := range []int{10000, 50, 432} {
recs := make(chan *log.Record)
done := make(chan int)
spec := SpeculativeHandler(100, log.ChannelHandler(recs))
go func() {
defer close(done)
expectedCount := int(math.Min(float64(count), float64(100)))
expectedIdx := count - expectedCount
for r := range recs {
if r.Ctx[1] != expectedIdx {
t.Errorf("Bad ctx 'i', got %d expected %d", r.Ctx[1], expectedIdx)
return
}
expectedIdx++
expectedCount--
if expectedCount == 0 {
// got everything we expected
break
}
}
select {
case <-recs:
t.Errorf("got an extra record we shouldn't have!")
default:
}
}()
lg := log.New()
lg.SetHandler(spec)
for i := 0; i < count; i++ {
lg.Debug("test speculative", "i", i)
}
go spec.Flush()
// wait for the go routine to finish
<-done
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable count used in defer or goroutine at line 52
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 0fdc37b8399dd62aba3cfa194b0f810e945b6a0e
| 1.0 | flynn/flynn-discovery: Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go; 41 LoC -
Found a possible issue in [flynn/flynn-discovery](https://www.github.com/flynn/flynn-discovery) at [Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/flynn/flynn-discovery/blob/0fdc37b8399dd62aba3cfa194b0f810e945b6a0e/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
<details>
<summary>Click here to show the 41 line(s) of Go which triggered the analyzer.</summary>
```go
for _, count := range []int{10000, 50, 432} {
recs := make(chan *log.Record)
done := make(chan int)
spec := SpeculativeHandler(100, log.ChannelHandler(recs))
go func() {
defer close(done)
expectedCount := int(math.Min(float64(count), float64(100)))
expectedIdx := count - expectedCount
for r := range recs {
if r.Ctx[1] != expectedIdx {
t.Errorf("Bad ctx 'i', got %d expected %d", r.Ctx[1], expectedIdx)
return
}
expectedIdx++
expectedCount--
if expectedCount == 0 {
// got everything we expected
break
}
}
select {
case <-recs:
t.Errorf("got an extra record we shouldn't have!")
default:
}
}()
lg := log.New()
lg.SetHandler(spec)
for i := 0; i < count; i++ {
lg.Debug("test speculative", "i", i)
}
go spec.Flush()
// wait for the go routine to finish
<-done
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable count used in defer or goroutine at line 52
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 0fdc37b8399dd62aba3cfa194b0f810e945b6a0e
| non_priority | flynn flynn discovery godeps workspace src gopkg in inconshreveable ext ext test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for count range int recs make chan log record done make chan int spec speculativehandler log channelhandler recs go func defer close done expectedcount int math min count expectedidx count expectedcount for r range recs if r ctx expectedidx t errorf bad ctx i got d expected d r ctx expectedidx return expectedidx expectedcount if expectedcount got everything we expected break select case recs t errorf got an extra record we shouldn t have default lg log new lg sethandler spec for i i count i lg debug test speculative i i go spec flush wait for the go routine to finish done below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable count used in defer or goroutine at line leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
3,136 | 2,810,658,659 | IssuesEvent | 2015-05-17 03:03:02 | HeavensGate/Eternal | https://api.github.com/repos/HeavensGate/Eternal | opened | Opt-in Verb | code Gate/writing | A verb which allows living players to indicate they are now willing to take a mid-round antag role. Preferably not spammable, since it'll bump the admins. | 1.0 | Opt-in Verb - A verb which allows living players to indicate they are now willing to take a mid-round antag role. Preferably not spammable, since it'll bump the admins. | non_priority | opt in verb a verb which allows living players to indicate they are now willing to take a mid round antag role preferably not spammable since it ll bump the admins | 0 |
398,680 | 11,742,175,264 | IssuesEvent | 2020-03-11 23:51:59 | thaliawww/concrexit | https://api.github.com/repos/thaliawww/concrexit | closed | Documents: 2011-2012 is not directly reachable | bug documents easy and fun priority: medium | In GitLab by gmulder on Jun 27, 2017, 01:21
### One-sentence description
The documents page should switch to the year 2011-2012 when it is selected from the drop-down box.
### Current behaviour
The documents page does not switch to the year 2011-2012 when it is selected from the drop-down box.
### Expected behaviour
The documents page switches to the year 2011-2012 when it is selected from the drop-down box.
### Steps to reproduce
1. Log in
2. Go to the `/documents/` page (e.g. Association -> Documents).
3. Navigate to "General meetings"
4. Click on `2011-2012` (a dropdown box appears)
5. Click on `2011-2012` in the dropdown box
6. Nothing happens.
Alternatively, replace step 5 by clicking on another item (e.g. `2010-2011`) and then execute the original step 4 and 5 again. In that case, step 6 is replaced by the expected behaviour.
Note: the period (i.e. `2011-2012`) is relative to the current period. When we switch to the next year, this period will be `2012-2013`. | 1.0 | Documents: 2011-2012 is not directly reachable - In GitLab by gmulder on Jun 27, 2017, 01:21
### One-sentence description
The documents page should switch to the year 2011-2012 when it is selected from the drop-down box.
### Current behaviour
The documents page does not switch to the year 2011-2012 when it is selected from the drop-down box.
### Expected behaviour
The documents page switches to the year 2011-2012 when it is selected from the drop-down box.
### Steps to reproduce
1. Log in
2. Go to the `/documents/` page (e.g. Association -> Documents).
3. Navigate to "General meetings"
4. Click on `2011-2012` (a dropdown box appears)
5. Click on `2011-2012` in the dropdown box
6. Nothing happens.
Alternatively, replace step 5 by clicking on another item (e.g. `2010-2011`) and then execute the original step 4 and 5 again. In that case, step 6 is replaced by the expected behaviour.
Note: the period (i.e. `2011-2012`) is relative to the current period. When we switch to the next year, this period will be `2012-2013`. | priority | documents is not directly reachable in gitlab by gmulder on jun one sentence description the documents page should switch to the year when it is selected from the drop down box current behaviour the documents page does not switch to the year when it is selected from the drop down box expected behaviour the documents page switches to the year when it is selected from the drop down box steps to reproduce log in go to the documents page e g association documents navigate to general meetings click on a dropdown box appears click on in the dropdown box nothing happens alternatively replace step by clicking on another item e g and then execute the original step and again in that case step is replaced by the expected behaviour note the period i e is relative to the current period when we switch to the next year this period will be | 1 |
207,359 | 15,811,822,634 | IssuesEvent | 2021-04-05 03:48:55 | lussierc/StockSwingPredictor | https://api.github.com/repos/lussierc/StockSwingPredictor | closed | Finalize Testing Suite | testing top-priority | As I near the completion of the project source code and the implementation of the prediction model, this code will need to be extensively tested to ensure it functions correctly. These test cases should help ensure a confidence in program correctness.
To do this, test cases should be written for every part of the program. These test files should include:
- `test_scraper.py`: ensure data is scraped and that it is in the correct format
- `test_data_cleaner.py`: ensure data is properly converted
- `test_UI.py`: test the CML and UI lightly to ensure they are working
- `test_prediction.py`: ensures predictions are correctly generated by the program
Now, I want to ensure all of my test cases I write are passing and I have a high success rate. However, in addition to ensuring what is being tested is working properly, I want to ensure that almost everything that can be tested is tested. This will ensure the tool as a whole functions well.
With this, I want to make sure that my tool's code test coverage is high. Code coverage is a measure that analyzes how much source code of a program is executed (or tested) when a particular test suite runs. There is a library called [Coverage](https://coverage.readthedocs.io/en/coverage-5.5/) that will automatically measure the code coverage of your Python test suite, making it perfect for my project. With this, I will be using a version of the library that integrates with Pytest called [pytest-cov](https://pypi.org/project/pytest-cov/). | 1.0 | Finalize Testing Suite - As I near the completion of the project source code and the implementation of the prediction model, this code will need to be extensively tested to ensure it functions correctly. These test cases should help ensure a confidence in program correctness.
To do this, test cases should be written for every part of the program. These test files should include:
- `test_scraper.py`: ensure data is scraped and that it is in the correct format
- `test_data_cleaner.py`: ensure data is properly converted
- `test_UI.py`: test the CML and UI lightly to ensure they are working
- `test_prediction.py`: ensures predictions are correctly generated by the program
Now, I want to ensure all of my test cases I write are passing and I have a high success rate. However, in addition to ensuring what is being tested is working properly, I want to ensure that almost everything that can be tested is tested. This will ensure the tool as a whole functions well.
With this, I want to make sure that my tool's code test coverage is high. Code coverage is a measure that analyzes how much source code of a program is executed (or tested) when a particular test suite runs. There is a library called [Coverage](https://coverage.readthedocs.io/en/coverage-5.5/) that will automatically measure the code coverage of your Python test suite, making it perfect for my project. With this, I will be using a version of the library that integrates with Pytest called [pytest-cov](https://pypi.org/project/pytest-cov/). | non_priority | finalize testing suite as i near the completion of the project source code and the implementation of the prediction model this code will need to be extensively tested to ensure it functions correctly these test cases should help ensure a confidence in program correctness to do this test cases should be written for every part of the program these test files should include test scraper py ensure data is scraped and that it is in the correct format test data cleaner py ensure data is properly converted test ui py test the cml and ui lightly to ensure they are working test prediction py ensures predictions are correctly generated by the program now i want to ensure all of my test cases i write are passing and i have a high success rate however in addition to ensuring what is being tested is working properly i want to ensure that almost everything that can be tested is tested this will ensure the tool as a whole functions well with this i want to make sure that my tool s code test coverage is high code coverage is a measure that analyzes how much source code of a program is executed or tested when a particular test suite runs there is a library called that will automatically measure the code coverage of your python test suite making it perfect for my project with this i will be using a version of the library that integrates with pytest called | 0 |
6,205 | 9,107,428,823 | IssuesEvent | 2019-02-21 04:21:48 | rubberduck-vba/Rubberduck | https://api.github.com/repos/rubberduck-vba/Rubberduck | closed | Getting a parse error with no result in search results window | bug parse-tree-processing | Version 2.2.6672.28001
OS: Microsoft Windows NT 10.0.16299.0, x64
Host Product: Microsoft Office 2013 x86
Host Version: 15.0.5045.1000
Host Executable: MSACCESS.EXE
This is persisting after fixing the error and getting a successful compile.
[RubberduckLog.txt](https://github.com/rubberduck-vba/Rubberduck/files/2506927/RubberduckLog.txt)
| 1.0 | Getting a parse error with no result in search results window - Version 2.2.6672.28001
OS: Microsoft Windows NT 10.0.16299.0, x64
Host Product: Microsoft Office 2013 x86
Host Version: 15.0.5045.1000
Host Executable: MSACCESS.EXE
This is persisting after fixing the error and getting a successful compile.
[RubberduckLog.txt](https://github.com/rubberduck-vba/Rubberduck/files/2506927/RubberduckLog.txt)
| non_priority | getting a parse error with no result in search results window version os microsoft windows nt host product microsoft office host version host executable msaccess exe this is persisting after fixing the error and getting a successful compile | 0 |
232,101 | 25,564,967,337 | IssuesEvent | 2022-11-30 13:40:42 | jtimberlake/pacbot | https://api.github.com/repos/jtimberlake/pacbot | opened | CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz | security vulnerability | ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /webapp/package.json</p>
<p>Path to vulnerable library: /webapp/node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- cli-1.6.8.tgz (Root Library)
- stylus-0.54.8.tgz
- css-parse-2.0.0.tgz
- css-2.2.4.tgz
- source-map-resolve-0.5.3.tgz
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/pacbot/commit/5a25dc15882324583860630e773af2e2647dd5fa">5a25dc15882324583860630e773af2e2647dd5fa</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| True | CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz - ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /webapp/package.json</p>
<p>Path to vulnerable library: /webapp/node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- cli-1.6.8.tgz (Root Library)
- stylus-0.54.8.tgz
- css-parse-2.0.0.tgz
- css-2.2.4.tgz
- source-map-resolve-0.5.3.tgz
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/pacbot/commit/5a25dc15882324583860630e773af2e2647dd5fa">5a25dc15882324583860630e773af2e2647dd5fa</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| non_priority | cve high detected in decode uri component tgz cve high severity vulnerability vulnerable library decode uri component tgz a better decodeuricomponent library home page a href path to dependency file webapp package json path to vulnerable library webapp node modules decode uri component package json dependency hierarchy cli tgz root library stylus tgz css parse tgz css tgz source map resolve tgz x decode uri component tgz vulnerable library found in head commit a href found in base branch master vulnerability details decode uri component is vulnerable to improper input validation resulting in dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href | 0 |
253,798 | 21,705,656,287 | IssuesEvent | 2022-05-10 09:21:29 | FutureNorthants/VirtualWorker | https://api.github.com/repos/FutureNorthants/VirtualWorker | closed | Missing 'Comment - Sent to Email' Field | enhancement North needs retesting | **Bug description**
A clear and concise description of what the bug is, what is happening and what the impact is. NORTH - the comments field which tells an advisor where the case has been forwarded to is not appearing in a certain scenario. The scenario is that a case is created in CXM, the BOT identifies the wrong sovereign, the Hub manually change the sovereign, the customer sends an update, the BOT sends the new update to the relevant sovereign (last chosen one on CXM) however the Comments field is not present to prove where it has been sent. The impact is that staff are having to phone sovereigns to see if cases have been forwarded due to the fact that bug #74 has meant they have not been forwarding, so when the fix for #74 is in place we need to have visual proof of where the case has been sent or not to prevent manual checking.
**Reference number(s)**
EMN000535 in LIVE
If you can - please attempt to replicate the error(s) in test and provide the TEST reference number.
**Bug reproduction**
Steps to reproduce the behaviour:
1. Complete a Contact Us form
2. From the Hub change the Sovereign in CXM
3. Send a customer update on the case
4. See in the case on CXM that the top Comments field, which shows where the case has been forwarded to, is not present.
**Expected behaviour**
1. Customer to complete Contact us form
2. If BOT chooses wrong sovereign (A) the Hub wil reallocate to the correct sovereign (B)
3. Customer sends an update to the case
4. BOT sends the update to the latest sovereign (B)
5. The comments field to be displayed showing the trail for point 4 in CXM so that advisors can see where the update has been sent to. (this bit is missing at the moment)
**Screenshots**
There is one on unitary chat but i can't copy and paste from there and my snipping tool won't open either
**Additional comments**
This will impact us monitoring the Bug fix for #74
| 1.0 | Missing 'Comment - Sent to Email' Field - **Bug description**
A clear and concise description of what the bug is, what is happening and what the impact is. NORTH - the comments field which tells an advisor where the case has been forwarded to is not appearing in a certain scenario. The scenario is that a case is created in CXM, the BOT identifies the wrong sovereign, the Hub manually change the sovereign, the customer sends an update, the BOT sends the new update to the relevant sovereign (last chosen one on CXM) however the Comments field is not present to prove where it has been sent. The impact is that staff are having to phone sovereigns to see if cases have been forwarded due to the fact that bug #74 has meant they have not been forwarding, so when the fix for #74 is in place we need to have visual proof of where the case has been sent or not to prevent manual checking.
**Reference number(s)**
EMN000535 in LIVE
If you can - please attempt to replicate the error(s) in test and provide the TEST reference number.
**Bug reproduction**
Steps to reproduce the behaviour:
1. Complete a Contact Us form
2. From the Hub change the Sovereign in CXM
3. Send a customer update on the case
4. See in the case on CXM that the top Comments field, which shows where the case has been forwarded to, is not present.
**Expected behaviour**
1. Customer to complete Contact us form
2. If BOT chooses wrong sovereign (A) the Hub wil reallocate to the correct sovereign (B)
3. Customer sends an update to the case
4. BOT sends the update to the latest sovereign (B)
5. The comments field to be displayed showing the trail for point 4 in CXM so that advisors can see where the update has been sent to. (this bit is missing at the moment)
**Screenshots**
There is one on unitary chat but i can't copy and paste from there and my snipping tool won't open either
**Additional comments**
This will impact us monitoring the Bug fix for #74
| non_priority | missing comment sent to email field bug description a clear and concise description of what the bug is what is happening and what the impact is north the comments field which tells an advisor where the case has been forwarded to is not appearing in a certain scenario the scenario is that a case is created in cxm the bot identifies the wrong sovereign the hub manually change the sovereign the customer sends an update the bot sends the new update to the relevant sovereign last chosen one on cxm however the comments field is not present to prove where it has been sent the impact is that staff are having to phone sovereigns to see if cases have been forwarded due to the fact that bug has meant they have not been forwarding so when the fix for is in place we need to have visual proof of where the case has been sent or not to prevent manual checking reference number s in live if you can please attempt to replicate the error s in test and provide the test reference number bug reproduction steps to reproduce the behaviour complete a contact us form from the hub change the sovereign in cxm send a customer update on the case see in the case on cxm that the top comments field which shows where the case has been forwarded to is not present expected behaviour customer to complete contact us form if bot chooses wrong sovereign a the hub wil reallocate to the correct sovereign b customer sends an update to the case bot sends the update to the latest sovereign b the comments field to be displayed showing the trail for point in cxm so that advisors can see where the update has been sent to this bit is missing at the moment screenshots there is one on unitary chat but i can t copy and paste from there and my snipping tool won t open either additional comments this will impact us monitoring the bug fix for | 0 |
223,986 | 24,760,213,277 | IssuesEvent | 2022-10-21 22:40:09 | TIBCOSoftware/js-workshops | https://api.github.com/repos/TIBCOSoftware/js-workshops | opened | CVE-2022-37598 (High) detected in uglify-js-3.6.0.tgz | security vulnerability | ## CVE-2022-37598 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-3.6.0.tgz</b></p></summary>
<p>JavaScript parser, mangler/compressor and beautifier toolkit</p>
<p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz</a></p>
<p>Path to dependency file: /UrbanRoast-tutorial/resources/d3-funnel-master/package.json</p>
<p>Path to vulnerable library: /UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/uglify-js/package.json</p>
<p>
Dependency Hierarchy:
- gulp-uglify-3.0.2.tgz (Root Library)
- :x: **uglify-js-3.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TIBCOSoftware/js-workshops/commits/8a027629e1d1d82ce0fb49a090c585960d0967cd">8a027629e1d1d82ce0fb49a090c585960d0967cd</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js.
<p>Publish Date: 2022-10-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37598>CVE-2022-37598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-20</p>
<p>Fix Resolution: uglify-js - 3.13.10</p>
</p>
</details>
<p></p>
| True | CVE-2022-37598 (High) detected in uglify-js-3.6.0.tgz - ## CVE-2022-37598 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-3.6.0.tgz</b></p></summary>
<p>JavaScript parser, mangler/compressor and beautifier toolkit</p>
<p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.6.0.tgz</a></p>
<p>Path to dependency file: /UrbanRoast-tutorial/resources/d3-funnel-master/package.json</p>
<p>Path to vulnerable library: /UrbanRoast-tutorial/resources/d3-funnel-master/node_modules/uglify-js/package.json</p>
<p>
Dependency Hierarchy:
- gulp-uglify-3.0.2.tgz (Root Library)
- :x: **uglify-js-3.6.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/TIBCOSoftware/js-workshops/commits/8a027629e1d1d82ce0fb49a090c585960d0967cd">8a027629e1d1d82ce0fb49a090c585960d0967cd</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js.
<p>Publish Date: 2022-10-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37598>CVE-2022-37598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-20</p>
<p>Fix Resolution: uglify-js - 3.13.10</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in uglify js tgz cve high severity vulnerability vulnerable library uglify js tgz javascript parser mangler compressor and beautifier toolkit library home page a href path to dependency file urbanroast tutorial resources funnel master package json path to vulnerable library urbanroast tutorial resources funnel master node modules uglify js package json dependency hierarchy gulp uglify tgz root library x uglify js tgz vulnerable library found in head commit a href vulnerability details prototype pollution vulnerability in function defnode in ast js in mishoo uglifyjs via the name variable in ast js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution uglify js | 0 |
351,031 | 25,010,838,218 | IssuesEvent | 2022-11-03 15:08:41 | AY2223S1-CS2103T-W16-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-W16-1/tp | closed | [PE-D][Tester B] Finding by a date range can return other irrelevant search results | type.Bug documentation severity.Low good to fix | Using this sample data:

After typing `find 01/01/23 - 07/01/23`:

This might be because the first search result has a “-“ as a request under room status, which the command detects. However, novice users might be confused by this.
<!--session: 1666944239122-5fa2ca7e-3be6-428b-8e43-2cc245ff74fe-->
<!--Version: Web v3.4.4-->
-------------
Labels: `type.FunctionalityBug` `severity.Low`
original: NicholasTYD/ped#7 | 1.0 | [PE-D][Tester B] Finding by a date range can return other irrelevant search results - Using this sample data:

After typing `find 01/01/23 - 07/01/23`:

This might be because the first search result has a “-“ as a request under room status, which the command detects. However, novice users might be confused by this.
<!--session: 1666944239122-5fa2ca7e-3be6-428b-8e43-2cc245ff74fe-->
<!--Version: Web v3.4.4-->
-------------
Labels: `type.FunctionalityBug` `severity.Low`
original: NicholasTYD/ped#7 | non_priority | finding by a date range can return other irrelevant search results using this sample data after typing find this might be because the first search result has a “ “ as a request under room status which the command detects however novice users might be confused by this labels type functionalitybug severity low original nicholastyd ped | 0 |
179,278 | 14,700,231,926 | IssuesEvent | 2021-01-04 09:53:31 | superwerker/superwerker | https://api.github.com/repos/superwerker/superwerker | opened | Prepare for AWS Quick Starts | documentation | ## Tasks
- [ ] Logo (Quick Start Owner _This Quick Start was developed by AWS solutions architects._ e.g.)
- [ ] High Level Briefing
- [ ] Documentation (_Deployment Guide_)
- Website ( https://aws-quickstart.github.io/quickstart-dotnetcore-cicd/ )
- PDF ( https://aws-quickstart.s3.amazonaws.com/quickstart-trek10-serverless-enterprise-cicd/doc/serverless-cicd-for-the-enterprise-on-the-aws-cloud.pdf )
- [ ] Architecture Diagramm
## Topics
- What you'll build
- How to deploy
- Cost and licenses
- Resources
## Examples
- https://aws.amazon.com/quickstart/architecture/serverless-cicd-for-enterprise/
- https://aws.amazon.com/quickstart/architecture/eks-snyk/
- https://aws.amazon.com/quickstart/architecture/terraform-modules-on-aws/
- https://aws.amazon.com/quickstart/architecture/ibm-maximo/ | 1.0 | Prepare for AWS Quick Starts - ## Tasks
- [ ] Logo (Quick Start Owner _This Quick Start was developed by AWS solutions architects._ e.g.)
- [ ] High Level Briefing
- [ ] Documentation (_Deployment Guide_)
- Website ( https://aws-quickstart.github.io/quickstart-dotnetcore-cicd/ )
- PDF ( https://aws-quickstart.s3.amazonaws.com/quickstart-trek10-serverless-enterprise-cicd/doc/serverless-cicd-for-the-enterprise-on-the-aws-cloud.pdf )
- [ ] Architecture Diagramm
## Topics
- What you'll build
- How to deploy
- Cost and licenses
- Resources
## Examples
- https://aws.amazon.com/quickstart/architecture/serverless-cicd-for-enterprise/
- https://aws.amazon.com/quickstart/architecture/eks-snyk/
- https://aws.amazon.com/quickstart/architecture/terraform-modules-on-aws/
- https://aws.amazon.com/quickstart/architecture/ibm-maximo/ | non_priority | prepare for aws quick starts tasks logo quick start owner this quick start was developed by aws solutions architects e g high level briefing documentation deployment guide website pdf architecture diagramm topics what you ll build how to deploy cost and licenses resources examples | 0 |
72,924 | 8,790,572,890 | IssuesEvent | 2018-12-21 09:32:19 | nextcloud/server | https://api.github.com/repos/nextcloud/server | closed | App icon disappears for overflowing settings page names | 3. to review bug design stale | ### Steps to reproduce
1. Create app with a rather long settings page name
2. Apply app-icon to app
3. App icon will not be shown in e.g. /index.php/settings/user
### Expected behaviour
Settings page name gets shortened, but app icon stays visible.
### Actual behaviour
Settings page name is shortened and app icon disappears.
### Nextcloud version
13.0.0RC1

This is an issue that probably needs refactoring of the HTML structure.
We have this
```
<a href="/nextcloud/index.php/settings/admin/my_app">
<img alt="" src="/nextcloud/apps/my_app/img/app-dark.svg">
<span>A very long app settings page name that is too long</span>
</a>
```
And because the a-tag gets a padding at the position where the img is placed, it will be hidden, once the span-text overflows, because of overflow: hidden. At this moment I have no simple idea how to solve this, maybe someone else has. | 1.0 | App icon disappears for overflowing settings page names - ### Steps to reproduce
1. Create app with a rather long settings page name
2. Apply app-icon to app
3. App icon will not be shown in e.g. /index.php/settings/user
### Expected behaviour
Settings page name gets shortened, but app icon stays visible.
### Actual behaviour
Settings page name is shortened and app icon disappears.
### Nextcloud version
13.0.0RC1

This is an issue that probably needs refactoring of the HTML structure.
We have this
```
<a href="/nextcloud/index.php/settings/admin/my_app">
<img alt="" src="/nextcloud/apps/my_app/img/app-dark.svg">
<span>A very long app settings page name that is too long</span>
</a>
```
And because the a-tag gets a padding at the position where the img is placed, it will be hidden, once the span-text overflows, because of overflow: hidden. At this moment I have no simple idea how to solve this, maybe someone else has. | non_priority | app icon disappears for overflowing settings page names steps to reproduce create app with a rather long settings page name apply app icon to app app icon will not be shown in e g index php settings user expected behaviour settings page name gets shortened but app icon stays visible actual behaviour settings page name is shortened and app icon disappears nextcloud version this is an issue that probably needs refactoring of the html structure we have this a very long app settings page name that is too long and because the a tag gets a padding at the position where the img is placed it will be hidden once the span text overflows because of overflow hidden at this moment i have no simple idea how to solve this maybe someone else has | 0 |
311,966 | 23,411,977,491 | IssuesEvent | 2022-08-12 18:36:10 | bcbi/PreprocessMD.jl | https://api.github.com/repos/bcbi/PreprocessMD.jl | closed | Examples should display function inputs | documentation good first issue | The docstring examples for each function display the output DataFrame. The input DataFrames should also be displayed so that the user can see how the output relates to the input, especially for `pivot()`. | 1.0 | Examples should display function inputs - The docstring examples for each function display the output DataFrame. The input DataFrames should also be displayed so that the user can see how the output relates to the input, especially for `pivot()`. | non_priority | examples should display function inputs the docstring examples for each function display the output dataframe the input dataframes should also be displayed so that the user can see how the output relates to the input especially for pivot | 0 |
784,904 | 27,588,995,226 | IssuesEvent | 2023-03-08 22:20:11 | AdrKacz/super-duper-guacamole | https://api.github.com/repos/AdrKacz/super-duper-guacamole | opened | ❓ Comment profiter du cercles d'amis des utilisateurs | question priority:P1 | # What do you need to discuss?
Chaque utilisateurs à des amis, qui pourraient être des clients potentiels, ce qui rendrait à terme l'application viral.
# Why is that important?
Si on pouvait faire circuler l'application entre les cercles d'amis, cela augmenterait le nombre de téléchargement. Hors, comme on travaille un peu sur un réseau social, plus il y a de monde plus c'est fun.
# What are the element we will need to discuss it?
Quand il est question de _referal_ chaque app à sa stratégie, par exemple Revolut te paie 60€ par utilisateur apporté et Snapshat accéde a tes contacts pour te faciliter la communication avec eux. | 1.0 | ❓ Comment profiter du cercles d'amis des utilisateurs - # What do you need to discuss?
Chaque utilisateurs à des amis, qui pourraient être des clients potentiels, ce qui rendrait à terme l'application viral.
# Why is that important?
Si on pouvait faire circuler l'application entre les cercles d'amis, cela augmenterait le nombre de téléchargement. Hors, comme on travaille un peu sur un réseau social, plus il y a de monde plus c'est fun.
# What are the element we will need to discuss it?
Quand il est question de _referal_ chaque app à sa stratégie, par exemple Revolut te paie 60€ par utilisateur apporté et Snapshat accéde a tes contacts pour te faciliter la communication avec eux. | priority | ❓ comment profiter du cercles d amis des utilisateurs what do you need to discuss chaque utilisateurs à des amis qui pourraient être des clients potentiels ce qui rendrait à terme l application viral why is that important si on pouvait faire circuler l application entre les cercles d amis cela augmenterait le nombre de téléchargement hors comme on travaille un peu sur un réseau social plus il y a de monde plus c est fun what are the element we will need to discuss it quand il est question de referal chaque app à sa stratégie par exemple revolut te paie € par utilisateur apporté et snapshat accéde a tes contacts pour te faciliter la communication avec eux | 1 |
16,976 | 3,587,160,365 | IssuesEvent | 2016-01-30 04:04:50 | tgstation/-tg-station | https://api.github.com/repos/tgstation/-tg-station | closed | You can teleport via mulebots | Bug In Game Exploit Tested / Reproduced | rev: 0debb6cb68829429f4d53f0e52e30dcbc0abcbf6
H2Repr0duce420noblazetelep0rt:
0. Turn off 'return home after delivery' on mulebot
1. Have your mulebot deliver you somewhere
2. Try to climb back onto the mule
3. Get blocked by flaps, you don't end up on the mulebot
4. Send mulebot home/somewhere else
5. Wait
6. Get teleported to the mule's latest drop point | 1.0 | You can teleport via mulebots - rev: 0debb6cb68829429f4d53f0e52e30dcbc0abcbf6
H2Repr0duce420noblazetelep0rt:
0. Turn off 'return home after delivery' on mulebot
1. Have your mulebot deliver you somewhere
2. Try to climb back onto the mule
3. Get blocked by flaps, you don't end up on the mulebot
4. Send mulebot home/somewhere else
5. Wait
6. Get teleported to the mule's latest drop point | non_priority | you can teleport via mulebots rev turn off return home after delivery on mulebot have your mulebot deliver you somewhere try to climb back onto the mule get blocked by flaps you don t end up on the mulebot send mulebot home somewhere else wait get teleported to the mule s latest drop point | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.