Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7,747 | 7,078,990,444 | IssuesEvent | 2018-01-10 07:35:52 | spring1944/spring1944 | https://api.github.com/repos/spring1944/spring1944 | closed | RFC: ditch wordpress/forum setup in favor of a static github page | infrastructure | Chatted with hoko about a direction for the site, and I think it's at least worth considering ditching our current setup.
Basically: We have three big pieces of software run on our behalf by koshi: wordpress, mediawiki, and phpbb. All of them are famously insecure, and we hardly use them (other than the wiki, which I'll get to).
I'm thinking about moving us to a static page hosted on github. Here's a quick breakdown of what we use in our current setup, and how we might replace it.
#### Wordpress --> static github project page
The few useful bits of content on the site (maps, media) can be moved over. For the other wordpress features -- we barely make use of wordpress for the authoring stuff, and the comments are mostly spam. Our image gallery is mostly dead links at this point, and changes so rarely that keeping screenshots in a branch of a repo would be perfectly fine. We're all sufficiently technical that writing newsposts by editing some markdown files and git push-ing is fine.
#### Mediawiki --> github wiki / curated content as part of the site
Mediawiki is probably the better wiki software, but porting the useful content wouldn't be very hard to do, since our wiki doesn't make extensive use of formatting, and it isn't themed in a meaningful way.
That said, I think I would promote the best bits of the wiki to proper pages on the static site, so we can polish and curate and promote the best resources in a highly visible place. Other stuff can go into the github wiki, for assorted 'things users want to contribute'.
#### Forum --> ???
This might be the most contentious point -- giving up our dedicated forum. I can grab a backup from koshi, so the archive would still be available, but I think our forum is actually fairly damaging to our image. Our repo is really active (at the moment :smiley:), but our forum is kind of a ghost town. One of the first things I do when checking out an open source game project is peek at the forums. No recent posts == no players, so I usually bail.
I think we would be better served by drawing people directly into `#s44` using the IRC bridge by having a web-IRC thingy available from the static page, maybe in addition to some public logging or other persistent messaging. I suspect that between github issues and #s44, our communication needs are pretty much covered.
Thoughts?
| 1.0 | RFC: ditch wordpress/forum setup in favor of a static github page - Chatted with hoko about a direction for the site, and I think it's at least worth considering ditching our current setup.
Basically: We have three big pieces of software run on our behalf by koshi: wordpress, mediawiki, and phpbb. All of them are famously insecure, and we hardly use them (other than the wiki, which I'll get to).
I'm thinking about moving us to a static page hosted on github. Here's a quick breakdown of what we use in our current setup, and how we might replace it.
#### Wordpress --> static github project page
The few useful bits of content on the site (maps, media) can be moved over. For the other wordpress features -- we barely make use of wordpress for the authoring stuff, and the comments are mostly spam. Our image gallery is mostly dead links at this point, and changes so rarely that keeping screenshots in a branch of a repo would be perfectly fine. We're all sufficiently technical that writing newsposts by editing some markdown files and git push-ing is fine.
#### Mediawiki --> github wiki / curated content as part of the site
Mediawiki is probably the better wiki software, but porting the useful content wouldn't be very hard to do, since our wiki doesn't make extensive use of formatting, and it isn't themed in a meaningful way.
That said, I think I would promote the best bits of the wiki to proper pages on the static site, so we can polish and curate and promote the best resources in a highly visible place. Other stuff can go into the github wiki, for assorted 'things users want to contribute'.
#### Forum --> ???
This might be the most contentious point -- giving up our dedicated forum. I can grab a backup from koshi, so the archive would still be available, but I think our forum is actually fairly damaging to our image. Our repo is really active (at the moment :smiley:), but our forum is kind of a ghost town. One of the first things I do when checking out an open source game project is peek at the forums. No recent posts == no players, so I usually bail.
I think we would be better served by drawing people directly into `#s44` using the IRC bridge by having a web-IRC thingy available from the static page, maybe in addition to some public logging or other persistent messaging. I suspect that between github issues and #s44, our communication needs are pretty much covered.
Thoughts?
| infrastructure | rfc ditch wordpress forum setup in favor of a static github page chatted with hoko about a direction for the site and i think it s at least worth considering ditching our current setup basically we have three big pieces of software run on our behalf by koshi wordpress mediawiki and phpbb all of them are famously insecure and we hardly use them other than the wiki which i ll get to i m thinking about moving us to a static page hosted on github here s a quick breakdown of what we use in our current setup and how we might replace it wordpress static github project page the few useful bits of content on the site maps media can be moved over for the other wordpress features we barely make use of wordpress for the authoring stuff and the comments are mostly spam our image gallery is mostly dead links at this point and changes so rarely that keeping screenshots in a branch of a repo would be perfectly fine we re all sufficiently technical that writing newsposts by editing some markdown files and git push ing is fine mediawiki github wiki curated content as part of the site mediawiki is probably the better wiki software but porting the useful content wouldn t be very hard to do since our wiki doesn t make extensive use of formatting and it isn t themed in a meaningful way that said i think i would promote the best bits of the wiki to proper pages on the static site so we can polish and curate and promote the best resources in a highly visible place other stuff can go into the github wiki for assorted things users want to contribute forum this might be the most contentious point giving up our dedicated forum i can grab a backup from koshi so the archive would still be available but i think our forum is actually fairly damaging to our image our repo is really active at the moment smiley but our forum is kind of a ghost town one of the first things i do when checking out an open source game project is peek at the forums no recent posts no players so i usually bail i think we would be better served by drawing people directly into using the irc bridge by having a web irc thingy available from the static page maybe in addition to some public logging or other persistent messaging i suspect that between github issues and our communication needs are pretty much covered thoughts | 1 |
60,823 | 8,468,406,688 | IssuesEvent | 2018-10-23 19:39:21 | devtools-html/perf.html | https://api.github.com/repos/devtools-html/perf.html | opened | Improve Enzyme best practice docs | documentation | From #1401, there were some updates to the Enzyme testing docs, but they could be better still.
@julienw wrote:
> I'd like some guidance of how to use enzyme; because enzyme is very powerful it's easy to use it in a bad way (eg: manipulating props), and I think we should take great care to test the components from the point of view of the user.
| 1.0 | Improve Enzyme best practice docs - From #1401, there were some updates to the Enzyme testing docs, but they could be better still.
@julienw wrote:
> I'd like some guidance of how to use enzyme; because enzyme is very powerful it's easy to use it in a bad way (eg: manipulating props), and I think we should take great care to test the components from the point of view of the user.
| non_infrastructure | improve enzyme best practice docs from there were some updates to the enzyme testing docs but they could be better still julienw wrote i d like some guidance of how to use enzyme because enzyme is very powerful it s easy to use it in a bad way eg manipulating props and i think we should take great care to test the components from the point of view of the user | 0 |
531,903 | 15,527,114,081 | IssuesEvent | 2021-03-13 04:19:55 | OnTopicCMS/OnTopic-Library | https://api.github.com/repos/OnTopicCMS/OnTopic-Library | closed | Bug: TrackedRecord<T> constructor doesn't accept null value | Area: Entity Priority: 3 Severity 1: Minor Status 5: Complete Type: Bug | The `TrackedRecord<T>.Value` property is nullable, but when constructing a new `TrackedRecord<T>` via the constructor, the `value` parameter is required—both in terms of its nullability annotation, as well as an explicit guard clause. The same is true of the derived `TopicReferenceRecord` and `AttributeRecord` constructors.
In practice, we generally prefer creating these via e.g. `TrackedRecord<T>.SetValue()`, which doesn't use the constructor, and can set a `null` value. But this mismatch isn't consistent with the data model. As such, the constructors should be updated to maintain parity with the underlying property return types they represent. | 1.0 | Bug: TrackedRecord<T> constructor doesn't accept null value - The `TrackedRecord<T>.Value` property is nullable, but when constructing a new `TrackedRecord<T>` via the constructor, the `value` parameter is required—both in terms of its nullability annotation, as well as an explicit guard clause. The same is true of the derived `TopicReferenceRecord` and `AttributeRecord` constructors.
In practice, we generally prefer creating these via e.g. `TrackedRecord<T>.SetValue()`, which doesn't use the constructor, and can set a `null` value. But this mismatch isn't consistent with the data model. As such, the constructors should be updated to maintain parity with the underlying property return types they represent. | non_infrastructure | bug trackedrecord constructor doesn t accept null value the trackedrecord value property is nullable but when constructing a new trackedrecord via the constructor the value parameter is required—both in terms of its nullability annotation as well as an explicit guard clause the same is true of the derived topicreferencerecord and attributerecord constructors in practice we generally prefer creating these via e g trackedrecord setvalue which doesn t use the constructor and can set a null value but this mismatch isn t consistent with the data model as such the constructors should be updated to maintain parity with the underlying property return types they represent | 0 |
11,418 | 9,181,319,045 | IssuesEvent | 2019-03-05 09:56:57 | elastic/beats | https://api.github.com/repos/elastic/beats | opened | [Metricbeat] Improve / add integration tests of Ceph | :infrastructure Metricbeat | In its current state Ceph module don't have any integration tests because Ceph containers take too long to start: over 1 minute in most situations and more than 5 even sometimes, producing a flaky test because `compose.EnsureUp` function we use defaults to 5 minutes. Probably, it has some problem if it reaches 5 minutes for starting up that wouldn't allow it to ever start properly.
`TestData` must be manually skipped for the same reason as it tries to reach the server that is not started as described here https://github.com/elastic/beats/pull/10990#discussion_r262406519 and here https://github.com/elastic/beats/pull/10993#discussion_r261627931
As soon as we have this merged https://github.com/elastic/beats/pull/10648 the module should be more friendly to add those tests. | 1.0 | [Metricbeat] Improve / add integration tests of Ceph - In its current state Ceph module don't have any integration tests because Ceph containers take too long to start: over 1 minute in most situations and more than 5 even sometimes, producing a flaky test because `compose.EnsureUp` function we use defaults to 5 minutes. Probably, it has some problem if it reaches 5 minutes for starting up that wouldn't allow it to ever start properly.
`TestData` must be manually skipped for the same reason as it tries to reach the server that is not started as described here https://github.com/elastic/beats/pull/10990#discussion_r262406519 and here https://github.com/elastic/beats/pull/10993#discussion_r261627931
As soon as we have this merged https://github.com/elastic/beats/pull/10648 the module should be more friendly to add those tests. | infrastructure | improve add integration tests of ceph in its current state ceph module don t have any integration tests because ceph containers take too long to start over minute in most situations and more than even sometimes producing a flaky test because compose ensureup function we use defaults to minutes probably it has some problem if it reaches minutes for starting up that wouldn t allow it to ever start properly testdata must be manually skipped for the same reason as it tries to reach the server that is not started as described here and here as soon as we have this merged the module should be more friendly to add those tests | 1 |
672,910 | 22,908,233,369 | IssuesEvent | 2022-07-16 00:01:29 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | Visualization: Bar chart not showing 0 values unlike line and area chart | Type:Bug Priority:P3 .Frontend Visualization/Charts | This is the only option when choosing bar chart.

Whilst below images are for Line and Area chart.


This is the same as for previous metabase version that I was on around v0.3+
However, there was a workaround (just switch to line/area chart then select **"Replace missing values" with Zero** then revert to bar chart, missing values would then be replaced with 0 value) that I did back then that is now not working on my current version.
{
"browser-info": {
"language": "en-US",
"platform": "Win32",
"userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:100.0) Gecko/20100101 Firefox/100.0",
"vendor": ""
},
"system-info": {
"file.encoding": "UTF-8",
"java.runtime.name": "OpenJDK Runtime Environment",
"java.runtime.version": "11.0.15+10-Ubuntu-0ubuntu0.18.04.1",
"java.vendor": "Private Build",
"java.vendor.url": "Unknown",
"java.version": "11.0.15",
"java.vm.name": "OpenJDK 64-Bit Server VM",
"java.vm.version": "11.0.15+10-Ubuntu-0ubuntu0.18.04.1",
"os.name": "Linux",
"os.version": "5.4.0-1077-azure",
"user.language": "en",
"user.timezone": "Etc/UTC"
},
"metabase-info": {
"databases": [
"mysql",
"mongo",
"googleanalytics"
],
"hosting-env": "unknown",
"application-database": "h2",
"application-database-details": {
"database": {
"name": "H2",
"version": "1.4.197 (2018-03-18)"
},
"jdbc-driver": {
"name": "H2 JDBC Driver",
"version": "1.4.197 (2018-03-18)"
}
},
"run-mode": "prod",
"version": {
"date": "2022-04-07",
"tag": "v0.42.4",
"branch": "release-x.42.x",
"hash": "7c3ce2d"
},
"settings": {
"report-timezone": "Asia/Hong_Kong"
}
}
}
| 1.0 | Visualization: Bar chart not showing 0 values unlike line and area chart - This is the only option when choosing bar chart.

Whilst below images are for Line and Area chart.


This is the same as for previous metabase version that I was on around v0.3+
However, there was a workaround (just switch to line/area chart then select **"Replace missing values" with Zero** then revert to bar chart, missing values would then be replaced with 0 value) that I did back then that is now not working on my current version.
{
"browser-info": {
"language": "en-US",
"platform": "Win32",
"userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:100.0) Gecko/20100101 Firefox/100.0",
"vendor": ""
},
"system-info": {
"file.encoding": "UTF-8",
"java.runtime.name": "OpenJDK Runtime Environment",
"java.runtime.version": "11.0.15+10-Ubuntu-0ubuntu0.18.04.1",
"java.vendor": "Private Build",
"java.vendor.url": "Unknown",
"java.version": "11.0.15",
"java.vm.name": "OpenJDK 64-Bit Server VM",
"java.vm.version": "11.0.15+10-Ubuntu-0ubuntu0.18.04.1",
"os.name": "Linux",
"os.version": "5.4.0-1077-azure",
"user.language": "en",
"user.timezone": "Etc/UTC"
},
"metabase-info": {
"databases": [
"mysql",
"mongo",
"googleanalytics"
],
"hosting-env": "unknown",
"application-database": "h2",
"application-database-details": {
"database": {
"name": "H2",
"version": "1.4.197 (2018-03-18)"
},
"jdbc-driver": {
"name": "H2 JDBC Driver",
"version": "1.4.197 (2018-03-18)"
}
},
"run-mode": "prod",
"version": {
"date": "2022-04-07",
"tag": "v0.42.4",
"branch": "release-x.42.x",
"hash": "7c3ce2d"
},
"settings": {
"report-timezone": "Asia/Hong_Kong"
}
}
}
| non_infrastructure | visualization bar chart not showing values unlike line and area chart this is the only option when choosing bar chart whilst below images are for line and area chart this is the same as for previous metabase version that i was on around however there was a workaround just switch to line area chart then select replace missing values with zero then revert to bar chart missing values would then be replaced with value that i did back then that is now not working on my current version browser info language en us platform useragent mozilla windows nt rv gecko firefox vendor system info file encoding utf java runtime name openjdk runtime environment java runtime version ubuntu java vendor private build java vendor url unknown java version java vm name openjdk bit server vm java vm version ubuntu os name linux os version azure user language en user timezone etc utc metabase info databases mysql mongo googleanalytics hosting env unknown application database application database details database name version jdbc driver name jdbc driver version run mode prod version date tag branch release x x hash settings report timezone asia hong kong | 0 |
3,627 | 14,672,547,338 | IssuesEvent | 2020-12-30 10:53:23 | Homebrew/homebrew-core | https://api.github.com/repos/Homebrew/homebrew-core | opened | luajit probably needs to be deprecated | help wanted maintainer feedback | - The latest release (stable OR beta) is from 2017
- It's heavily patched
- Every new macOS version requires an additional patch
- Upstream's recommendation is to “build from git HEAD”, and they won't apparently ship new releases: https://github.com/LuaJIT/LuaJIT/issues/648#issuecomment-752404043
The reason I'm not doing a pull request directly is that a lot of things depend on luajit, so I want to open a discussion and figure out the best way to handle this. Can some of these be migrated to one of the lua formulas? | True | luajit probably needs to be deprecated - - The latest release (stable OR beta) is from 2017
- It's heavily patched
- Every new macOS version requires an additional patch
- Upstream's recommendation is to “build from git HEAD”, and they won't apparently ship new releases: https://github.com/LuaJIT/LuaJIT/issues/648#issuecomment-752404043
The reason I'm not doing a pull request directly is that a lot of things depend on luajit, so I want to open a discussion and figure out the best way to handle this. Can some of these be migrated to one of the lua formulas? | non_infrastructure | luajit probably needs to be deprecated the latest release stable or beta is from it s heavily patched every new macos version requires an additional patch upstream s recommendation is to “build from git head” and they won t apparently ship new releases the reason i m not doing a pull request directly is that a lot of things depend on luajit so i want to open a discussion and figure out the best way to handle this can some of these be migrated to one of the lua formulas | 0 |
114,416 | 11,846,416,770 | IssuesEvent | 2020-03-24 10:09:36 | kyma-incubator/documentation-component | https://api.github.com/repos/kyma-incubator/documentation-component | reopened | Provide examples in JavaScript | area/documentation enhancement stale | **Description**
- have examples with both, JS and TS like we have here https://github.com/kyma-project/kyma/blob/master/docs/kyma/04-04-cluster-installation.md#prerequisites (raw https://raw.githubusercontent.com/kyma-project/kyma/master/docs/kyma/04-04-cluster-installation.md)
- JS is always first on the list
**Reasons**
At the moment all examples for the component are in TypeScript, except of sandbox projects.
Believe it or not, but there is still a great number of javascript developers that do not need TypeScript at all to write good quality code :D https://insights.stackoverflow.com/survey/2019#technology-_-programming-scripting-and-markup-languages | 1.0 | Provide examples in JavaScript - **Description**
- have examples with both, JS and TS like we have here https://github.com/kyma-project/kyma/blob/master/docs/kyma/04-04-cluster-installation.md#prerequisites (raw https://raw.githubusercontent.com/kyma-project/kyma/master/docs/kyma/04-04-cluster-installation.md)
- JS is always first on the list
**Reasons**
At the moment all examples for the component are in TypeScript, except of sandbox projects.
Believe it or not, but there is still a great number of javascript developers that do not need TypeScript at all to write good quality code :D https://insights.stackoverflow.com/survey/2019#technology-_-programming-scripting-and-markup-languages | non_infrastructure | provide examples in javascript description have examples with both js and ts like we have here raw js is always first on the list reasons at the moment all examples for the component are in typescript except of sandbox projects believe it or not but there is still a great number of javascript developers that do not need typescript at all to write good quality code d | 0 |
11,095 | 8,924,513,643 | IssuesEvent | 2019-01-21 18:58:37 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | metricbeat - filesystem & fsstat not collecting mount path on windows machines | :Windows :infrastructure Metricbeat module | Hi,
it's look like the metricbeat not collecting any data of devices without drive letter on windows machines (mount path), the reason for that is metricbeat relay wrong command:
get-psdrive -PSProvider filesystem
this powershell syntax by design not returning mount path.
they should relay on different command:
Get-WmiObject Win32_Volume | Format-Table Name, Label, FreeSpace, Capacity
metricbeat 6.2.4 | 1.0 | metricbeat - filesystem & fsstat not collecting mount path on windows machines - Hi,
it's look like the metricbeat not collecting any data of devices without drive letter on windows machines (mount path), the reason for that is metricbeat relay wrong command:
get-psdrive -PSProvider filesystem
this powershell syntax by design not returning mount path.
they should relay on different command:
Get-WmiObject Win32_Volume | Format-Table Name, Label, FreeSpace, Capacity
metricbeat 6.2.4 | infrastructure | metricbeat filesystem fsstat not collecting mount path on windows machines hi it s look like the metricbeat not collecting any data of devices without drive letter on windows machines mount path the reason for that is metricbeat relay wrong command get psdrive psprovider filesystem this powershell syntax by design not returning mount path they should relay on different command get wmiobject volume format table name label freespace capacity metricbeat | 1 |
38,871 | 6,712,075,254 | IssuesEvent | 2017-10-13 07:55:56 | Microsoft/WindowsTemplateStudio | https://api.github.com/repos/Microsoft/WindowsTemplateStudio | closed | Minimum requirements to install WTS have changed to VS 2017 Update 3 or higher and .NET 4.7 | Documentation fall-creators-update | Currently the minimum requirements to install WTS are VS 2017 Update 3 or higher and .NET 4.7.
By now, the TargetVersion for generated projects still point to 10.0.15063.0. When SDK for FCU ready (publicly available) we will update this too.
We need to update the requirements in the documentation. | 1.0 | Minimum requirements to install WTS have changed to VS 2017 Update 3 or higher and .NET 4.7 - Currently the minimum requirements to install WTS are VS 2017 Update 3 or higher and .NET 4.7.
By now, the TargetVersion for generated projects still point to 10.0.15063.0. When SDK for FCU ready (publicly available) we will update this too.
We need to update the requirements in the documentation. | non_infrastructure | minimum requirements to install wts have changed to vs update or higher and net currently the minimum requirements to install wts are vs update or higher and net by now the targetversion for generated projects still point to when sdk for fcu ready publicly available we will update this too we need to update the requirements in the documentation | 0 |
8,616 | 7,525,090,886 | IssuesEvent | 2018-04-13 09:26:09 | OCR-D/pyocrd | https://api.github.com/repos/OCR-D/pyocrd | closed | Encapsulate merging functionality in an API | infrastructure | `merge_ocr_txt.py` is a command line script containing actual functionality which should be available via API. | 1.0 | Encapsulate merging functionality in an API - `merge_ocr_txt.py` is a command line script containing actual functionality which should be available via API. | infrastructure | encapsulate merging functionality in an api merge ocr txt py is a command line script containing actual functionality which should be available via api | 1 |
15,392 | 11,493,089,346 | IssuesEvent | 2020-02-11 22:17:57 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | opened | [cms-ci] Sync files from upstream | ⭐️ Infrastructure | Sync files from upstream and have all new environments consume that tar file. Create similar command as `cms-db-download` e.g. `cms-files-download`.
This should be synced at same interval as DB, `10 * * * *`.
Current logs have a lot of "image download failed" in logs. | 1.0 | [cms-ci] Sync files from upstream - Sync files from upstream and have all new environments consume that tar file. Create similar command as `cms-db-download` e.g. `cms-files-download`.
This should be synced at same interval as DB, `10 * * * *`.
Current logs have a lot of "image download failed" in logs. | infrastructure | sync files from upstream sync files from upstream and have all new environments consume that tar file create similar command as cms db download e g cms files download this should be synced at same interval as db current logs have a lot of image download failed in logs | 1 |
14,826 | 11,184,393,250 | IssuesEvent | 2019-12-31 17:56:04 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | closed | Data Hydration | EDTR Prototype category:models-and-data-infrastructure | Investigate and implement logic for hydrating the Apollo Client cache using the data that is already being dumped into the DOM.
The relations data can be dumped separately as it is not stored in Apollo cache. -mw | 1.0 | Data Hydration - Investigate and implement logic for hydrating the Apollo Client cache using the data that is already being dumped into the DOM.
The relations data can be dumped separately as it is not stored in Apollo cache. -mw | infrastructure | data hydration investigate and implement logic for hydrating the apollo client cache using the data that is already being dumped into the dom the relations data can be dumped separately as it is not stored in apollo cache mw | 1 |
234,404 | 17,954,049,305 | IssuesEvent | 2021-09-13 04:10:09 | PySimpleGUI/PySimpleGUI | https://api.github.com/repos/PySimpleGUI/PySimpleGUI | closed | Problem when graphing a Line | documentation | ### Type of Issues (Enhancement, Error, Bug, Question)
Bug/Question
### Operating System
Windows 10
### Python version
3.7.0
### PySimpleGUI Port and Version
PySimpleGUIQt 0.26.0
### Code or partial code causing the problem
MainRadioCenter_elem = sg.Graph(canvas_size=(600,400),graph_bottom_left=(0,0), graph_top_right=(600,400), background_color='white' )
MainRadioCenter_elem.DrawLine(point_from=(78,240), point_to=(522,240), color='blue')
I don't quite understand whats happening or if I did something wrong but this error happens when attempting the part of the code above

| 1.0 | Problem when graphing a Line - ### Type of Issues (Enhancement, Error, Bug, Question)
Bug/Question
### Operating System
Windows 10
### Python version
3.7.0
### PySimpleGUI Port and Version
PySimpleGUIQt 0.26.0
### Code or partial code causing the problem
MainRadioCenter_elem = sg.Graph(canvas_size=(600,400),graph_bottom_left=(0,0), graph_top_right=(600,400), background_color='white' )
MainRadioCenter_elem.DrawLine(point_from=(78,240), point_to=(522,240), color='blue')
I don't quite understand whats happening or if I did something wrong but this error happens when attempting the part of the code above

| non_infrastructure | problem when graphing a line type of issues enhancement error bug question bug question operating system windows python version pysimplegui port and version pysimpleguiqt code or partial code causing the problem mainradiocenter elem sg graph canvas size graph bottom left graph top right background color white mainradiocenter elem drawline point from point to color blue i don t quite understand whats happening or if i did something wrong but this error happens when attempting the part of the code above | 0 |
556,615 | 16,487,120,469 | IssuesEvent | 2021-05-24 19:46:35 | venturemark/webclient | https://api.github.com/repos/venturemark/webclient | closed | Can't send out invite emails | priority/high | Kav was trying to use Venturemark to invite a user and was unable to send out an invite.
***update* It appears that his comment is based on the input not being cleared after inviting user.**

| 1.0 | Can't send out invite emails - Kav was trying to use Venturemark to invite a user and was unable to send out an invite.
***update* It appears that his comment is based on the input not being cleared after inviting user.**

| non_infrastructure | can t send out invite emails kav was trying to use venturemark to invite a user and was unable to send out an invite update it appears that his comment is based on the input not being cleared after inviting user | 0 |
712,633 | 24,501,544,017 | IssuesEvent | 2022-10-10 13:11:18 | AzisabaNetwork/Kuvel | https://api.github.com/repos/AzisabaNetwork/Kuvel | closed | Handling of Duplicate Server Names for Load Balancer | kind/feature priority/high area/discovery area/load-balancer | If Kuvel try to create a load balancer with a server name that is already currently registered, it will just warn you and not be created. It should be possible to set the registration strategy for duplicate names from config.
The planned values currently include
1. Only warning ( current behavior )
2. Hijack the server name if nobody is playing on the server
3. Hijack the server name even if someone is playing on the server | 1.0 | Handling of Duplicate Server Names for Load Balancer - If Kuvel try to create a load balancer with a server name that is already currently registered, it will just warn you and not be created. It should be possible to set the registration strategy for duplicate names from config.
The planned values currently include
1. Only warning ( current behavior )
2. Hijack the server name if nobody is playing on the server
3. Hijack the server name even if someone is playing on the server | non_infrastructure | handling of duplicate server names for load balancer if kuvel try to create a load balancer with a server name that is already currently registered it will just warn you and not be created it should be possible to set the registration strategy for duplicate names from config the planned values currently include only warning current behavior hijack the server name if nobody is playing on the server hijack the server name even if someone is playing on the server | 0 |
93,049 | 11,736,611,912 | IssuesEvent | 2020-03-11 13:22:35 | emergenzeHack/covid19italia_form | https://api.github.com/repos/emergenzeHack/covid19italia_form | opened | Nuovo form unico per segnalazione INIZIATIVE | aiuto necessario backend form design | Qui: https://ee.humanitarianresponse.info/x/#6KafBk33
Mi date un feedback? @Saraveg @cristigalas | 1.0 | Nuovo form unico per segnalazione INIZIATIVE - Qui: https://ee.humanitarianresponse.info/x/#6KafBk33
Mi date un feedback? @Saraveg @cristigalas | non_infrastructure | nuovo form unico per segnalazione iniziative qui mi date un feedback saraveg cristigalas | 0 |
815,004 | 30,533,242,864 | IssuesEvent | 2023-07-19 15:30:57 | Consiglio-Regionale-della-Lombardia/PEM | https://api.github.com/repos/Consiglio-Regionale-della-Lombardia/PEM | closed | Visualizzazione comandi griglia riepilogo atti | view low priority mobile | Riga singola
- Visualizza gli atti per i quali è richiesta la mia firma
Sulla stessa riga:
- Seleziona tutti gli atti
- Espandi tutti gli atti

| 1.0 | Visualizzazione comandi griglia riepilogo atti - Riga singola
- Visualizza gli atti per i quali è richiesta la mia firma
Sulla stessa riga:
- Seleziona tutti gli atti
- Espandi tutti gli atti

| non_infrastructure | visualizzazione comandi griglia riepilogo atti riga singola visualizza gli atti per i quali è richiesta la mia firma sulla stessa riga seleziona tutti gli atti espandi tutti gli atti | 0 |
4,478 | 2,610,094,804 | IssuesEvent | 2015-02-26 18:28:36 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳红蓝光怎么样治疗青春痘 | auto-migrated Priority-Medium Type-Defect | ```
深圳红蓝光怎么样治疗青春痘【深圳韩方科颜全国热线400-869-
1818,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机��
�以韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘�
��品,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“
不反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开��
�国内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾�
��脸上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:18 | 1.0 | 深圳红蓝光怎么样治疗青春痘 - ```
深圳红蓝光怎么样治疗青春痘【深圳韩方科颜全国热线400-869-
1818,24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机��
�以韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘�
��品,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“
不反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开��
�国内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾�
��脸上的痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 8:18 | non_infrastructure | 深圳红蓝光怎么样治疗青春痘 深圳红蓝光怎么样治疗青春痘【 , 】深圳韩方科颜专业祛痘连锁机构,机�� �以韩国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘� ��品,韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“ 不反弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开�� �国内专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾� ��脸上的痘痘。 original issue reported on code google com by szft com on may at | 0 |
690,193 | 23,648,910,181 | IssuesEvent | 2022-08-26 03:21:19 | Kong/gateway-operator | https://api.github.com/repos/Kong/gateway-operator | closed | Data race on setting the Cloud Flare logger singleton from multiple tests | bug priority/high | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
When running `make test.integration` I encountered the followig:
```
==================
WARNING: DATA RACE
Write at 0x000108380f20 by goroutine 703:
github.com/cloudflare/cfssl/log.SetLogger()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/github.com/cloudflare/cfssl@v1.6.1/log/log.go:62 +0x45c
github.com/kong/gateway-operator/controllers.signCertificate()
/Users/USER/code_/gateway-operator/controllers/utils.go:110 +0x28
github.com/kong/gateway-operator/controllers.maybeCreateCertificateSecret()
/Users/USER/code_/gateway-operator/controllers/utils.go:219 +0x9c0
github.com/kong/gateway-operator/controllers.(*DataPlaneReconciler).ensureCertificate()
/Users/USER/code_/gateway-operator/controllers/dataplane_controller_reconciler_utils.go:114 +0x190
github.com/kong/gateway-operator/controllers.(*DataPlaneReconciler).Reconcile()
/Users/USER/code_/gateway-operator/controllers/dataplane_controller.go:120 +0x578
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:121 +0xe4
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:320 +0x360
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:273 +0x240
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:234 +0x98
Previous write at 0x000108380f20 by goroutine 704:
github.com/cloudflare/cfssl/log.SetLogger()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/github.com/cloudflare/cfssl@v1.6.1/log/log.go:62 +0x45c
github.com/kong/gateway-operator/controllers.signCertificate()
/Users/USER/code_/gateway-operator/controllers/utils.go:110 +0x28
github.com/kong/gateway-operator/controllers.maybeCreateCertificateSecret()
/Users/USER/code_/gateway-operator/controllers/utils.go:219 +0x9c0
github.com/kong/gateway-operator/controllers.(*ControlPlaneReconciler).ensureCertificate()
/Users/USER/code_/gateway-operator/controllers/controlplane_controller_reconciler_utils.go:290 +0x19c
github.com/kong/gateway-operator/controllers.(*ControlPlaneReconciler).Reconcile()
/Users/USER/code_/gateway-operator/controllers/controlplane_controller.go:234 +0x1330
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:121 +0xe4
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:320 +0x360
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:273 +0x240
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:234 +0x98
Goroutine 703 (running) created at:
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:230 +0x364
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:241 +0x254
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func1()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:219 +0x158
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:222 +0x44
Goroutine 704 (running) created at:
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:230 +0x364
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:241 +0x254
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func1()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:219 +0x158
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:222 +0x44
==================
```
which points to the following line in code: https://github.com/Kong/gateway-operator/blob/efea71c1b334f4eb5d9be0e845127f54a2600834/controllers/utils.go#L110 which sets a singleton to a designated value in here: https://github.com/cloudflare/cfssl/blob/7614d6cad35dd6d33c8c1fc2e2db5d9ce111e56b/log/log.go#L62
### Expected Behavior
No data race happens.
### Steps To Reproduce
```markdown
1. Run `make test.integration`
2. Wait for data race to occur
```
### Kong Ingress Controller version
```shell
N/A
```
### Kubernetes version
```shell
N/A
```
### Anything else?
_No response_ | 1.0 | Data race on setting the Cloud Flare logger singleton from multiple tests - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
When running `make test.integration` I encountered the followig:
```
==================
WARNING: DATA RACE
Write at 0x000108380f20 by goroutine 703:
github.com/cloudflare/cfssl/log.SetLogger()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/github.com/cloudflare/cfssl@v1.6.1/log/log.go:62 +0x45c
github.com/kong/gateway-operator/controllers.signCertificate()
/Users/USER/code_/gateway-operator/controllers/utils.go:110 +0x28
github.com/kong/gateway-operator/controllers.maybeCreateCertificateSecret()
/Users/USER/code_/gateway-operator/controllers/utils.go:219 +0x9c0
github.com/kong/gateway-operator/controllers.(*DataPlaneReconciler).ensureCertificate()
/Users/USER/code_/gateway-operator/controllers/dataplane_controller_reconciler_utils.go:114 +0x190
github.com/kong/gateway-operator/controllers.(*DataPlaneReconciler).Reconcile()
/Users/USER/code_/gateway-operator/controllers/dataplane_controller.go:120 +0x578
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:121 +0xe4
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:320 +0x360
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:273 +0x240
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:234 +0x98
Previous write at 0x000108380f20 by goroutine 704:
github.com/cloudflare/cfssl/log.SetLogger()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/github.com/cloudflare/cfssl@v1.6.1/log/log.go:62 +0x45c
github.com/kong/gateway-operator/controllers.signCertificate()
/Users/USER/code_/gateway-operator/controllers/utils.go:110 +0x28
github.com/kong/gateway-operator/controllers.maybeCreateCertificateSecret()
/Users/USER/code_/gateway-operator/controllers/utils.go:219 +0x9c0
github.com/kong/gateway-operator/controllers.(*ControlPlaneReconciler).ensureCertificate()
/Users/USER/code_/gateway-operator/controllers/controlplane_controller_reconciler_utils.go:290 +0x19c
github.com/kong/gateway-operator/controllers.(*ControlPlaneReconciler).Reconcile()
/Users/USER/code_/gateway-operator/controllers/controlplane_controller.go:234 +0x1330
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:121 +0xe4
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:320 +0x360
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:273 +0x240
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:234 +0x98
Goroutine 703 (running) created at:
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:230 +0x364
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:241 +0x254
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func1()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:219 +0x158
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:222 +0x44
Goroutine 704 (running) created at:
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:230 +0x364
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/internal/controller/controller.go:241 +0x254
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func1()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:219 +0x158
sigs.k8s.io/controller-runtime/pkg/manager.(*runnableGroup).reconcile.func2()
/Users/USER/.gvm/pkgsets/go1.19/global/pkg/mod/sigs.k8s.io/controller-runtime@v0.12.3/pkg/manager/runnable_group.go:222 +0x44
==================
```
which points to the following line in code: https://github.com/Kong/gateway-operator/blob/efea71c1b334f4eb5d9be0e845127f54a2600834/controllers/utils.go#L110 which sets a singleton to a designated value in here: https://github.com/cloudflare/cfssl/blob/7614d6cad35dd6d33c8c1fc2e2db5d9ce111e56b/log/log.go#L62
### Expected Behavior
No data race happens.
### Steps To Reproduce
```markdown
1. Run `make test.integration`
2. Wait for data race to occur
```
### Kong Ingress Controller version
```shell
N/A
```
### Kubernetes version
```shell
N/A
```
### Anything else?
_No response_ | non_infrastructure | data race on setting the cloud flare logger singleton from multiple tests is there an existing issue for this i have searched the existing issues current behavior when running make test integration i encountered the followig warning data race write at by goroutine github com cloudflare cfssl log setlogger users user gvm pkgsets global pkg mod github com cloudflare cfssl log log go github com kong gateway operator controllers signcertificate users user code gateway operator controllers utils go github com kong gateway operator controllers maybecreatecertificatesecret users user code gateway operator controllers utils go github com kong gateway operator controllers dataplanereconciler ensurecertificate users user code gateway operator controllers dataplane controller reconciler utils go github com kong gateway operator controllers dataplanereconciler reconcile users user code gateway operator controllers dataplane controller go sigs io controller runtime pkg internal controller controller reconcile users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller reconcilehandler users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller processnextworkitem users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller start users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go previous write at by goroutine github com cloudflare cfssl log setlogger users user gvm pkgsets global pkg mod github com cloudflare cfssl log log go github com kong gateway operator controllers signcertificate users user code gateway operator controllers utils go github com kong gateway operator controllers maybecreatecertificatesecret users user code gateway operator controllers utils go github com kong gateway operator controllers controlplanereconciler ensurecertificate users user code gateway operator controllers controlplane controller reconciler utils go github com kong gateway operator controllers controlplanereconciler reconcile users user code gateway operator controllers controlplane controller go sigs io controller runtime pkg internal controller controller reconcile users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller reconcilehandler users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller processnextworkitem users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller start users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go goroutine running created at sigs io controller runtime pkg internal controller controller start users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller start users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg manager runnablegroup reconcile users user gvm pkgsets global pkg mod sigs io controller runtime pkg manager runnable group go sigs io controller runtime pkg manager runnablegroup reconcile users user gvm pkgsets global pkg mod sigs io controller runtime pkg manager runnable group go goroutine running created at sigs io controller runtime pkg internal controller controller start users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg internal controller controller start users user gvm pkgsets global pkg mod sigs io controller runtime pkg internal controller controller go sigs io controller runtime pkg manager runnablegroup reconcile users user gvm pkgsets global pkg mod sigs io controller runtime pkg manager runnable group go sigs io controller runtime pkg manager runnablegroup reconcile users user gvm pkgsets global pkg mod sigs io controller runtime pkg manager runnable group go which points to the following line in code which sets a singleton to a designated value in here expected behavior no data race happens steps to reproduce markdown run make test integration wait for data race to occur kong ingress controller version shell n a kubernetes version shell n a anything else no response | 0 |
19,267 | 13,211,305,516 | IssuesEvent | 2020-08-15 22:11:04 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | pybindings - don't auto-build dependencies (╯°□°)╯︵ ┻━┻ (Trac #1032) | Incomplete Migration Migrated from Trac enhancement infrastructure | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1032">https://code.icecube.wisc.edu/projects/icecube/ticket/1032</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:01:19",
"_ts": "1547251279109761",
"description": "this is fine with `make` or `make all` but when trying to do a partial/incremental build, you have to keep failing python scripts to see what pybindings you need (without looking into CMakeLists.txt)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2015-06-25T19:00:21",
"component": "infrastructure",
"summary": "pybindings - don't auto-build dependencies (\u256f\u00b0\u25a1\u00b0)\u256f\ufe35 \u253b\u2501\u253b",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "enhancement"
}
```
</p>
</details>
| 1.0 | pybindings - don't auto-build dependencies (╯°□°)╯︵ ┻━┻ (Trac #1032) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1032">https://code.icecube.wisc.edu/projects/icecube/ticket/1032</a>, reported by negaand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-01-12T00:01:19",
"_ts": "1547251279109761",
"description": "this is fine with `make` or `make all` but when trying to do a partial/incremental build, you have to keep failing python scripts to see what pybindings you need (without looking into CMakeLists.txt)",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2015-06-25T19:00:21",
"component": "infrastructure",
"summary": "pybindings - don't auto-build dependencies (\u256f\u00b0\u25a1\u00b0)\u256f\ufe35 \u253b\u2501\u253b",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "nega",
"type": "enhancement"
}
```
</p>
</details>
| infrastructure | pybindings don t auto build dependencies ╯°□° ╯︵ ┻━┻ trac migrated from json status closed changetime ts description this is fine with make or make all but when trying to do a partial incremental build you have to keep failing python scripts to see what pybindings you need without looking into cmakelists txt reporter nega cc resolution fixed time component infrastructure summary pybindings don t auto build dependencies priority normal keywords milestone owner nega type enhancement | 1 |
27,814 | 4,330,371,486 | IssuesEvent | 2016-07-26 19:47:55 | F5Networks/f5-common-python | https://api.github.com/repos/F5Networks/f5-common-python | closed | During functional testing in 11.5.4 device goes to forced offline and does not come back online | functional test | The tests live here, and the likely culprit is: https://github.com/F5Networks/f5-common-python/blob/60abd57feb5b457e441f6d940430a50458866762/test/functional/tm/sys/test_failover.py#L75
Maybe this is just as simple as having an addfinalizer at the end of the test to force it back online in case the test fails. | 1.0 | During functional testing in 11.5.4 device goes to forced offline and does not come back online - The tests live here, and the likely culprit is: https://github.com/F5Networks/f5-common-python/blob/60abd57feb5b457e441f6d940430a50458866762/test/functional/tm/sys/test_failover.py#L75
Maybe this is just as simple as having an addfinalizer at the end of the test to force it back online in case the test fails. | non_infrastructure | during functional testing in device goes to forced offline and does not come back online the tests live here and the likely culprit is maybe this is just as simple as having an addfinalizer at the end of the test to force it back online in case the test fails | 0 |
6,618 | 6,534,978,529 | IssuesEvent | 2017-08-31 13:05:14 | openshiftio/appdev-documentation | https://api.github.com/repos/openshiftio/appdev-documentation | opened | Deduplicate cico_build_deploy.sh and cico_build_test.sh | Component | Infrastructure Type | Enhancement | Due to #504, there is a new script called `cico_build_test.sh` in the repo root. This script is partially identical to `cico_build_deploy`, so it warrants refactoring to eliminate the code duplication. | 1.0 | Deduplicate cico_build_deploy.sh and cico_build_test.sh - Due to #504, there is a new script called `cico_build_test.sh` in the repo root. This script is partially identical to `cico_build_deploy`, so it warrants refactoring to eliminate the code duplication. | infrastructure | deduplicate cico build deploy sh and cico build test sh due to there is a new script called cico build test sh in the repo root this script is partially identical to cico build deploy so it warrants refactoring to eliminate the code duplication | 1 |
229,001 | 18,275,752,624 | IssuesEvent | 2021-10-04 18:36:20 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: sequelize failed | C-test-failure O-robot O-roachtest branch-master | roachtest.sequelize [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3524635&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3524635&tab=artifacts#/sequelize) on master @ [5f53feb1ce7070453e9ce52d99013934718cc9d7](https://github.com/cockroachdb/cockroach/commits/5f53feb1ce7070453e9ce52d99013934718cc9d7):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/sequelize/run_1
sequelize.go:101,sequelize.go:157,test_runner.go:777: all attempts failed for add nodesource repository due to error: output in run_122206.270795111_n1_curl: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-3524635-1633069195-81-n1cpu4:1 -- curl -sL https://deb.nodesource.com/setup_12.x | sudo -E bash - returned: exit status 20
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #70981 roachtest: sequelize failed [C-test-failure O-roachtest O-robot branch-release-21.2 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sequelize.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: sequelize failed - roachtest.sequelize [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3524635&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3524635&tab=artifacts#/sequelize) on master @ [5f53feb1ce7070453e9ce52d99013934718cc9d7](https://github.com/cockroachdb/cockroach/commits/5f53feb1ce7070453e9ce52d99013934718cc9d7):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/sequelize/run_1
sequelize.go:101,sequelize.go:157,test_runner.go:777: all attempts failed for add nodesource repository due to error: output in run_122206.270795111_n1_curl: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-3524635-1633069195-81-n1cpu4:1 -- curl -sL https://deb.nodesource.com/setup_12.x | sudo -E bash - returned: exit status 20
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #70981 roachtest: sequelize failed [C-test-failure O-roachtest O-robot branch-release-21.2 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sequelize.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_infrastructure | roachtest sequelize failed roachtest sequelize with on master the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts sequelize run sequelize go sequelize go test runner go all attempts failed for add nodesource repository due to error output in run curl home agent work go src github com cockroachdb cockroach bin roachprod run teamcity curl sl sudo e bash returned exit status reproduce see same failure on other branches roachtest sequelize failed cc cockroachdb sql experience | 0 |
92,553 | 11,682,359,033 | IssuesEvent | 2020-03-05 00:01:35 | quicwg/base-drafts | https://api.github.com/repos/quicwg/base-drafts | closed | Allow the Transport to Stop/Reset a Stream? | -transport design has-consensus | Per the current design for streams, the application MUST be the one to shutdown a stream, because the error code space for streams is purely application layer; and the QUIC transport has no knowledge of these error code, so it can't just pick one.
When writing a general purpose QUIC library, and integrating it into other general purpose libraries ([i.e. dotnet](https://github.com/dotnet/runtime/pull/427)) I constantly have to explain to non-QUIC folks that they can't just close their stream object/handle without first supplying an application layer specific error code; meaning any intermediate library needs input from the app even if some fatal error (memory allocation failure) happened along the way.
So, I've come to the point that it might be a good idea to allow for the transport layer to completely terminate a stream for it's own reason. Whether that means we add a flag to RESET_STREAM and STOP_SENDING to indicate the transport is specifying the error code, or we add a new frame entirely doesn't matter to me. It would just be a lot cleaner for general purpose libraries to be able to kill a stream when necessary, on its own (no app layer involvement). | 1.0 | Allow the Transport to Stop/Reset a Stream? - Per the current design for streams, the application MUST be the one to shutdown a stream, because the error code space for streams is purely application layer; and the QUIC transport has no knowledge of these error code, so it can't just pick one.
When writing a general purpose QUIC library, and integrating it into other general purpose libraries ([i.e. dotnet](https://github.com/dotnet/runtime/pull/427)) I constantly have to explain to non-QUIC folks that they can't just close their stream object/handle without first supplying an application layer specific error code; meaning any intermediate library needs input from the app even if some fatal error (memory allocation failure) happened along the way.
So, I've come to the point that it might be a good idea to allow for the transport layer to completely terminate a stream for it's own reason. Whether that means we add a flag to RESET_STREAM and STOP_SENDING to indicate the transport is specifying the error code, or we add a new frame entirely doesn't matter to me. It would just be a lot cleaner for general purpose libraries to be able to kill a stream when necessary, on its own (no app layer involvement). | non_infrastructure | allow the transport to stop reset a stream per the current design for streams the application must be the one to shutdown a stream because the error code space for streams is purely application layer and the quic transport has no knowledge of these error code so it can t just pick one when writing a general purpose quic library and integrating it into other general purpose libraries i constantly have to explain to non quic folks that they can t just close their stream object handle without first supplying an application layer specific error code meaning any intermediate library needs input from the app even if some fatal error memory allocation failure happened along the way so i ve come to the point that it might be a good idea to allow for the transport layer to completely terminate a stream for it s own reason whether that means we add a flag to reset stream and stop sending to indicate the transport is specifying the error code or we add a new frame entirely doesn t matter to me it would just be a lot cleaner for general purpose libraries to be able to kill a stream when necessary on its own no app layer involvement | 0 |
317,786 | 9,669,276,411 | IssuesEvent | 2019-05-21 16:57:47 | etternagame/etterna | https://api.github.com/repos/etternagame/etterna | closed | Old files with only .dwi sometimes use wrong image for song banner | Good First Issue Priority: Low Type: Bug | Some old files that only have a .dwi and not .sm or .ssc don't specify the file path to the song banner. It seems like Etterna tries to fall back by grabbing some other image from the folder, but it doesn't always grab the correct one. I think the desired behavior should be to first try and fall back to another image in the song folder that has the word "banner" in the file name, as I think specifying which image was which using the file name used to be the common practice.
Example: Red Zone in KBC3. The .ogg itself has some associated artwork (album cover?), and Etterna uses this image as the banner rather than the actual banner, which has "banner" in the file name. | 1.0 | Old files with only .dwi sometimes use wrong image for song banner - Some old files that only have a .dwi and not .sm or .ssc don't specify the file path to the song banner. It seems like Etterna tries to fall back by grabbing some other image from the folder, but it doesn't always grab the correct one. I think the desired behavior should be to first try and fall back to another image in the song folder that has the word "banner" in the file name, as I think specifying which image was which using the file name used to be the common practice.
Example: Red Zone in KBC3. The .ogg itself has some associated artwork (album cover?), and Etterna uses this image as the banner rather than the actual banner, which has "banner" in the file name. | non_infrastructure | old files with only dwi sometimes use wrong image for song banner some old files that only have a dwi and not sm or ssc don t specify the file path to the song banner it seems like etterna tries to fall back by grabbing some other image from the folder but it doesn t always grab the correct one i think the desired behavior should be to first try and fall back to another image in the song folder that has the word banner in the file name as i think specifying which image was which using the file name used to be the common practice example red zone in the ogg itself has some associated artwork album cover and etterna uses this image as the banner rather than the actual banner which has banner in the file name | 0 |
192,163 | 14,601,561,112 | IssuesEvent | 2020-12-21 08:53:05 | hbvhuwe/easy-travel | https://api.github.com/repos/hbvhuwe/easy-travel | closed | Add tests for Travel planning API | project:api test | Create a file `/tests/routes/travel.test.js` for testing Travel planning API.
Create a file `/tests/services/travel.test.js` for testing Travel planning API service.
Can start implementing, when the #95 pull request will be closed and merged. | 1.0 | Add tests for Travel planning API - Create a file `/tests/routes/travel.test.js` for testing Travel planning API.
Create a file `/tests/services/travel.test.js` for testing Travel planning API service.
Can start implementing, when the #95 pull request will be closed and merged. | non_infrastructure | add tests for travel planning api create a file tests routes travel test js for testing travel planning api create a file tests services travel test js for testing travel planning api service can start implementing when the pull request will be closed and merged | 0 |
12,082 | 9,582,704,396 | IssuesEvent | 2019-05-08 01:55:53 | nervosnetwork/ckb | https://api.github.com/repos/nervosnetwork/ckb | closed | Config file check and friendly error message | m:infrastructure p:should-have s:available t:enhancement | Before booting the node, check the config file first:
- Report any error with a user-friendly message.
- ~~Verify that all path points to existing files.~~ #387 | 1.0 | Config file check and friendly error message - Before booting the node, check the config file first:
- Report any error with a user-friendly message.
- ~~Verify that all path points to existing files.~~ #387 | infrastructure | config file check and friendly error message before booting the node check the config file first report any error with a user friendly message verify that all path points to existing files | 1 |
143,338 | 22,032,960,064 | IssuesEvent | 2022-05-28 05:48:24 | stores-cedcommerce/Internal---Liz-Store-dev.---31-May-2022 | https://api.github.com/repos/stores-cedcommerce/Internal---Liz-Store-dev.---31-May-2022 | closed | Slide navigation buttons getting disappear on clicking in mobile UI | Ready to test Home page content Design / UI / UX Mobile | Bug - Slide navigation buttons getting disappear on clicking in mobile UI
Exp - slide navigation button should not get disappear on clicking.
Ref Link - https://drive.google.com/file/d/1uEBSututgfn9wpAFZkUoYB0OevucOB1e/view | 1.0 | Slide navigation buttons getting disappear on clicking in mobile UI - Bug - Slide navigation buttons getting disappear on clicking in mobile UI
Exp - slide navigation button should not get disappear on clicking.
Ref Link - https://drive.google.com/file/d/1uEBSututgfn9wpAFZkUoYB0OevucOB1e/view | non_infrastructure | slide navigation buttons getting disappear on clicking in mobile ui bug slide navigation buttons getting disappear on clicking in mobile ui exp slide navigation button should not get disappear on clicking ref link | 0 |
6,590 | 6,525,653,887 | IssuesEvent | 2017-08-29 16:38:32 | eclipse/smarthome | https://api.github.com/repos/eclipse/smarthome | closed | Version bumps / releases | Infrastructure | migrated from Bugzilla [#478597](https://bugs.eclipse.org/bugs/show_bug.cgi?id=478597)
status UNCONFIRMED severity _normal_ in component _Infrastructure_ for _---_
Reported in version _unspecified_ on platform _All_
Assigned to: Project Inbox
On 2015-09-29 04:06:10 -0400, Markus Rathgeb wrote:
> Eclipse SmartHome is a framework to build end user solutions on top.
>
> If an "end user" would like to use that framework an do a release of his software, the software should not depend on a SNAPSHOT at all.
>
> The ESH framework themselves is not bound (AFAIK) to any consumer release cycles and this is IMHO fine.
>
> But I think it would be fine if we could produce releases from time to time so it could used as a reference for custom products.
> I cannot depend on 0.8-SNAPSHOT all the time.
>
> I do not mean a 1.0 stable framework release without API breaks anymore...
> I just want to use a fixed reference in my sources.
>
> Using custom repositories are possible, but this mean for every reference a new repository. That is not good.
>
> If the process for a release is a long task for Eclipse products, I will ask, if we could bump the snapshot version from time to time without a official release.
> So building a release could build by the consumers.
>
> I cannot build a 0.8.0, 0.8.1, ... release myself if Eclipse releases an official 0.8[.0] some time.
On 2015-09-29 04:39:35 -0400, Kai Kreuzer wrote:
> I agree that fixed releases would definitely help to have a reliable way to reference a specific code base from within a solution.
>
> Note that I am currently preparing the 0.8.0 release, see https://projects.eclipse.org/projects/iot.smarthome/releases/0.8.0/
>
> But creating such minor releases is some work; I would not expect more than 2-3 each year.
>
> Nonetheless, we could maybe check if we do more regular service releases, which should be a bit simpler (but still means work).
> We will have to check how much this can be automated (update version in sources, have documentation generated and published, adapt the build plans and download locations etc.)
On 2015-09-29 04:55:04 -0400, Kai Kreuzer wrote:
> Probably a better option than regular service releases might be to create milestones, e.g. 0.9.0M1, 0.9.0M2 etc. This seems to be possible without following the release process and it I think we also would not need to publish documentation on the website for such milestones.
On 2015-09-29 07:38:43 -0400, Markus Rathgeb wrote:
> > Probably a better option than regular service releases might be to create
> > milestones, e.g. 0.9.0M1, 0.9.0M2 etc. This seems to be possible without
> > following the release process and it I think we also would not need to
> > publish documentation on the website for such milestones.
>
> That would be enough. I just want fixed states we could point to. If this are alphas, betas, release candidates, milestones, releases, ... does not matter for me.
> As long as it is fixed for a specific code base.
>
> If the qualifier (M) is chosen correctly, Maven should handle this correctly, too.
> https://docs.oracle.com/middleware/1212/core/MAVEN/maven_version.htm#MAVEN400
On 2015-11-03 12:50:36 -0500, Markus Rathgeb wrote:
> I started doing it for myself:
> https://github.com/maggu2810/smarthome/commit/SHA: 04f8c9477fb896f77718a377f972f7f91876c639
>
> But is there something we can do to get a solution upstream?
| 1.0 | Version bumps / releases - migrated from Bugzilla [#478597](https://bugs.eclipse.org/bugs/show_bug.cgi?id=478597)
status UNCONFIRMED severity _normal_ in component _Infrastructure_ for _---_
Reported in version _unspecified_ on platform _All_
Assigned to: Project Inbox
On 2015-09-29 04:06:10 -0400, Markus Rathgeb wrote:
> Eclipse SmartHome is a framework to build end user solutions on top.
>
> If an "end user" would like to use that framework an do a release of his software, the software should not depend on a SNAPSHOT at all.
>
> The ESH framework themselves is not bound (AFAIK) to any consumer release cycles and this is IMHO fine.
>
> But I think it would be fine if we could produce releases from time to time so it could used as a reference for custom products.
> I cannot depend on 0.8-SNAPSHOT all the time.
>
> I do not mean a 1.0 stable framework release without API breaks anymore...
> I just want to use a fixed reference in my sources.
>
> Using custom repositories are possible, but this mean for every reference a new repository. That is not good.
>
> If the process for a release is a long task for Eclipse products, I will ask, if we could bump the snapshot version from time to time without a official release.
> So building a release could build by the consumers.
>
> I cannot build a 0.8.0, 0.8.1, ... release myself if Eclipse releases an official 0.8[.0] some time.
On 2015-09-29 04:39:35 -0400, Kai Kreuzer wrote:
> I agree that fixed releases would definitely help to have a reliable way to reference a specific code base from within a solution.
>
> Note that I am currently preparing the 0.8.0 release, see https://projects.eclipse.org/projects/iot.smarthome/releases/0.8.0/
>
> But creating such minor releases is some work; I would not expect more than 2-3 each year.
>
> Nonetheless, we could maybe check if we do more regular service releases, which should be a bit simpler (but still means work).
> We will have to check how much this can be automated (update version in sources, have documentation generated and published, adapt the build plans and download locations etc.)
On 2015-09-29 04:55:04 -0400, Kai Kreuzer wrote:
> Probably a better option than regular service releases might be to create milestones, e.g. 0.9.0M1, 0.9.0M2 etc. This seems to be possible without following the release process and it I think we also would not need to publish documentation on the website for such milestones.
On 2015-09-29 07:38:43 -0400, Markus Rathgeb wrote:
> > Probably a better option than regular service releases might be to create
> > milestones, e.g. 0.9.0M1, 0.9.0M2 etc. This seems to be possible without
> > following the release process and it I think we also would not need to
> > publish documentation on the website for such milestones.
>
> That would be enough. I just want fixed states we could point to. If this are alphas, betas, release candidates, milestones, releases, ... does not matter for me.
> As long as it is fixed for a specific code base.
>
> If the qualifier (M) is chosen correctly, Maven should handle this correctly, too.
> https://docs.oracle.com/middleware/1212/core/MAVEN/maven_version.htm#MAVEN400
On 2015-11-03 12:50:36 -0500, Markus Rathgeb wrote:
> I started doing it for myself:
> https://github.com/maggu2810/smarthome/commit/SHA: 04f8c9477fb896f77718a377f972f7f91876c639
>
> But is there something we can do to get a solution upstream?
| infrastructure | version bumps releases migrated from bugzilla status unconfirmed severity normal in component infrastructure for reported in version unspecified on platform all assigned to project inbox on markus rathgeb wrote eclipse smarthome is a framework to build end user solutions on top if an end user would like to use that framework an do a release of his software the software should not depend on a snapshot at all the esh framework themselves is not bound afaik to any consumer release cycles and this is imho fine but i think it would be fine if we could produce releases from time to time so it could used as a reference for custom products i cannot depend on snapshot all the time i do not mean a stable framework release without api breaks anymore i just want to use a fixed reference in my sources using custom repositories are possible but this mean for every reference a new repository that is not good if the process for a release is a long task for eclipse products i will ask if we could bump the snapshot version from time to time without a official release so building a release could build by the consumers i cannot build a release myself if eclipse releases an official some time on kai kreuzer wrote i agree that fixed releases would definitely help to have a reliable way to reference a specific code base from within a solution note that i am currently preparing the release see but creating such minor releases is some work i would not expect more than each year nonetheless we could maybe check if we do more regular service releases which should be a bit simpler but still means work we will have to check how much this can be automated update version in sources have documentation generated and published adapt the build plans and download locations etc on kai kreuzer wrote probably a better option than regular service releases might be to create milestones e g etc this seems to be possible without following the release process and it i think we also would not need to publish documentation on the website for such milestones on markus rathgeb wrote probably a better option than regular service releases might be to create milestones e g etc this seems to be possible without following the release process and it i think we also would not need to publish documentation on the website for such milestones that would be enough i just want fixed states we could point to if this are alphas betas release candidates milestones releases does not matter for me as long as it is fixed for a specific code base if the qualifier m is chosen correctly maven should handle this correctly too on markus rathgeb wrote i started doing it for myself but is there something we can do to get a solution upstream | 1 |
3,705 | 4,494,543,388 | IssuesEvent | 2016-08-31 06:44:30 | gammapy/gammapy | https://api.github.com/repos/gammapy/gammapy | opened | Add Sherpa Python 3 build on travis-ci | infrastructure tests | Since yesterday, Sherpa master supports Python 3.
- [ ] Add Sherpa Python 3 build that builds sherpa from source on travis-ci
- [ ] Change the main build that measures coverage to Python 3.5 on travis-ci
- [ ] Fix this test fail in `gammapy/scripts/tests/test_image_fit.py`: [gist](https://gist.github.com/cdeil/147d8c7ace94bceba9eb39d20d9b69d7)
- [ ] Fix this test fail in `gammapy/spectrum/tests/test_fit.py`: [gist](https://gist.github.com/cdeil/a96000e2471c40cedf89532535d38eab)
- [ ] Minor install docs and changelog update to mention that Sherpa supports Python 3
@joleroi - Any thoughts about the `gammapy/spectrum/tests/test_fit.py` fail? | 1.0 | Add Sherpa Python 3 build on travis-ci - Since yesterday, Sherpa master supports Python 3.
- [ ] Add Sherpa Python 3 build that builds sherpa from source on travis-ci
- [ ] Change the main build that measures coverage to Python 3.5 on travis-ci
- [ ] Fix this test fail in `gammapy/scripts/tests/test_image_fit.py`: [gist](https://gist.github.com/cdeil/147d8c7ace94bceba9eb39d20d9b69d7)
- [ ] Fix this test fail in `gammapy/spectrum/tests/test_fit.py`: [gist](https://gist.github.com/cdeil/a96000e2471c40cedf89532535d38eab)
- [ ] Minor install docs and changelog update to mention that Sherpa supports Python 3
@joleroi - Any thoughts about the `gammapy/spectrum/tests/test_fit.py` fail? | infrastructure | add sherpa python build on travis ci since yesterday sherpa master supports python add sherpa python build that builds sherpa from source on travis ci change the main build that measures coverage to python on travis ci fix this test fail in gammapy scripts tests test image fit py fix this test fail in gammapy spectrum tests test fit py minor install docs and changelog update to mention that sherpa supports python joleroi any thoughts about the gammapy spectrum tests test fit py fail | 1 |
31,987 | 26,333,711,645 | IssuesEvent | 2023-01-10 12:50:25 | celeritas-project/celeritas | https://api.github.com/repos/celeritas-project/celeritas | closed | Reformat for 0.2.0 | infrastructure housekeeping | - Don't align on assignment/initialization (because it ends up wreaking havoc with version control and is not very stable from one clang-format version to the next)
- [East const](https://mariusbancila.ro/blog/2018/11/23/join-the-east-const-revolution/)
- Apply IWYU (see #540, #541) | 1.0 | Reformat for 0.2.0 - - Don't align on assignment/initialization (because it ends up wreaking havoc with version control and is not very stable from one clang-format version to the next)
- [East const](https://mariusbancila.ro/blog/2018/11/23/join-the-east-const-revolution/)
- Apply IWYU (see #540, #541) | infrastructure | reformat for don t align on assignment initialization because it ends up wreaking havoc with version control and is not very stable from one clang format version to the next apply iwyu see | 1 |
1,767 | 2,571,886,071 | IssuesEvent | 2015-02-10 19:08:50 | mapbox/mapbox-gl-cocoa | https://api.github.com/repos/mapbox/mapbox-gl-cocoa | closed | find way to test rotation gestures | testing | Aside from making sure that they and `mapView.rotateEnabled` both work, we can prevent bugs like https://github.com/mapbox/mapbox-gl-cocoa/commit/42983b45190f853b19b430cc89b5c64368ce1d50.
Related: https://github.com/kif-framework/KIF/pull/410 | 1.0 | find way to test rotation gestures - Aside from making sure that they and `mapView.rotateEnabled` both work, we can prevent bugs like https://github.com/mapbox/mapbox-gl-cocoa/commit/42983b45190f853b19b430cc89b5c64368ce1d50.
Related: https://github.com/kif-framework/KIF/pull/410 | non_infrastructure | find way to test rotation gestures aside from making sure that they and mapview rotateenabled both work we can prevent bugs like related | 0 |
299,918 | 9,205,970,030 | IssuesEvent | 2019-03-08 12:16:45 | qissue-bot/QGIS | https://api.github.com/repos/qissue-bot/QGIS | closed | PostGIS: missing buttons to delete and add new columns in table editor | Category: GUI Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report | ---
Author Name: **Maciej Sieczka -** (Maciej Sieczka -)
Original Redmine Issue: 1273, https://issues.qgis.org/issues/1273
Original Assignee: nobody -
---
Why are the buttons to delete and add new columns for [[PostGIS]] layers missing in the table editor? I can bet they were there few revisions ago.
| 1.0 | PostGIS: missing buttons to delete and add new columns in table editor - ---
Author Name: **Maciej Sieczka -** (Maciej Sieczka -)
Original Redmine Issue: 1273, https://issues.qgis.org/issues/1273
Original Assignee: nobody -
---
Why are the buttons to delete and add new columns for [[PostGIS]] layers missing in the table editor? I can bet they were there few revisions ago.
| non_infrastructure | postgis missing buttons to delete and add new columns in table editor author name maciej sieczka maciej sieczka original redmine issue original assignee nobody why are the buttons to delete and add new columns for layers missing in the table editor i can bet they were there few revisions ago | 0 |
316,105 | 27,137,096,051 | IssuesEvent | 2023-02-16 13:59:46 | wazuh/wazuh-qa | https://api.github.com/repos/wazuh/wazuh-qa | opened | Add coverage to `group_sync_status` updated correctly | team/qa test/system type/test-development status/not-tracked subteam/qa-main | | Target version | Related issue | Related PR |
|--------------------|--------------------|-----------------|
| 4.4.0 | [#16188](https://github.com/wazuh/wazuh/issues/16188) [#16190](https://github.com/wazuh/wazuh/issues/16190) [#16196](https://github.com/wazuh/wazuh/issues/16196) | [#16202](https://github.com/wazuh/wazuh/pull/16202) |
<!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). -->
## Description
<!-- Description that puts into context and shows the QA tester the changes that have been implemented and have to be tested. -->
**System** tests should be developed to validate the `group_sync_status` of a big amount of `wazuh-agents` on every cluster nodes is updated correctly on each agent.
## Proposed test cases
<!-- Indicate the minimum test cases proposed by the developer. -->
<details><summary>Scenario Outline 1: Group sync status - syncreq status marked correctly</summary>
- **Given** a `wazuh-manager` and 50k `wazuh-agent` connected.
- **And** all `wazuh-agents` group sync status is `synreq`.
- **When** the status of the first chunk of agents is updated to `synced`.
- **Then** only the agents from the first chunk are **`synced`** the rest are still **`synreq`**.
</details>
| 2.0 | Add coverage to `group_sync_status` updated correctly - | Target version | Related issue | Related PR |
|--------------------|--------------------|-----------------|
| 4.4.0 | [#16188](https://github.com/wazuh/wazuh/issues/16188) [#16190](https://github.com/wazuh/wazuh/issues/16190) [#16196](https://github.com/wazuh/wazuh/issues/16196) | [#16202](https://github.com/wazuh/wazuh/pull/16202) |
<!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). -->
## Description
<!-- Description that puts into context and shows the QA tester the changes that have been implemented and have to be tested. -->
**System** tests should be developed to validate the `group_sync_status` of a big amount of `wazuh-agents` on every cluster nodes is updated correctly on each agent.
## Proposed test cases
<!-- Indicate the minimum test cases proposed by the developer. -->
<details><summary>Scenario Outline 1: Group sync status - syncreq status marked correctly</summary>
- **Given** a `wazuh-manager` and 50k `wazuh-agent` connected.
- **And** all `wazuh-agents` group sync status is `synreq`.
- **When** the status of the first chunk of agents is updated to `synced`.
- **Then** only the agents from the first chunk are **`synced`** the rest are still **`synreq`**.
</details>
| non_infrastructure | add coverage to group sync status updated correctly target version related issue related pr description system tests should be developed to validate the group sync status of a big amount of wazuh agents on every cluster nodes is updated correctly on each agent proposed test cases scenario outline group sync status syncreq status marked correctly given a wazuh manager and wazuh agent connected and all wazuh agents group sync status is synreq when the status of the first chunk of agents is updated to synced then only the agents from the first chunk are synced the rest are still synreq | 0 |
24,950 | 17,936,108,790 | IssuesEvent | 2021-09-10 15:30:50 | coq/coq | https://api.github.com/repos/coq/coq | closed | Add timing information to Windows packaging jobs. | kind: infrastructure platform: Windows | Our windows packaging jobs are fairly slow, however profiling them seems hard due to missing time information in the logs. Please, add such information. | 1.0 | Add timing information to Windows packaging jobs. - Our windows packaging jobs are fairly slow, however profiling them seems hard due to missing time information in the logs. Please, add such information. | infrastructure | add timing information to windows packaging jobs our windows packaging jobs are fairly slow however profiling them seems hard due to missing time information in the logs please add such information | 1 |
6,333 | 6,336,991,064 | IssuesEvent | 2017-07-26 22:29:39 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Adding Amazon Linux to the Runtime IDs | area-Infrastructure enhancement | I'd like to add Amazon Linux to list of Runtime IDs (RID). What are the steps to achieve this?
1. Does adding new Amazon Linux RIDs in [runtime.json](https://github.com/dotnet/corefx/blob/master/pkg/Microsoft.NETCore.Platforms/runtime.json) suffice? Are there any other steps involved?
2. The logic to create the RID string for a given Linux distro resides here in corehost [pal::get_current_os_rid_platform](https://github.com/dotnet/core-setup/blob/master/src/corehost/common/pal.unix.cpp#L295) , is that correct?
@Petermarcu, this is based on the discussion you and @normj had at NDC Oslo. | 1.0 | Adding Amazon Linux to the Runtime IDs - I'd like to add Amazon Linux to list of Runtime IDs (RID). What are the steps to achieve this?
1. Does adding new Amazon Linux RIDs in [runtime.json](https://github.com/dotnet/corefx/blob/master/pkg/Microsoft.NETCore.Platforms/runtime.json) suffice? Are there any other steps involved?
2. The logic to create the RID string for a given Linux distro resides here in corehost [pal::get_current_os_rid_platform](https://github.com/dotnet/core-setup/blob/master/src/corehost/common/pal.unix.cpp#L295) , is that correct?
@Petermarcu, this is based on the discussion you and @normj had at NDC Oslo. | infrastructure | adding amazon linux to the runtime ids i d like to add amazon linux to list of runtime ids rid what are the steps to achieve this does adding new amazon linux rids in suffice are there any other steps involved the logic to create the rid string for a given linux distro resides here in corehost is that correct petermarcu this is based on the discussion you and normj had at ndc oslo | 1 |
22,781 | 15,439,091,497 | IssuesEvent | 2021-03-07 22:52:40 | niconoe/pyinaturalist | https://api.github.com/repos/niconoe/pyinaturalist | closed | Build and deploy inaturalist-notebook Docker image for new pyinaturalist releases | docs examples infrastructure | Image repo: https://github.com/JWCook/inaturalist-notebook
I believe this can be triggered using the [repository_dispatch](https://github.com/peter-evans/repository-dispatch) event. | 1.0 | Build and deploy inaturalist-notebook Docker image for new pyinaturalist releases - Image repo: https://github.com/JWCook/inaturalist-notebook
I believe this can be triggered using the [repository_dispatch](https://github.com/peter-evans/repository-dispatch) event. | infrastructure | build and deploy inaturalist notebook docker image for new pyinaturalist releases image repo i believe this can be triggered using the event | 1 |
10,451 | 8,570,151,497 | IssuesEvent | 2018-11-11 17:33:11 | telepresenceio/telepresence | https://api.github.com/repos/telepresenceio/telepresence | closed | Telepresence should tell the user when a new version is available | enhancement infrastructure | Now that we have scout integration, Telepresence can tell the user about a new version being available.
First, as part of the release process, we must do something like this:
```bash
cat << EOF > app.json
{
"application": "telepresence",
"latest_version": "${TELEPRESENCE_VERSION}",
"notices": []
}
EOF
export AWS_DEFAULT_REGION=us-east-1
aws s3api put-object \
--bucket scout-datawire-io \
--key telepresence/app.json \
--body app.json
```
Then we can use the info scout passes back, specifically the `latest_version` key, to determine whether a new version is available and what it is.
| 1.0 | Telepresence should tell the user when a new version is available - Now that we have scout integration, Telepresence can tell the user about a new version being available.
First, as part of the release process, we must do something like this:
```bash
cat << EOF > app.json
{
"application": "telepresence",
"latest_version": "${TELEPRESENCE_VERSION}",
"notices": []
}
EOF
export AWS_DEFAULT_REGION=us-east-1
aws s3api put-object \
--bucket scout-datawire-io \
--key telepresence/app.json \
--body app.json
```
Then we can use the info scout passes back, specifically the `latest_version` key, to determine whether a new version is available and what it is.
| infrastructure | telepresence should tell the user when a new version is available now that we have scout integration telepresence can tell the user about a new version being available first as part of the release process we must do something like this bash cat app json application telepresence latest version telepresence version notices eof export aws default region us east aws put object bucket scout datawire io key telepresence app json body app json then we can use the info scout passes back specifically the latest version key to determine whether a new version is available and what it is | 1 |
11,880 | 14,006,292,290 | IssuesEvent | 2020-10-28 19:43:45 | oilshell/oil | https://api.github.com/repos/oilshell/oil | closed | errexit not disabled where it should be | compatibility divergence osh-language pending-release |
The causes of the test cases that ShellSpec still fails are 1 and 4. Other issues were resolved by fixing ShellSpec.
### 1. Problem with errexit
osh terminate where it must not.
```sh
set +e
foo() { set -e; return 123; }
foo &&:
echo "should be executed"
```
```sh
set +e
foo() { set -e; false; echo "should be executed"; }
foo &&:
echo "should be executed"
```
| True | errexit not disabled where it should be -
The causes of the test cases that ShellSpec still fails are 1 and 4. Other issues were resolved by fixing ShellSpec.
### 1. Problem with errexit
osh terminate where it must not.
```sh
set +e
foo() { set -e; return 123; }
foo &&:
echo "should be executed"
```
```sh
set +e
foo() { set -e; false; echo "should be executed"; }
foo &&:
echo "should be executed"
```
| non_infrastructure | errexit not disabled where it should be the causes of the test cases that shellspec still fails are and other issues were resolved by fixing shellspec problem with errexit osh terminate where it must not sh set e foo set e return foo echo should be executed sh set e foo set e false echo should be executed foo echo should be executed | 0 |
246,549 | 7,895,394,132 | IssuesEvent | 2018-06-29 02:59:44 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | The partition on the parallel pgpu profile is incorrect. | Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 2 - Minor Irritation Support Group: Any bug version: 2.7.1 | The partition in the parallel pgpu profile is set to pbatch instead of pgpu. This was most likely a cut and paste error.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 03/19/2014 11:51 am
Original update: 03/25/2014 03:52 pm
Ticket number: 1776 | 1.0 | The partition on the parallel pgpu profile is incorrect. - The partition in the parallel pgpu profile is set to pbatch instead of pgpu. This was most likely a cut and paste error.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 03/19/2014 11:51 am
Original update: 03/25/2014 03:52 pm
Ticket number: 1776 | non_infrastructure | the partition on the parallel pgpu profile is incorrect the partition in the parallel pgpu profile is set to pbatch instead of pgpu this was most likely a cut and paste error redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation am original update pm ticket number | 0 |
24,699 | 17,627,581,380 | IssuesEvent | 2021-08-19 01:07:53 | tgrx/alpha | https://api.github.com/repos/tgrx/alpha | closed | Docker-compose exit by C-c stops the running containers for sure | newcomer configuration management urgent infrastructure env-development | `restart: always` causes DB and DBA containers to continue running after exit by C-c | 1.0 | Docker-compose exit by C-c stops the running containers for sure - `restart: always` causes DB and DBA containers to continue running after exit by C-c | infrastructure | docker compose exit by c c stops the running containers for sure restart always causes db and dba containers to continue running after exit by c c | 1 |
20,970 | 14,265,423,042 | IssuesEvent | 2020-11-20 17:05:42 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | opened | AsyncCompletionSession.Dismiss can deadlock if called from PrioritizedTaskScheduler-AboveNormal | Area-Infrastructure Flaky Resolution-External Test | **PrioritizedTaskScheduler-AboveNormal** is a single-threaded synchronization context which is not tracked by JTF as the "main thread". When it calls `AsyncCompletionSession.Dismiss`, the thread can deadlock due to the use of JoinableTaskFactory.Run. This issue tracks test failures caused by this editor API bug. | 1.0 | AsyncCompletionSession.Dismiss can deadlock if called from PrioritizedTaskScheduler-AboveNormal - **PrioritizedTaskScheduler-AboveNormal** is a single-threaded synchronization context which is not tracked by JTF as the "main thread". When it calls `AsyncCompletionSession.Dismiss`, the thread can deadlock due to the use of JoinableTaskFactory.Run. This issue tracks test failures caused by this editor API bug. | infrastructure | asynccompletionsession dismiss can deadlock if called from prioritizedtaskscheduler abovenormal prioritizedtaskscheduler abovenormal is a single threaded synchronization context which is not tracked by jtf as the main thread when it calls asynccompletionsession dismiss the thread can deadlock due to the use of joinabletaskfactory run this issue tracks test failures caused by this editor api bug | 1 |
304,463 | 26,278,250,541 | IssuesEvent | 2023-01-07 02:22:13 | InsightSoftwareConsortium/ITK | https://api.github.com/repos/InsightSoftwareConsortium/ITK | opened | Check initialization values in tests | type:Testing Good first issue | ### Description
For any given class, we should test that their ivars are initialized to the values in their corresponding .h/.hxx files. This would allow to detect and track changes/bugs related to a given initialization value.
### Steps to Reproduce
Prior to testing the Set/Get methods, check that all member variables have been initialized to the values in their corresponding .h/.hxx files.
Some macros might be helpful, but the macro will need to be called manually for each member/value pair.
### Expected behavior
All members are initialized to some default value, and the exact value is tested.
### Actual behavior
No testing is done for default initialization values.
### Reproducibility
100%.
### Versions
`master`.
### Environment
Any.
### Additional Information
None. | 1.0 | Check initialization values in tests - ### Description
For any given class, we should test that their ivars are initialized to the values in their corresponding .h/.hxx files. This would allow to detect and track changes/bugs related to a given initialization value.
### Steps to Reproduce
Prior to testing the Set/Get methods, check that all member variables have been initialized to the values in their corresponding .h/.hxx files.
Some macros might be helpful, but the macro will need to be called manually for each member/value pair.
### Expected behavior
All members are initialized to some default value, and the exact value is tested.
### Actual behavior
No testing is done for default initialization values.
### Reproducibility
100%.
### Versions
`master`.
### Environment
Any.
### Additional Information
None. | non_infrastructure | check initialization values in tests description for any given class we should test that their ivars are initialized to the values in their corresponding h hxx files this would allow to detect and track changes bugs related to a given initialization value steps to reproduce prior to testing the set get methods check that all member variables have been initialized to the values in their corresponding h hxx files some macros might be helpful but the macro will need to be called manually for each member value pair expected behavior all members are initialized to some default value and the exact value is tested actual behavior no testing is done for default initialization values reproducibility versions master environment any additional information none | 0 |
101,896 | 21,871,886,463 | IssuesEvent | 2022-05-19 06:29:12 | apache/rocketmq | https://api.github.com/repos/apache/rocketmq | closed | Remove redundant exception | type/code style |
Remove redundant exception
**BUG REPORT**
1. Please describe the issue you observed:
- What did you do (The steps to reproduce)?
- What is expected to see?
- What did you see instead?
2. Please tell us about your environment:
3. Other information (e.g. detailed explanation, logs, related issues, suggestions on how to fix, etc):
| 1.0 | Remove redundant exception -
Remove redundant exception
**BUG REPORT**
1. Please describe the issue you observed:
- What did you do (The steps to reproduce)?
- What is expected to see?
- What did you see instead?
2. Please tell us about your environment:
3. Other information (e.g. detailed explanation, logs, related issues, suggestions on how to fix, etc):
| non_infrastructure | remove redundant exception remove redundant exception bug report please describe the issue you observed what did you do the steps to reproduce what is expected to see what did you see instead please tell us about your environment other information e g detailed explanation logs related issues suggestions on how to fix etc | 0 |
160,557 | 13,792,555,848 | IssuesEvent | 2020-10-09 13:46:43 | tomav/docker-mailserver | https://api.github.com/repos/tomav/docker-mailserver | closed | Prevent .dovecot.sieve folder from showing up in the IMAP folder list | documentation frozen due to age help wanted priority 3 [LOW] | I have sieve enabled, and I'm using a Roundcube plugin to configure it. What happens is that for the user I'm configuring, the following symlink is created in the `/var/mail/[domain]/[user]`folder:
`.dovecot.sieve -> sieve/roundcube.sieve`
This leads into my IMAP clients showing a `dovecot` folder with a `sieve` subfolder because the interpret everything in the email folder starting with a dot as an folder.
Did someone manage to prevent this?
This mailinglist posting describes that you should not put the sieve files into the same folder as the emails. They suggest to have a common root and then two subfolders: one for the emails and the other one for sieve. They also strongly suggest to have real home directories:
http://www.dovecot.org/pipermail/dovecot/2011-June/059875.html
| 1.0 | Prevent .dovecot.sieve folder from showing up in the IMAP folder list - I have sieve enabled, and I'm using a Roundcube plugin to configure it. What happens is that for the user I'm configuring, the following symlink is created in the `/var/mail/[domain]/[user]`folder:
`.dovecot.sieve -> sieve/roundcube.sieve`
This leads into my IMAP clients showing a `dovecot` folder with a `sieve` subfolder because the interpret everything in the email folder starting with a dot as an folder.
Did someone manage to prevent this?
This mailinglist posting describes that you should not put the sieve files into the same folder as the emails. They suggest to have a common root and then two subfolders: one for the emails and the other one for sieve. They also strongly suggest to have real home directories:
http://www.dovecot.org/pipermail/dovecot/2011-June/059875.html
| non_infrastructure | prevent dovecot sieve folder from showing up in the imap folder list i have sieve enabled and i m using a roundcube plugin to configure it what happens is that for the user i m configuring the following symlink is created in the var mail folder dovecot sieve sieve roundcube sieve this leads into my imap clients showing a dovecot folder with a sieve subfolder because the interpret everything in the email folder starting with a dot as an folder did someone manage to prevent this this mailinglist posting describes that you should not put the sieve files into the same folder as the emails they suggest to have a common root and then two subfolders one for the emails and the other one for sieve they also strongly suggest to have real home directories | 0 |
181,024 | 13,999,036,240 | IssuesEvent | 2020-10-28 10:16:16 | tomav/docker-mailserver | https://api.github.com/repos/tomav/docker-mailserver | closed | Separate test cases | enhancement priority 2 [MEDIUM] test related | ## Context
With the increasing amount of test cases testing is becoming more and more the bottleneck in development. Even a fast cached (`make all-fast`) build on a current PC currently takes around 17 minutes to complete.
My proposal would be to split the tests into separate test files according to the containers they are executed from.
The user could then decide to execute all containers and therefore all tests or just the particular container and its associated tests.
The `Makefile` would be getting a new optional argument `CONTAINER='container1 container2'`
Furthermore I would like to introduce a optional argument `PERF=(int)` setting, to respect different performance capabilities of different setups.
The setting would basically be a factor applied to any `sleep` command in the `Makefile`.
## Pros
- shorter debugging times(Performance factor and selective testing)
- smaller and therefore more readable test files
- more granular debugging
- less ressource heavy. Containers would be removed after their tests exited successfully.
- downward compatible
## Cons
- not one single searchable tests.bats file
| 1.0 | Separate test cases - ## Context
With the increasing amount of test cases testing is becoming more and more the bottleneck in development. Even a fast cached (`make all-fast`) build on a current PC currently takes around 17 minutes to complete.
My proposal would be to split the tests into separate test files according to the containers they are executed from.
The user could then decide to execute all containers and therefore all tests or just the particular container and its associated tests.
The `Makefile` would be getting a new optional argument `CONTAINER='container1 container2'`
Furthermore I would like to introduce a optional argument `PERF=(int)` setting, to respect different performance capabilities of different setups.
The setting would basically be a factor applied to any `sleep` command in the `Makefile`.
## Pros
- shorter debugging times(Performance factor and selective testing)
- smaller and therefore more readable test files
- more granular debugging
- less ressource heavy. Containers would be removed after their tests exited successfully.
- downward compatible
## Cons
- not one single searchable tests.bats file
| non_infrastructure | separate test cases context with the increasing amount of test cases testing is becoming more and more the bottleneck in development even a fast cached make all fast build on a current pc currently takes around minutes to complete my proposal would be to split the tests into separate test files according to the containers they are executed from the user could then decide to execute all containers and therefore all tests or just the particular container and its associated tests the makefile would be getting a new optional argument container furthermore i would like to introduce a optional argument perf int setting to respect different performance capabilities of different setups the setting would basically be a factor applied to any sleep command in the makefile pros shorter debugging times performance factor and selective testing smaller and therefore more readable test files more granular debugging less ressource heavy containers would be removed after their tests exited successfully downward compatible cons not one single searchable tests bats file | 0 |
31,280 | 25,512,598,895 | IssuesEvent | 2022-11-28 14:10:50 | cds-snc/scan-files | https://api.github.com/repos/cds-snc/scan-files | closed | S3 scan object Node.js 18 logging is broken | Bug | Bogue Infrastructure | # Summary
With the upgrade to Node.js 18 in #441 the Lambda is now broken trying to log the NPM debug info to `~/.npm` which is not writable in a Lambda environment.
Update the logging location to `/tmp`.
# Related
- #441
- cds-snc/platform-core-services#183 | 1.0 | S3 scan object Node.js 18 logging is broken - # Summary
With the upgrade to Node.js 18 in #441 the Lambda is now broken trying to log the NPM debug info to `~/.npm` which is not writable in a Lambda environment.
Update the logging location to `/tmp`.
# Related
- #441
- cds-snc/platform-core-services#183 | infrastructure | scan object node js logging is broken summary with the upgrade to node js in the lambda is now broken trying to log the npm debug info to npm which is not writable in a lambda environment update the logging location to tmp related cds snc platform core services | 1 |
5,475 | 5,693,550,231 | IssuesEvent | 2017-04-15 02:44:18 | mshenfield/subreddit_algebra | https://api.github.com/repos/mshenfield/subreddit_algebra | closed | Store models in a shared S3 bucket | infrastructure | Right now I upload them and they are copied to a special folder on the server that the web app knows to read from. I don't think this will work with [AWS AutoScaling](http://docs.aws.amazon.com/elasticbeanstalk/latest/dg/using-features.managing.as.html) because the new instance will include the most recent deployment, which may or may not include the latest models. | 1.0 | Store models in a shared S3 bucket - Right now I upload them and they are copied to a special folder on the server that the web app knows to read from. I don't think this will work with [AWS AutoScaling](http://docs.aws.amazon.com/elasticbeanstalk/latest/dg/using-features.managing.as.html) because the new instance will include the most recent deployment, which may or may not include the latest models. | infrastructure | store models in a shared bucket right now i upload them and they are copied to a special folder on the server that the web app knows to read from i don t think this will work with because the new instance will include the most recent deployment which may or may not include the latest models | 1 |
7,700 | 7,054,795,894 | IssuesEvent | 2018-01-04 03:25:27 | vistadataproject/VDM | https://api.github.com/repos/vistadataproject/VDM | closed | fileman.js prototype and isolation test suite | Infrastructure Track | __Background__: UPDATE^DIE is the main API we use for changing FileMan data behind VDM. We need to have a formal and isolated UPDATE^DIE test suite covering scenarios like mixing updates, creates and deletes in single calls.
__The Prototype space__: VDM/prototypes/fileman - README covers relevant API files.
__TODO__: create pure tests that have features of 100 and 52 and other complicated files. Include .11 cross references too and other indexes.
__Interplay with dmUtils__: which checks valid combinations of VDM input data that leads to fm input data. | 1.0 | fileman.js prototype and isolation test suite - __Background__: UPDATE^DIE is the main API we use for changing FileMan data behind VDM. We need to have a formal and isolated UPDATE^DIE test suite covering scenarios like mixing updates, creates and deletes in single calls.
__The Prototype space__: VDM/prototypes/fileman - README covers relevant API files.
__TODO__: create pure tests that have features of 100 and 52 and other complicated files. Include .11 cross references too and other indexes.
__Interplay with dmUtils__: which checks valid combinations of VDM input data that leads to fm input data. | infrastructure | fileman js prototype and isolation test suite background update die is the main api we use for changing fileman data behind vdm we need to have a formal and isolated update die test suite covering scenarios like mixing updates creates and deletes in single calls the prototype space vdm prototypes fileman readme covers relevant api files todo create pure tests that have features of and and other complicated files include cross references too and other indexes interplay with dmutils which checks valid combinations of vdm input data that leads to fm input data | 1 |
552,913 | 16,330,648,215 | IssuesEvent | 2021-05-12 08:48:51 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.mall.cz - design is broken | browser-firefox engine-gecko priority-normal | <!-- @browser: Firefox 78.10.1esr -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/73220 -->
**URL**: https://www.mall.cz/8k-uhd
**Browser / Version**: Firefox 78.10.1esr
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Design is broken
**Description**: Items are misaligned
**Steps to Reproduce**:
Left column (with filters) is too narrow, so it is impossible to filter commodities. Right column (with tiles of goods) exceeds right edge of the browser.(monitor resolution 3840x2160 pix)
The same issue happens in my old computer with windows 7 and the exact same browser Firefox 78.10.1 esr. (monitor resolution 1920x1080 pix)
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/1313910f-7e8a-4079-b008-6e6f25eee967.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.mall.cz - design is broken - <!-- @browser: Firefox 78.10.1esr -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Firefox/78.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/73220 -->
**URL**: https://www.mall.cz/8k-uhd
**Browser / Version**: Firefox 78.10.1esr
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Design is broken
**Description**: Items are misaligned
**Steps to Reproduce**:
Left column (with filters) is too narrow, so it is impossible to filter commodities. Right column (with tiles of goods) exceeds right edge of the browser.(monitor resolution 3840x2160 pix)
The same issue happens in my old computer with windows 7 and the exact same browser Firefox 78.10.1 esr. (monitor resolution 1920x1080 pix)
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2021/5/1313910f-7e8a-4079-b008-6e6f25eee967.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | design is broken url browser version firefox operating system windows tested another browser yes edge problem type design is broken description items are misaligned steps to reproduce left column with filters is too narrow so it is impossible to filter commodities right column with tiles of goods exceeds right edge of the browser monitor resolution pix the same issue happens in my old computer with windows and the exact same browser firefox esr monitor resolution pix view the screenshot img alt screenshot src browser configuration none from with ❤️ | 0 |
9,983 | 8,303,718,408 | IssuesEvent | 2018-09-21 18:32:32 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | opened | All Configurations Package validation leg not failing CI | area-Infrastructure packaging | https://ci3.dot.net/job/dotnet_corefx/job/master/job/windows-TGroup_all+CGroup_Debug+AGroup_x64+TestOuter_false_prtest/16272/consoleFull
The All Configurations leg has been passing in CI but there are failures in the testing.
```
11:01:50 D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\tools\sdk\2.1.401\NuGet.targets(114,5): error : Cannot create a file when that file already exists [D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\projects\System.Composition.TypedParts\netcoreapp1.1\project.csproj] [D:\j\workspace\windows-TGrou---0d2c9ac4\pkg\test\testPackages.proj]
```
and
```
11:02:00 D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\projects\System.Data.SqlClient\netcoreapp2.0\project.csproj : error NU1603: System.Data.SqlClient 4.7.0-preview1-26921-0 depends on System.Buffers (>= 4.6.0-preview1-26921-0) but System.Buffers 4.6.0-preview1-26921-0 was not found. An approximate best match of System.Buffers 4.6.0-preview3-26501-01 was resolved. [D:\j\workspace\windows-TGrou---0d2c9ac4\pkg\test\testPackages.proj]
```
and
```
11:02:10 D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\projects\System.Diagnostics.EventLog\netcoreapp2.1\project.csproj : error NU1603: Microsoft.Win32.Registry 4.6.0-preview1-26921-0 depends on System.Buffers (>= 4.6.0-preview1-26921-0) but System.Buffers 4.6.0-preview1-26921-0 was not found. An approximate best match of System.Buffers 4.6.0-preview3-26501-01 was resolved. [D:\j\workspace\windows-TGrou---0d2c9ac4\pkg\test\testPackages.proj]
```
@joperezr can you have a look at the first error and figure out why this isn't causing CI to fail?
@safern can you look at the other 2 errors as they are related to the Buffers package removal.
| 1.0 | All Configurations Package validation leg not failing CI - https://ci3.dot.net/job/dotnet_corefx/job/master/job/windows-TGroup_all+CGroup_Debug+AGroup_x64+TestOuter_false_prtest/16272/consoleFull
The All Configurations leg has been passing in CI but there are failures in the testing.
```
11:01:50 D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\tools\sdk\2.1.401\NuGet.targets(114,5): error : Cannot create a file when that file already exists [D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\projects\System.Composition.TypedParts\netcoreapp1.1\project.csproj] [D:\j\workspace\windows-TGrou---0d2c9ac4\pkg\test\testPackages.proj]
```
and
```
11:02:00 D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\projects\System.Data.SqlClient\netcoreapp2.0\project.csproj : error NU1603: System.Data.SqlClient 4.7.0-preview1-26921-0 depends on System.Buffers (>= 4.6.0-preview1-26921-0) but System.Buffers 4.6.0-preview1-26921-0 was not found. An approximate best match of System.Buffers 4.6.0-preview3-26501-01 was resolved. [D:\j\workspace\windows-TGrou---0d2c9ac4\pkg\test\testPackages.proj]
```
and
```
11:02:10 D:\j\workspace\windows-TGrou---0d2c9ac4\bin\testPkg\projects\System.Diagnostics.EventLog\netcoreapp2.1\project.csproj : error NU1603: Microsoft.Win32.Registry 4.6.0-preview1-26921-0 depends on System.Buffers (>= 4.6.0-preview1-26921-0) but System.Buffers 4.6.0-preview1-26921-0 was not found. An approximate best match of System.Buffers 4.6.0-preview3-26501-01 was resolved. [D:\j\workspace\windows-TGrou---0d2c9ac4\pkg\test\testPackages.proj]
```
@joperezr can you have a look at the first error and figure out why this isn't causing CI to fail?
@safern can you look at the other 2 errors as they are related to the Buffers package removal.
| infrastructure | all configurations package validation leg not failing ci the all configurations leg has been passing in ci but there are failures in the testing d j workspace windows tgrou bin testpkg tools sdk nuget targets error cannot create a file when that file already exists and d j workspace windows tgrou bin testpkg projects system data sqlclient project csproj error system data sqlclient depends on system buffers but system buffers was not found an approximate best match of system buffers was resolved and d j workspace windows tgrou bin testpkg projects system diagnostics eventlog project csproj error microsoft registry depends on system buffers but system buffers was not found an approximate best match of system buffers was resolved joperezr can you have a look at the first error and figure out why this isn t causing ci to fail safern can you look at the other errors as they are related to the buffers package removal | 1 |
12,572 | 9,854,403,205 | IssuesEvent | 2019-06-19 16:45:58 | SNLComputation/Albany | https://api.github.com/repos/SNLComputation/Albany | closed | Replacing Epetra dgdp in PHAL::Workset with Tpetra dgdpT throughout Albany | ATO infrastructure | I have been making an effort to remove Epetra member variables in PHAL::Workset, as a part of the conversion of ATO to Tpetra/Thyra. This includes in particular responses and their derivatives (workset.g, workset.dgdp, and workset.overlapped_dgdp). I was able to successfully remove the Epetra versions of g and overlapped_dgdp in PHAL::Workset and replace them with Tpetra analogs throughout Albany. I am having issues with workset.dgdp, however. The only place where workset.dgdp appears is:
src/evaluators/PHAL_SeparableScatterScalarResponse_Def.hpp
responses/Albany_FieldManagerScalarResponseFunction.cpp
The switch from workset.dgdp to workset.dgdpT should be trivial. However, when I make the change, the ATO::Constraint2D_adj test fails. @mperego, can you please have a look at this, as the author of distributed responses?
| 1.0 | Replacing Epetra dgdp in PHAL::Workset with Tpetra dgdpT throughout Albany - I have been making an effort to remove Epetra member variables in PHAL::Workset, as a part of the conversion of ATO to Tpetra/Thyra. This includes in particular responses and their derivatives (workset.g, workset.dgdp, and workset.overlapped_dgdp). I was able to successfully remove the Epetra versions of g and overlapped_dgdp in PHAL::Workset and replace them with Tpetra analogs throughout Albany. I am having issues with workset.dgdp, however. The only place where workset.dgdp appears is:
src/evaluators/PHAL_SeparableScatterScalarResponse_Def.hpp
responses/Albany_FieldManagerScalarResponseFunction.cpp
The switch from workset.dgdp to workset.dgdpT should be trivial. However, when I make the change, the ATO::Constraint2D_adj test fails. @mperego, can you please have a look at this, as the author of distributed responses?
| infrastructure | replacing epetra dgdp in phal workset with tpetra dgdpt throughout albany i have been making an effort to remove epetra member variables in phal workset as a part of the conversion of ato to tpetra thyra this includes in particular responses and their derivatives workset g workset dgdp and workset overlapped dgdp i was able to successfully remove the epetra versions of g and overlapped dgdp in phal workset and replace them with tpetra analogs throughout albany i am having issues with workset dgdp however the only place where workset dgdp appears is src evaluators phal separablescatterscalarresponse def hpp responses albany fieldmanagerscalarresponsefunction cpp the switch from workset dgdp to workset dgdpt should be trivial however when i make the change the ato adj test fails mperego can you please have a look at this as the author of distributed responses | 1 |
2,607 | 8,033,111,091 | IssuesEvent | 2018-07-29 00:12:54 | City-Bureau/city-scrapers | https://api.github.com/repos/City-Bureau/city-scrapers | closed | Detroit Land Bank Authority | architecture: spiders location: detroit new spider needed | New spider for Detroit Land Bank Authority (det_land_bank) from
https://buildingdetroit.org/events/meetings
This spider is currently part of ticketing group 4 on the [Public Meetings Aggregator](https://docs.google.com/spreadsheets/d/1Uu9FmcQAy8AXhBR2bj84LbSdlfgMNNroa9P89rsDJSQ/edit#gid=1214169297). | 1.0 | Detroit Land Bank Authority - New spider for Detroit Land Bank Authority (det_land_bank) from
https://buildingdetroit.org/events/meetings
This spider is currently part of ticketing group 4 on the [Public Meetings Aggregator](https://docs.google.com/spreadsheets/d/1Uu9FmcQAy8AXhBR2bj84LbSdlfgMNNroa9P89rsDJSQ/edit#gid=1214169297). | non_infrastructure | detroit land bank authority new spider for detroit land bank authority det land bank from this spider is currently part of ticketing group on the | 0 |
21,179 | 28,149,381,438 | IssuesEvent | 2023-04-02 21:19:32 | AbdElAziz333/Pluto | https://api.github.com/repos/AbdElAziz333/Pluto | closed | Mod causes server connection issues | bug vanilla parity in processing | Game version: Forge 40.1.84 (Minecraft 1.18.2)
Mod version: 0.0.2 (https://www.curseforge.com/minecraft/mc-mods/pluto/files/4049858)
Steps to reproduce:
-Install Pluto on both server and client
-Join server
-Disconnect from server
You will now be unable to reconnect to the server until the client is restarted. | 1.0 | Mod causes server connection issues - Game version: Forge 40.1.84 (Minecraft 1.18.2)
Mod version: 0.0.2 (https://www.curseforge.com/minecraft/mc-mods/pluto/files/4049858)
Steps to reproduce:
-Install Pluto on both server and client
-Join server
-Disconnect from server
You will now be unable to reconnect to the server until the client is restarted. | non_infrastructure | mod causes server connection issues game version forge minecraft mod version steps to reproduce install pluto on both server and client join server disconnect from server you will now be unable to reconnect to the server until the client is restarted | 0 |
6,264 | 6,278,924,202 | IssuesEvent | 2017-07-18 15:15:38 | dart-lang/site-webdev | https://api.github.com/repos/dart-lang/site-webdev | opened | Ensure that basic project setup and build process is beginner friendly | Infrastructure | - README install instructions need updating #829
- `scripts/serve_local.sh`
- Switch to using superstatic so as to avoid having to deal with firebase project access permissions (this is what we did for the Travis build process).
- Setup for use of [browsersync](https://www.browsersync.io). Also see [this fb post](https://firebase.googleblog.com/2015/12/a-host-of-improvements_61.html).
| 1.0 | Ensure that basic project setup and build process is beginner friendly - - README install instructions need updating #829
- `scripts/serve_local.sh`
- Switch to using superstatic so as to avoid having to deal with firebase project access permissions (this is what we did for the Travis build process).
- Setup for use of [browsersync](https://www.browsersync.io). Also see [this fb post](https://firebase.googleblog.com/2015/12/a-host-of-improvements_61.html).
| infrastructure | ensure that basic project setup and build process is beginner friendly readme install instructions need updating scripts serve local sh switch to using superstatic so as to avoid having to deal with firebase project access permissions this is what we did for the travis build process setup for use of also see | 1 |
2,140 | 3,522,062,032 | IssuesEvent | 2016-01-13 07:14:33 | SatelliteQE/robottelo | https://api.github.com/repos/SatelliteQE/robottelo | closed | Update jenkins for new pxe-less discovery tests. | Infrastructure | From https://github.com/SatelliteQE/robottelo/pull/3204 :
Please note that robottelo-ci and robottelo.properties.sample still needs to be updated with discovery_iso parameter. Idea here is to share the approach.
I'll update robottelo-ci and sample-properties file, If all changes defined in this PR looks good to reviewers. | 1.0 | Update jenkins for new pxe-less discovery tests. - From https://github.com/SatelliteQE/robottelo/pull/3204 :
Please note that robottelo-ci and robottelo.properties.sample still needs to be updated with discovery_iso parameter. Idea here is to share the approach.
I'll update robottelo-ci and sample-properties file, If all changes defined in this PR looks good to reviewers. | infrastructure | update jenkins for new pxe less discovery tests from please note that robottelo ci and robottelo properties sample still needs to be updated with discovery iso parameter idea here is to share the approach i ll update robottelo ci and sample properties file if all changes defined in this pr looks good to reviewers | 1 |
55,835 | 23,616,011,474 | IssuesEvent | 2022-08-24 15:55:44 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Update RAS QC Viewer to Enhance Filtering | Workgroup: SMB Type: Data Service: Geo | Filter needs to be updated to more easily distinguish features with linked photos. New data needs to be configured in attribute table | 1.0 | Update RAS QC Viewer to Enhance Filtering - Filter needs to be updated to more easily distinguish features with linked photos. New data needs to be configured in attribute table | non_infrastructure | update ras qc viewer to enhance filtering filter needs to be updated to more easily distinguish features with linked photos new data needs to be configured in attribute table | 0 |
328,673 | 9,998,455,384 | IssuesEvent | 2019-07-12 08:14:06 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.backcountry.com - design is broken | browser-fenix engine-gecko priority-normal | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.backcountry.com/explore/my-climbing-partner-eats-chicken-liver
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: no blog pics
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.backcountry.com - design is broken - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.backcountry.com/explore/my-climbing-partner-eats-chicken-liver
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: no blog pics
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | design is broken url browser version firefox mobile operating system android tested another browser no problem type design is broken description no blog pics steps to reproduce browser configuration none from with ❤️ | 0 |
41,693 | 10,762,733,544 | IssuesEvent | 2019-11-01 00:50:24 | randombit/botan | https://api.github.com/repos/randombit/botan | closed | Building with jom/nmake only works if using MSVC | build problem | Cannot build the library because there is no libbotan-2.a

| 1.0 | Building with jom/nmake only works if using MSVC - Cannot build the library because there is no libbotan-2.a

| non_infrastructure | building with jom nmake only works if using msvc cannot build the library because there is no libbotan a | 0 |
184,256 | 14,283,712,148 | IssuesEvent | 2020-11-23 11:25:05 | Puzzlepart/prosjektportalen365 | https://api.github.com/repos/Puzzlepart/prosjektportalen365 | opened | Test av v1.2.4 | release test | ## Testing the new version before creating new release
Checklist for testing the additions, changes and fixes
# Version 1.2.4
## Added
- [ ] Added "default" option for extensions, similar to list content #328
- [ ] Added info message if there are unpublished statusreports #340
- [ ] Added published/unpublished indicators for statusreports in dropdown and ribbon #341
- [ ] Added possiblity to delete unpublished statusreports #343
- [ ] Added PNG snapshot when publishing project status #337
## Fixed
- [ ] Improved failure handling for PlannerConfiguration task in Project Setup #329
- [ ] Support adding AD groups to get porfolio insights from SP group #332
- [ ] Change to latest statusreport when creating a new statusreport #343
After everything has been checked and approved a release of the new version can be created. It is important that the changelog, this issue and the release notes are equal.
| 1.0 | Test av v1.2.4 - ## Testing the new version before creating new release
Checklist for testing the additions, changes and fixes
# Version 1.2.4
## Added
- [ ] Added "default" option for extensions, similar to list content #328
- [ ] Added info message if there are unpublished statusreports #340
- [ ] Added published/unpublished indicators for statusreports in dropdown and ribbon #341
- [ ] Added possiblity to delete unpublished statusreports #343
- [ ] Added PNG snapshot when publishing project status #337
## Fixed
- [ ] Improved failure handling for PlannerConfiguration task in Project Setup #329
- [ ] Support adding AD groups to get porfolio insights from SP group #332
- [ ] Change to latest statusreport when creating a new statusreport #343
After everything has been checked and approved a release of the new version can be created. It is important that the changelog, this issue and the release notes are equal.
| non_infrastructure | test av testing the new version before creating new release checklist for testing the additions changes and fixes version added added default option for extensions similar to list content added info message if there are unpublished statusreports added published unpublished indicators for statusreports in dropdown and ribbon added possiblity to delete unpublished statusreports added png snapshot when publishing project status fixed improved failure handling for plannerconfiguration task in project setup support adding ad groups to get porfolio insights from sp group change to latest statusreport when creating a new statusreport after everything has been checked and approved a release of the new version can be created it is important that the changelog this issue and the release notes are equal | 0 |
30,338 | 24,755,116,274 | IssuesEvent | 2022-10-21 16:56:16 | GoogleForCreators/web-stories-wp | https://api.github.com/repos/GoogleForCreators/web-stories-wp | closed | E2E Tests: Fix flakey .mov insertion test | Type: Bug Type: Infrastructure P3 Pod: WP Package: E2E Tests | <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Task Description
<!-- A clear and concise description of what this task is about. -->
Unfortunately #8215 didn't help much for this tests, it continues to be very flakey.
Some observations:
* Chromium has issues with displaying MOV and MP4 reliably (which makes sense, as both aren't supported in that browser)
* Tests work more reliably in non-headless mode
* On Chromium, the video element briefly disappears from the canvas during transcoding | 1.0 | E2E Tests: Fix flakey .mov insertion test - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Task Description
<!-- A clear and concise description of what this task is about. -->
Unfortunately #8215 didn't help much for this tests, it continues to be very flakey.
Some observations:
* Chromium has issues with displaying MOV and MP4 reliably (which makes sense, as both aren't supported in that browser)
* Tests work more reliably in non-headless mode
* On Chromium, the video element briefly disappears from the canvas during transcoding | infrastructure | tests fix flakey mov insertion test task description unfortunately didn t help much for this tests it continues to be very flakey some observations chromium has issues with displaying mov and reliably which makes sense as both aren t supported in that browser tests work more reliably in non headless mode on chromium the video element briefly disappears from the canvas during transcoding | 1 |
326,869 | 9,961,849,259 | IssuesEvent | 2019-07-07 09:13:26 | Ash258/Scoop-GithubActions | https://api.github.com/repos/Ash258/Scoop-GithubActions | closed | Pull request trigger on comment | enhancement high-priority pull-request-action | Trigger Pull request action when PR comment `/verify` is added
https://developer.github.com/v3/issues/comments/
https://developer.github.com/actions/managing-workflows/workflow-configuration-options/#events-supported-in-workflow-files | 1.0 | Pull request trigger on comment - Trigger Pull request action when PR comment `/verify` is added
https://developer.github.com/v3/issues/comments/
https://developer.github.com/actions/managing-workflows/workflow-configuration-options/#events-supported-in-workflow-files | non_infrastructure | pull request trigger on comment trigger pull request action when pr comment verify is added | 0 |
158,827 | 24,902,232,278 | IssuesEvent | 2022-10-28 22:32:44 | Office-of-Digital-Services/California-State-Web-Template-Website | https://api.github.com/repos/Office-of-Digital-Services/California-State-Web-Template-Website | closed | Create new Visual Design landing page that will combine content from Typography, Color Schemes, and Icons pages | eng design P1 | Combine content from typography, color schemes and icons pages into one Visual Design page. For icons section make some intro text and the link to separate icons page.
Add "On this page" page navigation with anchors to Typography, Color Schemes and Icons sections | 1.0 | Create new Visual Design landing page that will combine content from Typography, Color Schemes, and Icons pages - Combine content from typography, color schemes and icons pages into one Visual Design page. For icons section make some intro text and the link to separate icons page.
Add "On this page" page navigation with anchors to Typography, Color Schemes and Icons sections | non_infrastructure | create new visual design landing page that will combine content from typography color schemes and icons pages combine content from typography color schemes and icons pages into one visual design page for icons section make some intro text and the link to separate icons page add on this page page navigation with anchors to typography color schemes and icons sections | 0 |
614,016 | 19,122,187,829 | IssuesEvent | 2021-12-01 00:38:15 | kubernetes/website | https://api.github.com/repos/kubernetes/website | closed | DNS for Services and Pods docs improvement suggestions | kind/bug priority/backlog lifecycle/rotten triage/accepted | Here's the text I see in the Pod's DNS policy section today:
https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy
```
Pod's DNS Policy
DNS policies can be set on a per-pod basis. Currently Kubernetes supports the following pod-specific DNS policies. These policies are specified in the dnsPolicy field of a Pod Spec.
"Default": The Pod inherits the name resolution configuration from the node that the pods run on. See related discussion for more details.
"ClusterFirst": Any DNS query that does not match the configured cluster domain suffix, such as "www.kubernetes.io", is forwarded to the upstream nameserver inherited from the node. Cluster administrators may have extra stub-domain and upstream DNS servers configured. See related discussion for details on how DNS queries are handled in those cases.
"ClusterFirstWithHostNet": For Pods running with hostNetwork, you should explicitly set its DNS policy "ClusterFirstWithHostNet".
"None": It allows a Pod to ignore DNS settings from the Kubernetes environment. All DNS settings are supposed to be provided using the dnsConfig field in the Pod Spec. See Pod's DNS config subsection below.
Note: "Default" is not the default DNS policy. If dnsPolicy is not explicitly specified, then "ClusterFirst" is used.
```
I've found a few issues with this:
* `ClusterFirstWithHostNet` doesn't explain what it does, it just says to use it. As far as I could tell from some discussion with colleagues, it's the same as `ClusterFirst` but using search domains from the node on the pod as well as the search domains from `ClusterFirst`. Is that the case?
* The link for "See related discussion for details on how DNS queries are handled in those cases." leads here: https://kubernetes.io/docs/tasks/administer-cluster/dns-custom-nameservers/#effects-on-pods . However, I cannot find any "effects on pods" section. Perhaps this was removed or moved at some point? I'd be curious what that discussion was about, but I don't know where it's supposed to be or what the discussion says so it's a bit hard to find without a valid link. | 1.0 | DNS for Services and Pods docs improvement suggestions - Here's the text I see in the Pod's DNS policy section today:
https://kubernetes.io/docs/concepts/services-networking/dns-pod-service/#pod-s-dns-policy
```
Pod's DNS Policy
DNS policies can be set on a per-pod basis. Currently Kubernetes supports the following pod-specific DNS policies. These policies are specified in the dnsPolicy field of a Pod Spec.
"Default": The Pod inherits the name resolution configuration from the node that the pods run on. See related discussion for more details.
"ClusterFirst": Any DNS query that does not match the configured cluster domain suffix, such as "www.kubernetes.io", is forwarded to the upstream nameserver inherited from the node. Cluster administrators may have extra stub-domain and upstream DNS servers configured. See related discussion for details on how DNS queries are handled in those cases.
"ClusterFirstWithHostNet": For Pods running with hostNetwork, you should explicitly set its DNS policy "ClusterFirstWithHostNet".
"None": It allows a Pod to ignore DNS settings from the Kubernetes environment. All DNS settings are supposed to be provided using the dnsConfig field in the Pod Spec. See Pod's DNS config subsection below.
Note: "Default" is not the default DNS policy. If dnsPolicy is not explicitly specified, then "ClusterFirst" is used.
```
I've found a few issues with this:
* `ClusterFirstWithHostNet` doesn't explain what it does, it just says to use it. As far as I could tell from some discussion with colleagues, it's the same as `ClusterFirst` but using search domains from the node on the pod as well as the search domains from `ClusterFirst`. Is that the case?
* The link for "See related discussion for details on how DNS queries are handled in those cases." leads here: https://kubernetes.io/docs/tasks/administer-cluster/dns-custom-nameservers/#effects-on-pods . However, I cannot find any "effects on pods" section. Perhaps this was removed or moved at some point? I'd be curious what that discussion was about, but I don't know where it's supposed to be or what the discussion says so it's a bit hard to find without a valid link. | non_infrastructure | dns for services and pods docs improvement suggestions here s the text i see in the pod s dns policy section today pod s dns policy dns policies can be set on a per pod basis currently kubernetes supports the following pod specific dns policies these policies are specified in the dnspolicy field of a pod spec default the pod inherits the name resolution configuration from the node that the pods run on see related discussion for more details clusterfirst any dns query that does not match the configured cluster domain suffix such as is forwarded to the upstream nameserver inherited from the node cluster administrators may have extra stub domain and upstream dns servers configured see related discussion for details on how dns queries are handled in those cases clusterfirstwithhostnet for pods running with hostnetwork you should explicitly set its dns policy clusterfirstwithhostnet none it allows a pod to ignore dns settings from the kubernetes environment all dns settings are supposed to be provided using the dnsconfig field in the pod spec see pod s dns config subsection below note default is not the default dns policy if dnspolicy is not explicitly specified then clusterfirst is used i ve found a few issues with this clusterfirstwithhostnet doesn t explain what it does it just says to use it as far as i could tell from some discussion with colleagues it s the same as clusterfirst but using search domains from the node on the pod as well as the search domains from clusterfirst is that the case the link for see related discussion for details on how dns queries are handled in those cases leads here however i cannot find any effects on pods section perhaps this was removed or moved at some point i d be curious what that discussion was about but i don t know where it s supposed to be or what the discussion says so it s a bit hard to find without a valid link | 0 |
13,858 | 10,500,391,914 | IssuesEvent | 2019-09-26 10:23:20 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Couldn't get all the processes data running in Metricbeats Windows Module | :Windows :infrastructure [zube]: In Review | Hi @narph -
I took the snapshot build and tested the metricbeats, I encountered an issue with the Windows Perfmon Module, metricbeats couldn't able to pull all the processes data running on the server, lets say if I restart a process, metricbeats is not able to capture the data for that particular restarted process unless I restart the metricbeats, Please let me know if you need any additional information. Thanks !
| 1.0 | Couldn't get all the processes data running in Metricbeats Windows Module - Hi @narph -
I took the snapshot build and tested the metricbeats, I encountered an issue with the Windows Perfmon Module, metricbeats couldn't able to pull all the processes data running on the server, lets say if I restart a process, metricbeats is not able to capture the data for that particular restarted process unless I restart the metricbeats, Please let me know if you need any additional information. Thanks !
| infrastructure | couldn t get all the processes data running in metricbeats windows module hi narph i took the snapshot build and tested the metricbeats i encountered an issue with the windows perfmon module metricbeats couldn t able to pull all the processes data running on the server lets say if i restart a process metricbeats is not able to capture the data for that particular restarted process unless i restart the metricbeats please let me know if you need any additional information thanks | 1 |
101,513 | 16,512,296,360 | IssuesEvent | 2021-05-26 06:28:45 | valtech-ch/microservice-kubernetes-cluster | https://api.github.com/repos/valtech-ch/microservice-kubernetes-cluster | opened | CVE-2015-4852 (High) detected in commons-collections-3.2.1.jar | security vulnerability | ## CVE-2015-4852 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary>
<p>Types that extend and augment the Java Collections Framework.</p>
<p>Path to dependency file: microservice-kubernetes-cluster/persistence/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-collections/commons-collections/3.2.1/761ea405b9b37ced573d2df0d1e3a4e0f9edc668/commons-collections-3.2.1.jar</p>
<p>
Dependency Hierarchy:
- springfox-staticdocs-2.6.1.jar (Root Library)
- swagger2markup-0.9.2.jar
- :x: **commons-collections-3.2.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/valtech-ch/microservice-kubernetes-cluster/commit/eb274179a823f7d17154880d5a503973bae259a0">eb274179a823f7d17154880d5a503973bae259a0</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.
<p>Publish Date: 2015-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852>CVE-2015-4852</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p>
<p>Release Date: 2015-11-18</p>
<p>Fix Resolution: commons-collections:commons-collections:3.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-4852 (High) detected in commons-collections-3.2.1.jar - ## CVE-2015-4852 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary>
<p>Types that extend and augment the Java Collections Framework.</p>
<p>Path to dependency file: microservice-kubernetes-cluster/persistence/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-collections/commons-collections/3.2.1/761ea405b9b37ced573d2df0d1e3a4e0f9edc668/commons-collections-3.2.1.jar</p>
<p>
Dependency Hierarchy:
- springfox-staticdocs-2.6.1.jar (Root Library)
- swagger2markup-0.9.2.jar
- :x: **commons-collections-3.2.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/valtech-ch/microservice-kubernetes-cluster/commit/eb274179a823f7d17154880d5a503973bae259a0">eb274179a823f7d17154880d5a503973bae259a0</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The WLS Security component in Oracle WebLogic Server 10.3.6.0, 12.1.2.0, 12.1.3.0, and 12.2.1.0 allows remote attackers to execute arbitrary commands via a crafted serialized Java object in T3 protocol traffic to TCP port 7001, related to oracle_common/modules/com.bea.core.apache.commons.collections.jar. NOTE: the scope of this CVE is limited to the WebLogic Server product.
<p>Publish Date: 2015-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-4852>CVE-2015-4852</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2015/11/17/19">https://www.openwall.com/lists/oss-security/2015/11/17/19</a></p>
<p>Release Date: 2015-11-18</p>
<p>Fix Resolution: commons-collections:commons-collections:3.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in commons collections jar cve high severity vulnerability vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file microservice kubernetes cluster persistence build gradle path to vulnerable library home wss scanner gradle caches modules files commons collections commons collections commons collections jar dependency hierarchy springfox staticdocs jar root library jar x commons collections jar vulnerable library found in head commit a href found in base branch develop vulnerability details the wls security component in oracle weblogic server and allows remote attackers to execute arbitrary commands via a crafted serialized java object in protocol traffic to tcp port related to oracle common modules com bea core apache commons collections jar note the scope of this cve is limited to the weblogic server product publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections step up your open source security game with whitesource | 0 |
19,522 | 3,461,596,878 | IssuesEvent | 2015-12-20 06:58:09 | featherplain/amethyst | https://api.github.com/repos/featherplain/amethyst | opened | 一部スタイルの修正 | bug design | - [ ] 先頭固定表示に `featured` のタグをつける
- [ ] 個別投稿のときに余白がない (画像参照)
- [ ] `<code>` と `<pre>` の `color` を `$colorBase` に修正 | 1.0 | 一部スタイルの修正 - - [ ] 先頭固定表示に `featured` のタグをつける
- [ ] 個別投稿のときに余白がない (画像参照)
- [ ] `<code>` と `<pre>` の `color` を `$colorBase` に修正 | non_infrastructure | 一部スタイルの修正 先頭固定表示に featured のタグをつける 個別投稿のときに余白がない 画像参照 と の color を colorbase に修正 | 0 |
32,879 | 27,063,638,188 | IssuesEvent | 2023-02-13 21:58:09 | katefike/sage | https://api.github.com/repos/katefike/sage | closed | Configure Postfix to use TLS Certificates | infrastructure | Possibly have the ansible script run the commands from inside the mailserver to make the cert and key file. | 1.0 | Configure Postfix to use TLS Certificates - Possibly have the ansible script run the commands from inside the mailserver to make the cert and key file. | infrastructure | configure postfix to use tls certificates possibly have the ansible script run the commands from inside the mailserver to make the cert and key file | 1 |
52,969 | 13,249,541,341 | IssuesEvent | 2020-08-19 21:00:00 | ophrescue/RescueRails | https://api.github.com/repos/ophrescue/RescueRails | opened | AdopterSearcher not finding some applications | Defect | I think this bug was introduced in #1649
Users reported issues when searching for Adoption Applications, from initial testing it seems that Running and AdopterSearch will only return Unassigned applications.
Need to confirm bug, write tests to replicate and apply fix. | 1.0 | AdopterSearcher not finding some applications - I think this bug was introduced in #1649
Users reported issues when searching for Adoption Applications, from initial testing it seems that Running and AdopterSearch will only return Unassigned applications.
Need to confirm bug, write tests to replicate and apply fix. | non_infrastructure | adoptersearcher not finding some applications i think this bug was introduced in users reported issues when searching for adoption applications from initial testing it seems that running and adoptersearch will only return unassigned applications need to confirm bug write tests to replicate and apply fix | 0 |
11,098 | 8,925,199,084 | IssuesEvent | 2019-01-21 21:39:57 | angular/material2 | https://api.github.com/repos/angular/material2 | closed | Package docs content into an npm package | in progress infrastructure | We currently publish the docs content to https://github.com/angular/material2-docs-content in an ad hoc structure. It's not an npm package, meaning we need a bash script to pull in the content to the right place.
**Things that are pending:**
- [X] Build material-examples using Bazel
- [x] Generate highlighted files from example source w/ Bazel
- [x] Docs SCSS should no longer expect a class for each markdown HTML element
- [x] Package all together in a NPM package:
* Example package (needed for show-casing examples in the docs)
* StackBlitz example source files (needed for external stackblitz examples)
* Highlighted example source files (in `.html`, needed for source view of examples)
* Guides which will be shown in the docs
* Components overview files | 1.0 | Package docs content into an npm package - We currently publish the docs content to https://github.com/angular/material2-docs-content in an ad hoc structure. It's not an npm package, meaning we need a bash script to pull in the content to the right place.
**Things that are pending:**
- [X] Build material-examples using Bazel
- [x] Generate highlighted files from example source w/ Bazel
- [x] Docs SCSS should no longer expect a class for each markdown HTML element
- [x] Package all together in a NPM package:
* Example package (needed for show-casing examples in the docs)
* StackBlitz example source files (needed for external stackblitz examples)
* Highlighted example source files (in `.html`, needed for source view of examples)
* Guides which will be shown in the docs
* Components overview files | infrastructure | package docs content into an npm package we currently publish the docs content to in an ad hoc structure it s not an npm package meaning we need a bash script to pull in the content to the right place things that are pending build material examples using bazel generate highlighted files from example source w bazel docs scss should no longer expect a class for each markdown html element package all together in a npm package example package needed for show casing examples in the docs stackblitz example source files needed for external stackblitz examples highlighted example source files in html needed for source view of examples guides which will be shown in the docs components overview files | 1 |
21,143 | 14,398,794,995 | IssuesEvent | 2020-12-03 10:03:40 | radicle-dev/radicle-upstream | https://api.github.com/repos/radicle-dev/radicle-upstream | closed | Prefix proxy process to make it clear it belongs to radicle | improvement infrastructure proxy | Currently when debugging, esp. in scenarios where the proxy process is lingering, it's not always easy to spot that it belongs to radicle. A simple `radilce-` prefix would help to aid with that.
Ref. #1440 | 1.0 | Prefix proxy process to make it clear it belongs to radicle - Currently when debugging, esp. in scenarios where the proxy process is lingering, it's not always easy to spot that it belongs to radicle. A simple `radilce-` prefix would help to aid with that.
Ref. #1440 | infrastructure | prefix proxy process to make it clear it belongs to radicle currently when debugging esp in scenarios where the proxy process is lingering it s not always easy to spot that it belongs to radicle a simple radilce prefix would help to aid with that ref | 1 |
15,912 | 11,760,837,298 | IssuesEvent | 2020-03-13 20:30:05 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Intellisense is not working for vscode | area-Infrastructure-libraries untriaged | The new config system broke the intellisense for vscode. This could be verified from the omnisharp log
```
[info]: OmniSharp.WorkspaceInitializer
Configuration finished.
[info]: OmniSharp.Stdio.Host
Omnisharp server running using Stdio at location 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl' on host 6192.
[fail]: OmniSharp.MSBuild.ProjectLoader
The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[warn]: OmniSharp.MSBuild.ProjectManager
Failed to load project file 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl\tests\System.IO.FileSystem.AccessControl.Tests.csproj'.
d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\tests\System.IO.FileSystem.AccessControl.Tests.csproj
C:\Program Files\dotnet\sdk\3.1.200-preview-014971\Sdks\Microsoft.NET.Sdk\targets\Microsoft.NET.TargetFrameworkInference.targets(93,5): Error: The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[info]: OmniSharp.MSBuild.ProjectManager
Loading project: d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\src\System.IO.FileSystem.AccessControl.csproj
[fail]: OmniSharp.MSBuild.ProjectLoader
The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[warn]: OmniSharp.MSBuild.ProjectManager
Failed to load project file 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl\src\System.IO.FileSystem.AccessControl.csproj'.
d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\src\System.IO.FileSystem.AccessControl.csproj
C:\Program Files\dotnet\sdk\3.1.200-preview-014971\Sdks\Microsoft.NET.Sdk\targets\Microsoft.NET.TargetFrameworkInference.targets(93,5): Error: The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[info]: OmniSharp.MSBuild.ProjectManager
Loading project: d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\ref\System.IO.FileSystem.AccessControl.csproj
[fail]: OmniSharp.MSBuild.ProjectLoader
The "ChooseBestP2PTargetFrameworkTask" task was not given a value for the required parameter "TargetFramework".
[warn]: OmniSharp.MSBuild.ProjectManager
Failed to load project file 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl\ref\System.IO.FileSystem.AccessControl.csproj'.
d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\ref\System.IO.FileSystem.AccessControl.csproj
C:\Users\calope\.nuget\packages\microsoft.dotnet.build.tasks.targetframework.sdk\5.0.0-beta.20124.3\build\Microsoft.DotNet.Build.Tasks.TargetFramework.Sdk.targets(76,5): Error: The "ChooseBestP2PTargetFrameworkTask" task was not given a value for the required parameter "TargetFramework".
```
cc @ericstj @ViktorHofer @safern @danmosemsft @carlossanlop @danmosemsft | 1.0 | Intellisense is not working for vscode - The new config system broke the intellisense for vscode. This could be verified from the omnisharp log
```
[info]: OmniSharp.WorkspaceInitializer
Configuration finished.
[info]: OmniSharp.Stdio.Host
Omnisharp server running using Stdio at location 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl' on host 6192.
[fail]: OmniSharp.MSBuild.ProjectLoader
The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[warn]: OmniSharp.MSBuild.ProjectManager
Failed to load project file 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl\tests\System.IO.FileSystem.AccessControl.Tests.csproj'.
d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\tests\System.IO.FileSystem.AccessControl.Tests.csproj
C:\Program Files\dotnet\sdk\3.1.200-preview-014971\Sdks\Microsoft.NET.Sdk\targets\Microsoft.NET.TargetFrameworkInference.targets(93,5): Error: The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[info]: OmniSharp.MSBuild.ProjectManager
Loading project: d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\src\System.IO.FileSystem.AccessControl.csproj
[fail]: OmniSharp.MSBuild.ProjectLoader
The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[warn]: OmniSharp.MSBuild.ProjectManager
Failed to load project file 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl\src\System.IO.FileSystem.AccessControl.csproj'.
d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\src\System.IO.FileSystem.AccessControl.csproj
C:\Program Files\dotnet\sdk\3.1.200-preview-014971\Sdks\Microsoft.NET.Sdk\targets\Microsoft.NET.TargetFrameworkInference.targets(93,5): Error: The TargetFramework value 'netcoreapp5.0-Windows_NT' was not recognized. It may be misspelled. If not, then the TargetFrameworkIdentifier and/or TargetFrameworkVersion properties must be specified explicitly.
[info]: OmniSharp.MSBuild.ProjectManager
Loading project: d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\ref\System.IO.FileSystem.AccessControl.csproj
[fail]: OmniSharp.MSBuild.ProjectLoader
The "ChooseBestP2PTargetFrameworkTask" task was not given a value for the required parameter "TargetFramework".
[warn]: OmniSharp.MSBuild.ProjectManager
Failed to load project file 'd:\runtime\src\libraries\System.IO.FileSystem.AccessControl\ref\System.IO.FileSystem.AccessControl.csproj'.
d:\runtime\src\libraries\System.IO.FileSystem.AccessControl\ref\System.IO.FileSystem.AccessControl.csproj
C:\Users\calope\.nuget\packages\microsoft.dotnet.build.tasks.targetframework.sdk\5.0.0-beta.20124.3\build\Microsoft.DotNet.Build.Tasks.TargetFramework.Sdk.targets(76,5): Error: The "ChooseBestP2PTargetFrameworkTask" task was not given a value for the required parameter "TargetFramework".
```
cc @ericstj @ViktorHofer @safern @danmosemsft @carlossanlop @danmosemsft | infrastructure | intellisense is not working for vscode the new config system broke the intellisense for vscode this could be verified from the omnisharp log omnisharp workspaceinitializer configuration finished omnisharp stdio host omnisharp server running using stdio at location d runtime src libraries system io filesystem accesscontrol on host omnisharp msbuild projectloader the targetframework value windows nt was not recognized it may be misspelled if not then the targetframeworkidentifier and or targetframeworkversion properties must be specified explicitly omnisharp msbuild projectmanager failed to load project file d runtime src libraries system io filesystem accesscontrol tests system io filesystem accesscontrol tests csproj d runtime src libraries system io filesystem accesscontrol tests system io filesystem accesscontrol tests csproj c program files dotnet sdk preview sdks microsoft net sdk targets microsoft net targetframeworkinference targets error the targetframework value windows nt was not recognized it may be misspelled if not then the targetframeworkidentifier and or targetframeworkversion properties must be specified explicitly omnisharp msbuild projectmanager loading project d runtime src libraries system io filesystem accesscontrol src system io filesystem accesscontrol csproj omnisharp msbuild projectloader the targetframework value windows nt was not recognized it may be misspelled if not then the targetframeworkidentifier and or targetframeworkversion properties must be specified explicitly omnisharp msbuild projectmanager failed to load project file d runtime src libraries system io filesystem accesscontrol src system io filesystem accesscontrol csproj d runtime src libraries system io filesystem accesscontrol src system io filesystem accesscontrol csproj c program files dotnet sdk preview sdks microsoft net sdk targets microsoft net targetframeworkinference targets error the targetframework value windows nt was not recognized it may be misspelled if not then the targetframeworkidentifier and or targetframeworkversion properties must be specified explicitly omnisharp msbuild projectmanager loading project d runtime src libraries system io filesystem accesscontrol ref system io filesystem accesscontrol csproj omnisharp msbuild projectloader the task was not given a value for the required parameter targetframework omnisharp msbuild projectmanager failed to load project file d runtime src libraries system io filesystem accesscontrol ref system io filesystem accesscontrol csproj d runtime src libraries system io filesystem accesscontrol ref system io filesystem accesscontrol csproj c users calope nuget packages microsoft dotnet build tasks targetframework sdk beta build microsoft dotnet build tasks targetframework sdk targets error the task was not given a value for the required parameter targetframework cc ericstj viktorhofer safern danmosemsft carlossanlop danmosemsft | 1 |
520,041 | 15,077,759,157 | IssuesEvent | 2021-02-05 07:34:08 | wso2/cellery | https://api.github.com/repos/wso2/cellery | closed | Support for Cron jobs, and more advanced configurations of Jobs. | Priority/High Resolution/Won’t Fix Severity/Major Type/Improvement | **Description:**
Currently we have a basic support for non-parallel jobs in cellery, where if a component is defined without any ingresses, then it's deployed as jobs by default. But we haven't addressed all different usecases of jobs yet, including job specific configurations such as as parallelism, active dead line, ttl after finished, etc. Though developers can include the cron job related tasks within the application code it self, it will require the job to run continuously. These can be eliminated with cron jobs, therefore I think it's convenient to deploy cron jobs for the developers in such cases.
We need to design how to support more advanced configurations of jobs and cron jobs. | 1.0 | Support for Cron jobs, and more advanced configurations of Jobs. - **Description:**
Currently we have a basic support for non-parallel jobs in cellery, where if a component is defined without any ingresses, then it's deployed as jobs by default. But we haven't addressed all different usecases of jobs yet, including job specific configurations such as as parallelism, active dead line, ttl after finished, etc. Though developers can include the cron job related tasks within the application code it self, it will require the job to run continuously. These can be eliminated with cron jobs, therefore I think it's convenient to deploy cron jobs for the developers in such cases.
We need to design how to support more advanced configurations of jobs and cron jobs. | non_infrastructure | support for cron jobs and more advanced configurations of jobs description currently we have a basic support for non parallel jobs in cellery where if a component is defined without any ingresses then it s deployed as jobs by default but we haven t addressed all different usecases of jobs yet including job specific configurations such as as parallelism active dead line ttl after finished etc though developers can include the cron job related tasks within the application code it self it will require the job to run continuously these can be eliminated with cron jobs therefore i think it s convenient to deploy cron jobs for the developers in such cases we need to design how to support more advanced configurations of jobs and cron jobs | 0 |
30,772 | 25,062,859,582 | IssuesEvent | 2022-11-07 04:38:08 | zer0Kerbal/PapaKerballinisPizza | https://api.github.com/repos/zer0Kerbal/PapaKerballinisPizza | closed | Update License | issue: license type: infrastructure | # Update License
<!--
updateLicense v1.0.0.0
created: 17 Aug 2022
updated:
-->
* Update License
* [ ] Updated License: {newLicense}
* was: {CC BY 4.0}
this file: This file: All Rights Reserved by zer0Kerbal
| 1.0 | Update License - # Update License
<!--
updateLicense v1.0.0.0
created: 17 Aug 2022
updated:
-->
* Update License
* [ ] Updated License: {newLicense}
* was: {CC BY 4.0}
this file: This file: All Rights Reserved by zer0Kerbal
| infrastructure | update license update license updatelicense created aug updated update license updated license newlicense was cc by this file this file all rights reserved by | 1 |
6,812 | 6,620,964,962 | IssuesEvent | 2017-09-21 17:21:31 | trusktr/infamous | https://api.github.com/repos/trusktr/infamous | opened | Convert to TypeScript | enhancement infrastructure | I've loved the pure-js approach, but I think TypeScript would offer a level of sanity checking as the code base grows. Let's investigate adding TypeScript.
Benefits:
- Tools like UML diagram generators: https://github.com/fsahmad/typescript-uml/issues/16
- Checking of types before runtime, preventing some errors
Downsides:
- Having to write or find type definitions for existing libraries, and updating the definition if we update library version. Here's [types for Three.js](https://www.npmjs.com/package/@types/three)
- Slows down development initially (but once we have our types all set up and devs up to speed, it should increase productivity) | 1.0 | Convert to TypeScript - I've loved the pure-js approach, but I think TypeScript would offer a level of sanity checking as the code base grows. Let's investigate adding TypeScript.
Benefits:
- Tools like UML diagram generators: https://github.com/fsahmad/typescript-uml/issues/16
- Checking of types before runtime, preventing some errors
Downsides:
- Having to write or find type definitions for existing libraries, and updating the definition if we update library version. Here's [types for Three.js](https://www.npmjs.com/package/@types/three)
- Slows down development initially (but once we have our types all set up and devs up to speed, it should increase productivity) | infrastructure | convert to typescript i ve loved the pure js approach but i think typescript would offer a level of sanity checking as the code base grows let s investigate adding typescript benefits tools like uml diagram generators checking of types before runtime preventing some errors downsides having to write or find type definitions for existing libraries and updating the definition if we update library version here s slows down development initially but once we have our types all set up and devs up to speed it should increase productivity | 1 |
19,986 | 13,616,197,849 | IssuesEvent | 2020-09-23 15:20:09 | patternfly/patternfly-org | https://api.github.com/repos/patternfly/patternfly-org | closed | Provide guidance around specific pull request description formatting for breaking change notes | infrastructure | We'd like to extract breaking change notes related to developer impact from pull requests. Please provide a specific format that can be used to pull text from the description of a PR (wrapping in tildes for example - ~~Rename pf-var-one to pf-var-two~~). This should allow us to run a script against github pr's regularly during breaking change release cycles to provide a digest of information for consumers that might want to be proactively keeping up. | 1.0 | Provide guidance around specific pull request description formatting for breaking change notes - We'd like to extract breaking change notes related to developer impact from pull requests. Please provide a specific format that can be used to pull text from the description of a PR (wrapping in tildes for example - ~~Rename pf-var-one to pf-var-two~~). This should allow us to run a script against github pr's regularly during breaking change release cycles to provide a digest of information for consumers that might want to be proactively keeping up. | infrastructure | provide guidance around specific pull request description formatting for breaking change notes we d like to extract breaking change notes related to developer impact from pull requests please provide a specific format that can be used to pull text from the description of a pr wrapping in tildes for example rename pf var one to pf var two this should allow us to run a script against github pr s regularly during breaking change release cycles to provide a digest of information for consumers that might want to be proactively keeping up | 1 |
345,662 | 10,371,171,911 | IssuesEvent | 2019-09-08 18:19:51 | raceintospace/raceintospace | https://api.github.com/repos/raceintospace/raceintospace | opened | Escape out of Prestige First video | Low Priority enhancement | When the game shows you a prestige first that your opponent just accomplished, you can advance to the next clip by pressing a key. I think it would also be good if the movie could be canceled by hitting Escape. Sometimes you just don't care to watch it. | 1.0 | Escape out of Prestige First video - When the game shows you a prestige first that your opponent just accomplished, you can advance to the next clip by pressing a key. I think it would also be good if the movie could be canceled by hitting Escape. Sometimes you just don't care to watch it. | non_infrastructure | escape out of prestige first video when the game shows you a prestige first that your opponent just accomplished you can advance to the next clip by pressing a key i think it would also be good if the movie could be canceled by hitting escape sometimes you just don t care to watch it | 0 |
4,729 | 5,246,541,381 | IssuesEvent | 2017-02-01 09:57:14 | lukewyman/ecommerce | https://api.github.com/repos/lukewyman/ecommerce | closed | Create Boot (main class) and Service Support for Orchestrator | enhancement infrastructure | Currently, the Orchestrator doesn't have the neccessary infrastructture to "fire up" the service
Depends on: #9
Requirements:
- Create Boot class
- Create application.conf
- Create OrchestratorServiceSupport | 1.0 | Create Boot (main class) and Service Support for Orchestrator - Currently, the Orchestrator doesn't have the neccessary infrastructture to "fire up" the service
Depends on: #9
Requirements:
- Create Boot class
- Create application.conf
- Create OrchestratorServiceSupport | infrastructure | create boot main class and service support for orchestrator currently the orchestrator doesn t have the neccessary infrastructture to fire up the service depends on requirements create boot class create application conf create orchestratorservicesupport | 1 |
31,557 | 25,868,502,568 | IssuesEvent | 2022-12-13 23:28:14 | cal-itp/eligibility-server | https://api.github.com/repos/cal-itp/eligibility-server | closed | Restrict server to respond only to a list of IP addresses | security infrastructure | ## Options
### IP restrictions
This could be done via `ip_restrictions` on the App Service. See #181 for the work done so far on this.
Theoretically, we should be able to use the values from [outbound_ip_address_list](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/linux_web_app#outbound_ip_address_list) for the Benefits app service, but this was not working so we ended up [removing the variable value](https://github.com/cal-itp/eligibility-server/issues/189) for the initial launch.
### Front door
Do some research to see if using a [front door](https://azure.microsoft.com/en-us/products/frontdoor/) on eligibility-server would simplify the networking details for IP restriction. Also see if the Front door provides other security measures that we'd like to have.
| 1.0 | Restrict server to respond only to a list of IP addresses - ## Options
### IP restrictions
This could be done via `ip_restrictions` on the App Service. See #181 for the work done so far on this.
Theoretically, we should be able to use the values from [outbound_ip_address_list](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/linux_web_app#outbound_ip_address_list) for the Benefits app service, but this was not working so we ended up [removing the variable value](https://github.com/cal-itp/eligibility-server/issues/189) for the initial launch.
### Front door
Do some research to see if using a [front door](https://azure.microsoft.com/en-us/products/frontdoor/) on eligibility-server would simplify the networking details for IP restriction. Also see if the Front door provides other security measures that we'd like to have.
| infrastructure | restrict server to respond only to a list of ip addresses options ip restrictions this could be done via ip restrictions on the app service see for the work done so far on this theoretically we should be able to use the values from for the benefits app service but this was not working so we ended up for the initial launch front door do some research to see if using a on eligibility server would simplify the networking details for ip restriction also see if the front door provides other security measures that we d like to have | 1 |
198,894 | 15,012,636,310 | IssuesEvent | 2021-02-01 02:10:16 | AleoHQ/aleo-setup | https://api.github.com/repos/AleoHQ/aleo-setup | closed | Test if coordinator replaces all dropped participants and completes round correctly | test | Write a test to ensure that the coordinator correctly executes the drop logic for failing participants and recovers to finish the round of contributions. | 1.0 | Test if coordinator replaces all dropped participants and completes round correctly - Write a test to ensure that the coordinator correctly executes the drop logic for failing participants and recovers to finish the round of contributions. | non_infrastructure | test if coordinator replaces all dropped participants and completes round correctly write a test to ensure that the coordinator correctly executes the drop logic for failing participants and recovers to finish the round of contributions | 0 |
9,673 | 8,074,058,718 | IssuesEvent | 2018-08-06 21:30:38 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Several build issues around referencing MS.VS.Text.Internal.dll | Area-Infrastructure | - [ ] This line should have backslashes and not forward slashes: https://github.com/dotnet/roslyn/blob/08a22f8557a7eecbf1862bdcd92d25a9b1ca3b4e/build/Targets/Versions.props#L61
- [ ] 6 Test projects at EditorFeatures level use this line to reference MS.VS.Text.Internal: https://github.com/dotnet/roslyn/blob/08a22f8557a7eecbf1862bdcd92d25a9b1ca3b4e/src/EditorFeatures/VisualBasicTest/BasicEditorServicesTest.vbproj#L61-L64
- [ ] The Editor's NuGet package for MS.VS.Text.Internal has an empty ref folder - the Editor needs to fix that.
Instead of referencing the .dll by specifying the full path in Include as above, here's what it should be doing instead:
```
<Reference Include="Microsoft.VisualStudio.Text.Internal">
<HintPath>$(NuGetPackageRoot)\microsoft.visualstudio.text.internal\$(MicrosoftVisualStudioTextInternalVersion)\lib\net46\Microsoft.VisualStudio.Text.Internal.dll</HintPath>
</Reference>
``` | 1.0 | Several build issues around referencing MS.VS.Text.Internal.dll - - [ ] This line should have backslashes and not forward slashes: https://github.com/dotnet/roslyn/blob/08a22f8557a7eecbf1862bdcd92d25a9b1ca3b4e/build/Targets/Versions.props#L61
- [ ] 6 Test projects at EditorFeatures level use this line to reference MS.VS.Text.Internal: https://github.com/dotnet/roslyn/blob/08a22f8557a7eecbf1862bdcd92d25a9b1ca3b4e/src/EditorFeatures/VisualBasicTest/BasicEditorServicesTest.vbproj#L61-L64
- [ ] The Editor's NuGet package for MS.VS.Text.Internal has an empty ref folder - the Editor needs to fix that.
Instead of referencing the .dll by specifying the full path in Include as above, here's what it should be doing instead:
```
<Reference Include="Microsoft.VisualStudio.Text.Internal">
<HintPath>$(NuGetPackageRoot)\microsoft.visualstudio.text.internal\$(MicrosoftVisualStudioTextInternalVersion)\lib\net46\Microsoft.VisualStudio.Text.Internal.dll</HintPath>
</Reference>
``` | infrastructure | several build issues around referencing ms vs text internal dll this line should have backslashes and not forward slashes test projects at editorfeatures level use this line to reference ms vs text internal the editor s nuget package for ms vs text internal has an empty ref folder the editor needs to fix that instead of referencing the dll by specifying the full path in include as above here s what it should be doing instead nugetpackageroot microsoft visualstudio text internal microsoftvisualstudiotextinternalversion lib microsoft visualstudio text internal dll | 1 |
25,715 | 19,029,907,081 | IssuesEvent | 2021-11-24 09:34:47 | mwvgroup/Pitt-Google-Broker | https://api.github.com/repos/mwvgroup/Pitt-Google-Broker | opened | Automate Deployment | Cloud Infrastructure Pipeline: Admin Goal | ### Goal: Refactor deployment code and workflow to simplify and automate.
- [ ] #65
- [ ] #80
- [ ] Automate with Terraform | 1.0 | Automate Deployment - ### Goal: Refactor deployment code and workflow to simplify and automate.
- [ ] #65
- [ ] #80
- [ ] Automate with Terraform | infrastructure | automate deployment goal refactor deployment code and workflow to simplify and automate automate with terraform | 1 |
12,103 | 9,595,251,274 | IssuesEvent | 2019-05-09 15:38:36 | laopinion/siteLaopinion | https://api.github.com/repos/laopinion/siteLaopinion | closed | Create components, page, entries | Doing Infrastructure | # Create components, page and entries
Se mencionan component page y entries mas importantes.
## Components
* Aperturas.js
* Header.js
* Footer.js
* Loading.js
* MasGalerias.js
* Notas.js
* VideosGaleria.js
## Pages
* Article.js
* Galeria.js
* Video.js
* Infografia.js
* Podcast.js
> Separar los components de **stateless** of the **stateful** se deben diferenciar por llevar container in the name of component.
## Entries
* Article.js
* Galeria.js
* Video.js
* Infografia.js
* Podcast.js
## Components Stateful
* ArticleContainer.js
* VideoContainer.js
> No estan todos los components mensionados.
## Stateless vs Stateful Components
### Stateful
```javascript
import React from "react";
import Show from "./show";
class Input extends React.Component {
constructor(props) { //<----Method
super(props);
this.state = { //<----Initialize state
value: ""
};
this.handleChange = this.handleChange.bind(this);
}
handleChange(event) { //<---- Method/Set State
this.setState({ value: event.target.value });
}
render() { . //<-----Method/Return JSX/HTML
return (
<div>
<input
name="firstName"
type="text"
value={this.state.value}
onChange={this.handleChange}
/>
<Show value={this.state.value} />
</div>
);
}
}
export default Input;
```
### Stateless
```javascript
import React from "react";
function Show(props) {
return( //<----Return JSX/HTML
<p>{props.value}</p>
)
}
export default Show;
```
| 1.0 | Create components, page, entries - # Create components, page and entries
Se mencionan component page y entries mas importantes.
## Components
* Aperturas.js
* Header.js
* Footer.js
* Loading.js
* MasGalerias.js
* Notas.js
* VideosGaleria.js
## Pages
* Article.js
* Galeria.js
* Video.js
* Infografia.js
* Podcast.js
> Separar los components de **stateless** of the **stateful** se deben diferenciar por llevar container in the name of component.
## Entries
* Article.js
* Galeria.js
* Video.js
* Infografia.js
* Podcast.js
## Components Stateful
* ArticleContainer.js
* VideoContainer.js
> No estan todos los components mensionados.
## Stateless vs Stateful Components
### Stateful
```javascript
import React from "react";
import Show from "./show";
class Input extends React.Component {
constructor(props) { //<----Method
super(props);
this.state = { //<----Initialize state
value: ""
};
this.handleChange = this.handleChange.bind(this);
}
handleChange(event) { //<---- Method/Set State
this.setState({ value: event.target.value });
}
render() { . //<-----Method/Return JSX/HTML
return (
<div>
<input
name="firstName"
type="text"
value={this.state.value}
onChange={this.handleChange}
/>
<Show value={this.state.value} />
</div>
);
}
}
export default Input;
```
### Stateless
```javascript
import React from "react";
function Show(props) {
return( //<----Return JSX/HTML
<p>{props.value}</p>
)
}
export default Show;
```
| infrastructure | create components page entries create components page and entries se mencionan component page y entries mas importantes components aperturas js header js footer js loading js masgalerias js notas js videosgaleria js pages article js galeria js video js infografia js podcast js separar los components de stateless of the stateful se deben diferenciar por llevar container in the name of component entries article js galeria js video js infografia js podcast js components stateful articlecontainer js videocontainer js no estan todos los components mensionados stateless vs stateful components stateful javascript import react from react import show from show class input extends react component constructor props method super props this state initialize state value this handlechange this handlechange bind this handlechange event method set state this setstate value event target value render method return jsx html return input name firstname type text value this state value onchange this handlechange export default input stateless javascript import react from react function show props return return jsx html props value export default show | 1 |
357,499 | 10,607,520,209 | IssuesEvent | 2019-10-11 04:09:50 | kubeflow/katib | https://api.github.com/repos/kubeflow/katib | closed | vizier-db needs enhanced liveness/readiness checks | priority/p1 | vizier-db pod is shown as "Running" after deployment but it actually is not fully configured if there was a problem with the katib-mysql PV not being writable.
OpenShift 3.11
```
$ oc logs vizier-db-7b4475c695-7vwt9
2019-01-21T14:33:28.470292Z 27562 [Note] Access denied for user 'UNKNOWN_USER'@'localhost' (using password: YES)
2019-01-21T14:33:30.469526Z 27563 [Note] Access denied for user 'UNKNOWN_USER'@'localhost' (using password: YES)
2019-01-21T14:33:32.471171Z 27564 [Note] Access denied for user 'UNKNOWN_USER'@'localhost' (using password: YES)
``` | 1.0 | vizier-db needs enhanced liveness/readiness checks - vizier-db pod is shown as "Running" after deployment but it actually is not fully configured if there was a problem with the katib-mysql PV not being writable.
OpenShift 3.11
```
$ oc logs vizier-db-7b4475c695-7vwt9
2019-01-21T14:33:28.470292Z 27562 [Note] Access denied for user 'UNKNOWN_USER'@'localhost' (using password: YES)
2019-01-21T14:33:30.469526Z 27563 [Note] Access denied for user 'UNKNOWN_USER'@'localhost' (using password: YES)
2019-01-21T14:33:32.471171Z 27564 [Note] Access denied for user 'UNKNOWN_USER'@'localhost' (using password: YES)
``` | non_infrastructure | vizier db needs enhanced liveness readiness checks vizier db pod is shown as running after deployment but it actually is not fully configured if there was a problem with the katib mysql pv not being writable openshift oc logs vizier db access denied for user unknown user localhost using password yes access denied for user unknown user localhost using password yes access denied for user unknown user localhost using password yes | 0 |
379,914 | 11,244,182,590 | IssuesEvent | 2020-01-10 06:15:26 | wso2/product-microgateway | https://api.github.com/repos/wso2/product-microgateway | opened | Not able to redirect a request to an endpoint using interceptor | Priority/Normal Type/Bug | ### Description:
We are using open API definition and interceptors to make request to dedicated backend but we are unable to make request to an endpoint other than that mentioned in the open API definition.
Is there a way to update the requested url through the Request object in the interceptor so that the end URL that is being hit is different from that mentioned in the open API definition.
### Steps to reproduce:
Attaching code:
OPEN API :
x-wso2-basePath: /Mobiquity/1.0.0
x-wso2-request-interceptor: validateGlobalRequest
x-wso2-response-interceptor: validateGlobalResponse
x-wso2-production-endpoints:
urls:
- http://localhost:7099/Mobiquity/
paths:
"/callJigsaw/V2":
post:
tags:
- callJigsaw
summary: check for update
description: callJigsaw
operationId: callJigsaw
x-wso2-request-interceptor: validateRequestDemo
x-wso2-response-interceptor: validateResponse
responses:
'200':
description: successful operation
'400':
description: Invalid status value
x-wso2-disable-security: true
"/callService/V2":
post:
tags:
- callService
summary: callService
description: callService
operationId: callService
x-wso2-request-interceptor: validateRequestDemo
x-wso2-response-interceptor: validateResponse
responses:
'200':
description: successful operation
'400':
description: Invalid status value
x-wso2-disable-security: true
For example considering the above open API : if call comes to /Mobiquity/1.0.0/callJigsaw/V2 then the endpoint that is hit is http://localhost:7099/Mobiquity/callJigsaw/V2 but what i want is the endpoint should be changed dynamically to http://localhost:7099/Mobiquity/callService/V2 or may be any url that i have written in the **interceptor** through the request object
public function validateRequest (http:Caller outboundEp, http:Request req) {
// Backend requires X-API-KEY header. No point in passing the request to the backend
// if the header is not present in the request.
boolean hasApiKey = req.hasHeader("token");
if (!hasApiKey) {
http:Response res = new;
res.statusCode = 400;
json message ={"serviceResponse":{"COMMAND":{"TYPE":"GENERAL","TXNSTATUS":"400","MESSAGE":"Missing required header","TXNMESSAGE":"Missing required header"}}};
res.setPayload(message);
var status = outboundEp->respond(res);
} else {
//set your url here that should ovveride the url defined in the callJigsaw path
req.setUrl("http://your own url");
}
}
### Affected Product Version:
<!-- Members can use Affected/*** labels -->
3.0.2
### Environment details (with versions):
- OS: windows
- Client:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| 1.0 | Not able to redirect a request to an endpoint using interceptor - ### Description:
We are using open API definition and interceptors to make request to dedicated backend but we are unable to make request to an endpoint other than that mentioned in the open API definition.
Is there a way to update the requested url through the Request object in the interceptor so that the end URL that is being hit is different from that mentioned in the open API definition.
### Steps to reproduce:
Attaching code:
OPEN API :
x-wso2-basePath: /Mobiquity/1.0.0
x-wso2-request-interceptor: validateGlobalRequest
x-wso2-response-interceptor: validateGlobalResponse
x-wso2-production-endpoints:
urls:
- http://localhost:7099/Mobiquity/
paths:
"/callJigsaw/V2":
post:
tags:
- callJigsaw
summary: check for update
description: callJigsaw
operationId: callJigsaw
x-wso2-request-interceptor: validateRequestDemo
x-wso2-response-interceptor: validateResponse
responses:
'200':
description: successful operation
'400':
description: Invalid status value
x-wso2-disable-security: true
"/callService/V2":
post:
tags:
- callService
summary: callService
description: callService
operationId: callService
x-wso2-request-interceptor: validateRequestDemo
x-wso2-response-interceptor: validateResponse
responses:
'200':
description: successful operation
'400':
description: Invalid status value
x-wso2-disable-security: true
For example considering the above open API : if call comes to /Mobiquity/1.0.0/callJigsaw/V2 then the endpoint that is hit is http://localhost:7099/Mobiquity/callJigsaw/V2 but what i want is the endpoint should be changed dynamically to http://localhost:7099/Mobiquity/callService/V2 or may be any url that i have written in the **interceptor** through the request object
public function validateRequest (http:Caller outboundEp, http:Request req) {
// Backend requires X-API-KEY header. No point in passing the request to the backend
// if the header is not present in the request.
boolean hasApiKey = req.hasHeader("token");
if (!hasApiKey) {
http:Response res = new;
res.statusCode = 400;
json message ={"serviceResponse":{"COMMAND":{"TYPE":"GENERAL","TXNSTATUS":"400","MESSAGE":"Missing required header","TXNMESSAGE":"Missing required header"}}};
res.setPayload(message);
var status = outboundEp->respond(res);
} else {
//set your url here that should ovveride the url defined in the callJigsaw path
req.setUrl("http://your own url");
}
}
### Affected Product Version:
<!-- Members can use Affected/*** labels -->
3.0.2
### Environment details (with versions):
- OS: windows
- Client:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| non_infrastructure | not able to redirect a request to an endpoint using interceptor description we are using open api definition and interceptors to make request to dedicated backend but we are unable to make request to an endpoint other than that mentioned in the open api definition is there a way to update the requested url through the request object in the interceptor so that the end url that is being hit is different from that mentioned in the open api definition steps to reproduce attaching code open api x basepath mobiquity x request interceptor validateglobalrequest x response interceptor validateglobalresponse x production endpoints urls paths calljigsaw post tags calljigsaw summary check for update description calljigsaw operationid calljigsaw x request interceptor validaterequestdemo x response interceptor validateresponse responses description successful operation description invalid status value x disable security true callservice post tags callservice summary callservice description callservice operationid callservice x request interceptor validaterequestdemo x response interceptor validateresponse responses description successful operation description invalid status value x disable security true for example considering the above open api if call comes to mobiquity calljigsaw then the endpoint that is hit is but what i want is the endpoint should be changed dynamically to or may be any url that i have written in the interceptor through the request object public function validaterequest http caller outboundep http request req backend requires x api key header no point in passing the request to the backend if the header is not present in the request boolean hasapikey req hasheader token if hasapikey http response res new res statuscode json message serviceresponse command type general txnstatus message missing required header txnmessage missing required header res setpayload message var status outboundep respond res else set your url here that should ovveride the url defined in the calljigsaw path req seturl own url affected product version environment details with versions os windows client env docker optional fields related issues suggested labels suggested assignees | 0 |
13,620 | 10,348,947,382 | IssuesEvent | 2019-09-04 21:03:17 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | VisualBasicCompletionCommandHandlerTests.TestBangFiltersInDocComment | Area-Infrastructure Test | Failures in PR validation [Test Logs](https://dev.azure.com/dnceng/public/_build/results?buildId=328771&view=ms.vss-test-web.build-test-results-tab&runId=9620558&resultId=100452&paneView=debug)
```
Error message
AssertCompletionSession\r\nExpected: True\r\nActual: False
Stack trace
at Microsoft.CodeAnalysis.Editor.UnitTests.IntelliSense.ModernCompletionTestState.VB$StateMachine_22_AssertCompletionSession.MoveNext() in /_/src/EditorFeatures/TestUtilities2/Intellisense/ModernCompletionTestState.vb:line 191
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at Microsoft.CodeAnalysis.Editor.UnitTests.IntelliSense.VisualBasicCompletionCommandHandlerTests.VB$StateMachine_59_TestBangFiltersInDocComment.MoveNext() in /_/src/EditorFeatures/Test2/IntelliSense/VisualBasicCompletionCommandHandlerTests.vb:line 1369
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
``` | 1.0 | VisualBasicCompletionCommandHandlerTests.TestBangFiltersInDocComment - Failures in PR validation [Test Logs](https://dev.azure.com/dnceng/public/_build/results?buildId=328771&view=ms.vss-test-web.build-test-results-tab&runId=9620558&resultId=100452&paneView=debug)
```
Error message
AssertCompletionSession\r\nExpected: True\r\nActual: False
Stack trace
at Microsoft.CodeAnalysis.Editor.UnitTests.IntelliSense.ModernCompletionTestState.VB$StateMachine_22_AssertCompletionSession.MoveNext() in /_/src/EditorFeatures/TestUtilities2/Intellisense/ModernCompletionTestState.vb:line 191
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.GetResult()
at Microsoft.CodeAnalysis.Editor.UnitTests.IntelliSense.VisualBasicCompletionCommandHandlerTests.VB$StateMachine_59_TestBangFiltersInDocComment.MoveNext() in /_/src/EditorFeatures/Test2/IntelliSense/VisualBasicCompletionCommandHandlerTests.vb:line 1369
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task)
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
``` | infrastructure | visualbasiccompletioncommandhandlertests testbangfiltersindoccomment failures in pr validation error message assertcompletionsession r nexpected true r nactual false stack trace at microsoft codeanalysis editor unittests intellisense moderncompletionteststate vb statemachine assertcompletionsession movenext in src editorfeatures intellisense moderncompletionteststate vb line end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices taskawaiter getresult at microsoft codeanalysis editor unittests intellisense visualbasiccompletioncommandhandlertests vb statemachine testbangfiltersindoccomment movenext in src editorfeatures intellisense visualbasiccompletioncommandhandlertests vb line end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task | 1 |
4,255 | 4,932,050,845 | IssuesEvent | 2016-11-28 12:20:56 | KCreate/charly-lang | https://api.github.com/repos/KCreate/charly-lang | opened | Auto-Update functionality via the CLI | Enhancement Infrastructure | You should be able to write `charly -f update` to update the charly installation to it's newest version.
It should use the `CHARLYDIR` environment variable to find the path to the git repository, do a git pull on the master branch and run the `install.sh` script. | 1.0 | Auto-Update functionality via the CLI - You should be able to write `charly -f update` to update the charly installation to it's newest version.
It should use the `CHARLYDIR` environment variable to find the path to the git repository, do a git pull on the master branch and run the `install.sh` script. | infrastructure | auto update functionality via the cli you should be able to write charly f update to update the charly installation to it s newest version it should use the charlydir environment variable to find the path to the git repository do a git pull on the master branch and run the install sh script | 1 |
245,040 | 26,503,781,721 | IssuesEvent | 2023-01-18 12:21:15 | RG4421/multi-juicer | https://api.github.com/repos/RG4421/multi-juicer | opened | CVE-2022-25901 (Medium) detected in cookiejar-2.1.2.tgz | security vulnerability | ## CVE-2022-25901 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookiejar-2.1.2.tgz</b></p></summary>
<p>simple persistent cookiejar system</p>
<p>Library home page: <a href="https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz">https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz</a></p>
<p>Path to dependency file: /juice-balancer/package.json</p>
<p>Path to vulnerable library: /juice-balancer/node_modules/cookiejar/package.json</p>
<p>
Dependency Hierarchy:
- supertest-6.1.3.tgz (Root Library)
- superagent-6.1.0.tgz
- :x: **cookiejar-2.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/multi-juicer/commit/8576a853ef65531c776b5ea4aac618580e1b0354">8576a853ef65531c776b5ea4aac618580e1b0354</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package cookiejar before 2.1.4 are vulnerable to Regular Expression Denial of Service (ReDoS) via the Cookie.parse function, which uses an insecure regular expression.
<p>Publish Date: 2023-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25901>CVE-2022-25901</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-01-18</p>
<p>Fix Resolution: cookiejar - 2.1.4</p>
</p>
</details>
<p></p>
| True | CVE-2022-25901 (Medium) detected in cookiejar-2.1.2.tgz - ## CVE-2022-25901 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cookiejar-2.1.2.tgz</b></p></summary>
<p>simple persistent cookiejar system</p>
<p>Library home page: <a href="https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz">https://registry.npmjs.org/cookiejar/-/cookiejar-2.1.2.tgz</a></p>
<p>Path to dependency file: /juice-balancer/package.json</p>
<p>Path to vulnerable library: /juice-balancer/node_modules/cookiejar/package.json</p>
<p>
Dependency Hierarchy:
- supertest-6.1.3.tgz (Root Library)
- superagent-6.1.0.tgz
- :x: **cookiejar-2.1.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/multi-juicer/commit/8576a853ef65531c776b5ea4aac618580e1b0354">8576a853ef65531c776b5ea4aac618580e1b0354</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of the package cookiejar before 2.1.4 are vulnerable to Regular Expression Denial of Service (ReDoS) via the Cookie.parse function, which uses an insecure regular expression.
<p>Publish Date: 2023-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25901>CVE-2022-25901</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2023-01-18</p>
<p>Fix Resolution: cookiejar - 2.1.4</p>
</p>
</details>
<p></p>
| non_infrastructure | cve medium detected in cookiejar tgz cve medium severity vulnerability vulnerable library cookiejar tgz simple persistent cookiejar system library home page a href path to dependency file juice balancer package json path to vulnerable library juice balancer node modules cookiejar package json dependency hierarchy supertest tgz root library superagent tgz x cookiejar tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of the package cookiejar before are vulnerable to regular expression denial of service redos via the cookie parse function which uses an insecure regular expression publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution cookiejar | 0 |
32,618 | 26,833,718,660 | IssuesEvent | 2023-02-02 17:47:15 | GoogleCloudPlatform/solutions-template | https://api.github.com/repos/GoogleCloudPlatform/solutions-template | closed | error when deploying optional cloudrun application | Infrastructure Tools/Scripts | when i uncomment the optional cloudrun service in main.tf (./terraform/environments/dev/main.tf), and run terraform apply, i get the following error:
```
module.cloudrun-sample.null_resource.deploy-cloudrun-image: Creation complete after 2m34s [id=6824479577047599322]
╷
│ Error: Error applying IAM policy for cloudrun service "v1/projects/solution1-373714/locations/us-central1/services/cloudrun-sample": Error setting IAM policy for cloudrun service "v1/projects/solution1-373714/locations/us-central1/services/cloudrun-sample": googleapi: Error 404: Resource 'cloudrun-sample' of kind 'SERVICE' in region 'us-central1' in project 'solution1-373714' does not exist.
│
│ with module.cloudrun-sample.google_cloud_run_service_iam_member.member[0],
│ on ../../modules/cloudrun/main.tf line 12, in resource "google_cloud_run_service_iam_member" "member":
│ 12: resource "google_cloud_run_service_iam_member" "member" {
│
╵
```
note that the service does deploy and is functional, but the error still exists... not sure why.
thanks
k | 1.0 | error when deploying optional cloudrun application - when i uncomment the optional cloudrun service in main.tf (./terraform/environments/dev/main.tf), and run terraform apply, i get the following error:
```
module.cloudrun-sample.null_resource.deploy-cloudrun-image: Creation complete after 2m34s [id=6824479577047599322]
╷
│ Error: Error applying IAM policy for cloudrun service "v1/projects/solution1-373714/locations/us-central1/services/cloudrun-sample": Error setting IAM policy for cloudrun service "v1/projects/solution1-373714/locations/us-central1/services/cloudrun-sample": googleapi: Error 404: Resource 'cloudrun-sample' of kind 'SERVICE' in region 'us-central1' in project 'solution1-373714' does not exist.
│
│ with module.cloudrun-sample.google_cloud_run_service_iam_member.member[0],
│ on ../../modules/cloudrun/main.tf line 12, in resource "google_cloud_run_service_iam_member" "member":
│ 12: resource "google_cloud_run_service_iam_member" "member" {
│
╵
```
note that the service does deploy and is functional, but the error still exists... not sure why.
thanks
k | infrastructure | error when deploying optional cloudrun application when i uncomment the optional cloudrun service in main tf terraform environments dev main tf and run terraform apply i get the following error module cloudrun sample null resource deploy cloudrun image creation complete after ╷ │ error error applying iam policy for cloudrun service projects locations us services cloudrun sample error setting iam policy for cloudrun service projects locations us services cloudrun sample googleapi error resource cloudrun sample of kind service in region us in project does not exist │ │ with module cloudrun sample google cloud run service iam member member │ on modules cloudrun main tf line in resource google cloud run service iam member member │ resource google cloud run service iam member member │ ╵ note that the service does deploy and is functional but the error still exists not sure why thanks k | 1 |
26,917 | 20,871,211,654 | IssuesEvent | 2022-03-22 12:09:04 | mitodl/ol-infrastructure | https://api.github.com/repos/mitodl/ol-infrastructure | opened | Progamatically Create users for micromasters | DevOps Infrastructure | Use `odl-micromasters-ci-opensearch-ro' as a guide.
Write pulumi code somewhere? Does it go in the stack for opensearch or find another stack / location for semi-random things like this. | 1.0 | Progamatically Create users for micromasters - Use `odl-micromasters-ci-opensearch-ro' as a guide.
Write pulumi code somewhere? Does it go in the stack for opensearch or find another stack / location for semi-random things like this. | infrastructure | progamatically create users for micromasters use odl micromasters ci opensearch ro as a guide write pulumi code somewhere does it go in the stack for opensearch or find another stack location for semi random things like this | 1 |
239,223 | 7,787,548,108 | IssuesEvent | 2018-06-06 23:02:44 | googleapis/nodejs-spanner | https://api.github.com/repos/googleapis/nodejs-spanner | closed | Error: 2 UNKNOWN: Transaction outcome unknown | priority: p0 type: bug |
#### Environment details
- OS: CoreOS
- Node.js version: v8.9.4
- npm version: v6.0.1
- `@google-cloud/spanner` version: v1.4.1
#### Steps to reproduce
1. ???
2. ???
Not sure what causes this but these errors started trickling in this morning. There's been about ~75 throughout the course of the day. Mainly wondering what would cause this sort of error and how I can go about debugging it.
Error info:
```
meta: {
code: 2
details: "Transaction outcome unknown."
message: "2 UNKNOWN: Transaction outcome unknown."
metadata: {
_internal_repr: {
}
}
note: "Exception occurred in retry method that was not classified as transient"
stack: [
0: "Error: 2 UNKNOWN: Transaction outcome unknown."
1: "at Object.exports.createStatusError (/opt/app/node_modules/google-gax/node_modules/grpc/src/common.js:87:15)"
2: "at Object.onReceiveStatus (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:1214:28)"
3: "at InterceptingListener._callNext (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:590:42)"
4: "at InterceptingListener.onReceiveStatus (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:640:8)"
5: "at callback (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:867:24)"
6: ""
]
statusCode: 500
}
```
| 1.0 | Error: 2 UNKNOWN: Transaction outcome unknown -
#### Environment details
- OS: CoreOS
- Node.js version: v8.9.4
- npm version: v6.0.1
- `@google-cloud/spanner` version: v1.4.1
#### Steps to reproduce
1. ???
2. ???
Not sure what causes this but these errors started trickling in this morning. There's been about ~75 throughout the course of the day. Mainly wondering what would cause this sort of error and how I can go about debugging it.
Error info:
```
meta: {
code: 2
details: "Transaction outcome unknown."
message: "2 UNKNOWN: Transaction outcome unknown."
metadata: {
_internal_repr: {
}
}
note: "Exception occurred in retry method that was not classified as transient"
stack: [
0: "Error: 2 UNKNOWN: Transaction outcome unknown."
1: "at Object.exports.createStatusError (/opt/app/node_modules/google-gax/node_modules/grpc/src/common.js:87:15)"
2: "at Object.onReceiveStatus (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:1214:28)"
3: "at InterceptingListener._callNext (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:590:42)"
4: "at InterceptingListener.onReceiveStatus (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:640:8)"
5: "at callback (/opt/app/node_modules/google-gax/node_modules/grpc/src/client_interceptors.js:867:24)"
6: ""
]
statusCode: 500
}
```
| non_infrastructure | error unknown transaction outcome unknown environment details os coreos node js version npm version google cloud spanner version steps to reproduce not sure what causes this but these errors started trickling in this morning there s been about throughout the course of the day mainly wondering what would cause this sort of error and how i can go about debugging it error info meta code details transaction outcome unknown message unknown transaction outcome unknown metadata internal repr note exception occurred in retry method that was not classified as transient stack error unknown transaction outcome unknown at object exports createstatuserror opt app node modules google gax node modules grpc src common js at object onreceivestatus opt app node modules google gax node modules grpc src client interceptors js at interceptinglistener callnext opt app node modules google gax node modules grpc src client interceptors js at interceptinglistener onreceivestatus opt app node modules google gax node modules grpc src client interceptors js at callback opt app node modules google gax node modules grpc src client interceptors js statuscode | 0 |
33,191 | 27,293,967,915 | IssuesEvent | 2023-02-23 18:42:07 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | llvm-symbolizer invocation failure | area-Infrastructure-coreclr | Seen in https://github.com/dotnet/runtime/issues/82414
And here: https://dev.azure.com/dnceng-public/public/_build/results?buildId=181701&view=ms.vss-test-web.build-test-results-tab&runId=3542266&resultId=107033&paneView=debug
```
Test failed. Trying to see if dump file was created in /home/helixbot/dotnetbuild/dumps since 2/23/2023 8:33:19 AM
Processing /home/helixbot/dotnetbuild/dumps/coredump.12662.dmp.crashreport.json
Printing stacktrace from '/home/helixbot/dotnetbuild/dumps/coredump.12662.dmp.crashreport.json'
Invoking llvm-symbolizer --pretty-print
Errors while running llvm-symbolizer --pretty-print
System.ComponentModel.Win32Exception (2): An error occurred trying to start process 'llvm-symbolizer' with working directory '/root/helix/work/workitem/e'. No such file or directory
at System.Diagnostics.Process.ForkAndExecProcess(ProcessStartInfo startInfo, String resolvedFilename, String[] argv, String[] envp, String cwd, Boolean setCredentials, UInt32 userId, UInt32 groupId, UInt32[] groups, Int32& stdinFd, Int32& stdoutFd, Int32& stderrFd, Boolean usesTerminal, Boolean throwOnNoExec)
at System.Diagnostics.Process.StartCore(ProcessStartInfo startInfo)
at CoreclrTestLib.CoreclrTestWrapperLib.TryPrintStackTraceFromCrashReport(String crashReportJsonFile, StreamWriter outputWriter)
Test Harness Exitcode is : 1
```
Related: https://github.com/dotnet/runtime/issues/77918 | 1.0 | llvm-symbolizer invocation failure - Seen in https://github.com/dotnet/runtime/issues/82414
And here: https://dev.azure.com/dnceng-public/public/_build/results?buildId=181701&view=ms.vss-test-web.build-test-results-tab&runId=3542266&resultId=107033&paneView=debug
```
Test failed. Trying to see if dump file was created in /home/helixbot/dotnetbuild/dumps since 2/23/2023 8:33:19 AM
Processing /home/helixbot/dotnetbuild/dumps/coredump.12662.dmp.crashreport.json
Printing stacktrace from '/home/helixbot/dotnetbuild/dumps/coredump.12662.dmp.crashreport.json'
Invoking llvm-symbolizer --pretty-print
Errors while running llvm-symbolizer --pretty-print
System.ComponentModel.Win32Exception (2): An error occurred trying to start process 'llvm-symbolizer' with working directory '/root/helix/work/workitem/e'. No such file or directory
at System.Diagnostics.Process.ForkAndExecProcess(ProcessStartInfo startInfo, String resolvedFilename, String[] argv, String[] envp, String cwd, Boolean setCredentials, UInt32 userId, UInt32 groupId, UInt32[] groups, Int32& stdinFd, Int32& stdoutFd, Int32& stderrFd, Boolean usesTerminal, Boolean throwOnNoExec)
at System.Diagnostics.Process.StartCore(ProcessStartInfo startInfo)
at CoreclrTestLib.CoreclrTestWrapperLib.TryPrintStackTraceFromCrashReport(String crashReportJsonFile, StreamWriter outputWriter)
Test Harness Exitcode is : 1
```
Related: https://github.com/dotnet/runtime/issues/77918 | infrastructure | llvm symbolizer invocation failure seen in and here test failed trying to see if dump file was created in home helixbot dotnetbuild dumps since am processing home helixbot dotnetbuild dumps coredump dmp crashreport json printing stacktrace from home helixbot dotnetbuild dumps coredump dmp crashreport json invoking llvm symbolizer pretty print errors while running llvm symbolizer pretty print system componentmodel an error occurred trying to start process llvm symbolizer with working directory root helix work workitem e no such file or directory at system diagnostics process forkandexecprocess processstartinfo startinfo string resolvedfilename string argv string envp string cwd boolean setcredentials userid groupid groups stdinfd stdoutfd stderrfd boolean usesterminal boolean throwonnoexec at system diagnostics process startcore processstartinfo startinfo at coreclrtestlib coreclrtestwrapperlib tryprintstacktracefromcrashreport string crashreportjsonfile streamwriter outputwriter test harness exitcode is related | 1 |
1,221 | 3,082,769,958 | IssuesEvent | 2015-08-24 01:31:10 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | Add Servo's memory reporting to rustc | A-infrastructure A-instrumentation A-tools I-compiletime | This probably means pulling out Servo's stuff into a crate of its own (if that is possible and amenable to the Servo folks), then using that crate in rustc (either as a submodule or by just moving the reporting stuff into the rust distro). Adding the actual hooks into rustc shouldn't be too tricky, presumably we'd need to add a -Z flag and code to setup the system reporter and to call that, probably in the same places we use `time_passes`. Finer-grained memory reporting can come later.
We really need decent stats on compiler memory usage before we start doing big refactorings of the compiler.
One thing I'm not sure about is if we can do memory reporting at the end of a phase and get the peak memory usage during that phase. That seems to me to be the most useful statistic, but it doesn't look like we can get that - only the amount of memory currently allocated, which is probably minimal between phases.
cc @jdm @nnethercote | 1.0 | Add Servo's memory reporting to rustc - This probably means pulling out Servo's stuff into a crate of its own (if that is possible and amenable to the Servo folks), then using that crate in rustc (either as a submodule or by just moving the reporting stuff into the rust distro). Adding the actual hooks into rustc shouldn't be too tricky, presumably we'd need to add a -Z flag and code to setup the system reporter and to call that, probably in the same places we use `time_passes`. Finer-grained memory reporting can come later.
We really need decent stats on compiler memory usage before we start doing big refactorings of the compiler.
One thing I'm not sure about is if we can do memory reporting at the end of a phase and get the peak memory usage during that phase. That seems to me to be the most useful statistic, but it doesn't look like we can get that - only the amount of memory currently allocated, which is probably minimal between phases.
cc @jdm @nnethercote | infrastructure | add servo s memory reporting to rustc this probably means pulling out servo s stuff into a crate of its own if that is possible and amenable to the servo folks then using that crate in rustc either as a submodule or by just moving the reporting stuff into the rust distro adding the actual hooks into rustc shouldn t be too tricky presumably we d need to add a z flag and code to setup the system reporter and to call that probably in the same places we use time passes finer grained memory reporting can come later we really need decent stats on compiler memory usage before we start doing big refactorings of the compiler one thing i m not sure about is if we can do memory reporting at the end of a phase and get the peak memory usage during that phase that seems to me to be the most useful statistic but it doesn t look like we can get that only the amount of memory currently allocated which is probably minimal between phases cc jdm nnethercote | 1 |
69,396 | 9,301,958,997 | IssuesEvent | 2019-03-24 04:03:15 | MoonHighway/pet-library | https://api.github.com/repos/MoonHighway/pet-library | closed | Create a Readme.md | documentation | Create Readme.md file with overview of project and installation instructions.
- [ ] Description
- [ ] Large Pet emoji or image
- [ ] requirements
- [ ] Installation instructions
- [ ] running development
| 1.0 | Create a Readme.md - Create Readme.md file with overview of project and installation instructions.
- [ ] Description
- [ ] Large Pet emoji or image
- [ ] requirements
- [ ] Installation instructions
- [ ] running development
| non_infrastructure | create a readme md create readme md file with overview of project and installation instructions description large pet emoji or image requirements installation instructions running development | 0 |
9,661 | 8,070,956,505 | IssuesEvent | 2018-08-06 11:33:40 | mindfulness-at-the-computer/mindfulness-at-the-computer | https://api.github.com/repos/mindfulness-at-the-computer/mindfulness-at-the-computer | closed | Setting up automated code coverage | infrastructure | We have successfully used CodeCov before but would like to find something that is less intrusive in the github PR system. Maybe Codacy's code coverage can be used:
https://github.com/codacy/python-codacy-coverage#setup
It may also be useful to start using coverage config files: https://coverage.readthedocs.io/en/coverage-4.5.1/config.html | 1.0 | Setting up automated code coverage - We have successfully used CodeCov before but would like to find something that is less intrusive in the github PR system. Maybe Codacy's code coverage can be used:
https://github.com/codacy/python-codacy-coverage#setup
It may also be useful to start using coverage config files: https://coverage.readthedocs.io/en/coverage-4.5.1/config.html | infrastructure | setting up automated code coverage we have successfully used codecov before but would like to find something that is less intrusive in the github pr system maybe codacy s code coverage can be used it may also be useful to start using coverage config files | 1 |
9,935 | 8,256,937,119 | IssuesEvent | 2018-09-13 01:57:08 | astroML/astroML | https://api.github.com/repos/astroML/astroML | opened | Include examples in the test suite | infrastructure | Examples are part of the user docs, they should be CI tested | 1.0 | Include examples in the test suite - Examples are part of the user docs, they should be CI tested | infrastructure | include examples in the test suite examples are part of the user docs they should be ci tested | 1 |
22,023 | 14,966,207,300 | IssuesEvent | 2021-01-27 14:19:38 | gonczor/black-sheep-learns | https://api.github.com/repos/gonczor/black-sheep-learns | closed | Add Celery task to resize course image | Backend Infrastructure enhancement | This will require setting up proper AWS infrastructure, integrating Celery and writing the task itself. | 1.0 | Add Celery task to resize course image - This will require setting up proper AWS infrastructure, integrating Celery and writing the task itself. | infrastructure | add celery task to resize course image this will require setting up proper aws infrastructure integrating celery and writing the task itself | 1 |
19,109 | 13,187,634,951 | IssuesEvent | 2020-08-13 04:03:36 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | Cobalt nodes 05, 06, 07, and 08 are missing some software packages (Trac #1074) | Migrated from Trac infrastructure task | cobalt05, cobalt06, cobalt07, and cobalt08 are missing some software that is available on other nodes. I have noticed the absence of the "'''pdflatex'''" package, but I am not sure if there are others.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1074">https://code.icecube.wisc.edu/ticket/1074</a>, reported by ddouglas and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-07-29T19:01:57",
"description": "cobalt05, cobalt06, cobalt07, and cobalt08 are missing some software that is available on other nodes. I have noticed the absence of the \"'''pdflatex'''\" package, but I am not sure if there are others.",
"reporter": "ddouglas",
"cc": "",
"resolution": "wontfix",
"_ts": "1438196517852643",
"component": "infrastructure",
"summary": "Cobalt nodes 05, 06, 07, and 08 are missing some software packages",
"priority": "minor",
"keywords": "pdflatex",
"time": "2015-07-29T19:01:00",
"milestone": "",
"owner": "",
"type": "task"
}
```
</p>
</details>
| 1.0 | Cobalt nodes 05, 06, 07, and 08 are missing some software packages (Trac #1074) - cobalt05, cobalt06, cobalt07, and cobalt08 are missing some software that is available on other nodes. I have noticed the absence of the "'''pdflatex'''" package, but I am not sure if there are others.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1074">https://code.icecube.wisc.edu/ticket/1074</a>, reported by ddouglas and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-07-29T19:01:57",
"description": "cobalt05, cobalt06, cobalt07, and cobalt08 are missing some software that is available on other nodes. I have noticed the absence of the \"'''pdflatex'''\" package, but I am not sure if there are others.",
"reporter": "ddouglas",
"cc": "",
"resolution": "wontfix",
"_ts": "1438196517852643",
"component": "infrastructure",
"summary": "Cobalt nodes 05, 06, 07, and 08 are missing some software packages",
"priority": "minor",
"keywords": "pdflatex",
"time": "2015-07-29T19:01:00",
"milestone": "",
"owner": "",
"type": "task"
}
```
</p>
</details>
| infrastructure | cobalt nodes and are missing some software packages trac and are missing some software that is available on other nodes i have noticed the absence of the pdflatex package but i am not sure if there are others migrated from json status closed changetime description and are missing some software that is available on other nodes i have noticed the absence of the pdflatex package but i am not sure if there are others reporter ddouglas cc resolution wontfix ts component infrastructure summary cobalt nodes and are missing some software packages priority minor keywords pdflatex time milestone owner type task | 1 |
34,093 | 28,215,573,707 | IssuesEvent | 2023-04-05 08:42:46 | pytest-dev/pytest | https://api.github.com/repos/pytest-dev/pytest | opened | Use separate GitHub environment for deploy workflows | type: infrastructure | Reading this page https://docs.pypi.org/trusted-publishers/security-model/#considerations they mention:
> [Use a dedicated environment](https://docs.github.com/en/actions/deployment/targeting-different-environments/using-environments-for-deployment): GitHub Actions supports "environments," which can be used to isolate secrets to specific workflows. OIDC publishing doesn't use any pre-configured secrets, but a dedicated publish or deploy environment is a general best practice.
>
> Dedicated environments allow for additional protections like [required reviewers](https://docs.github.com/en/actions/deployment/targeting-different-environments/using-environments-for-deployment#required-reviewers), which can be used to require manual approval for a workflow using the environment.
Seems like a good idea to segregate the deployment secrets to their own environment and add required reviewers (probably pytest core?) to better protect our releases.
I can try to set it up before the next release (#10869) if we agree. | 1.0 | Use separate GitHub environment for deploy workflows - Reading this page https://docs.pypi.org/trusted-publishers/security-model/#considerations they mention:
> [Use a dedicated environment](https://docs.github.com/en/actions/deployment/targeting-different-environments/using-environments-for-deployment): GitHub Actions supports "environments," which can be used to isolate secrets to specific workflows. OIDC publishing doesn't use any pre-configured secrets, but a dedicated publish or deploy environment is a general best practice.
>
> Dedicated environments allow for additional protections like [required reviewers](https://docs.github.com/en/actions/deployment/targeting-different-environments/using-environments-for-deployment#required-reviewers), which can be used to require manual approval for a workflow using the environment.
Seems like a good idea to segregate the deployment secrets to their own environment and add required reviewers (probably pytest core?) to better protect our releases.
I can try to set it up before the next release (#10869) if we agree. | infrastructure | use separate github environment for deploy workflows reading this page they mention github actions supports environments which can be used to isolate secrets to specific workflows oidc publishing doesn t use any pre configured secrets but a dedicated publish or deploy environment is a general best practice dedicated environments allow for additional protections like which can be used to require manual approval for a workflow using the environment seems like a good idea to segregate the deployment secrets to their own environment and add required reviewers probably pytest core to better protect our releases i can try to set it up before the next release if we agree | 1 |
9,346 | 7,929,137,781 | IssuesEvent | 2018-07-06 14:08:24 | aiidateam/aiida_core | https://api.github.com/repos/aiidateam/aiida_core | opened | pgtest fails to start a postgres server on travis | priority/critical-blocking topic/TestInfrastructure | This seems to have been induced by a change in the travis worker setup around 2018-07-05. | 1.0 | pgtest fails to start a postgres server on travis - This seems to have been induced by a change in the travis worker setup around 2018-07-05. | infrastructure | pgtest fails to start a postgres server on travis this seems to have been induced by a change in the travis worker setup around | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.