Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
19,081
25,127,606,887
IssuesEvent
2022-11-09 12:58:17
inmanta/inmanta-core
https://api.github.com/repos/inmanta/inmanta-core
reopened
Enable timestamp on the log messages produced by the test cases
process tiny
In order to debug a test case that times out on a wait condition, it's handy to see the timestamps in the log messages. That way it easier to see in which stage a lot of time is being lost. Adding the timestamps can be achieved by providing the following options to pytest: `--log-format="%(asctime)s.%(msecs)03d %(levelname)s %(message)s"`. If `pytest.ini` supports it, add it there.
1.0
Enable timestamp on the log messages produced by the test cases - In order to debug a test case that times out on a wait condition, it's handy to see the timestamps in the log messages. That way it easier to see in which stage a lot of time is being lost. Adding the timestamps can be achieved by providing the following options to pytest: `--log-format="%(asctime)s.%(msecs)03d %(levelname)s %(message)s"`. If `pytest.ini` supports it, add it there.
process
enable timestamp on the log messages produced by the test cases in order to debug a test case that times out on a wait condition it s handy to see the timestamps in the log messages that way it easier to see in which stage a lot of time is being lost adding the timestamps can be achieved by providing the following options to pytest log format asctime s msecs levelname s message s if pytest ini supports it add it there
1
139,101
12,837,281,236
IssuesEvent
2020-07-07 15:33:10
rust-bio/rust-bio
https://api.github.com/repos/rust-bio/rust-bio
closed
Improve docs: `seq_analysis::gc`
documentation
* [ ] add at least one example * [ ] describe time and/or memory complexity
1.0
Improve docs: `seq_analysis::gc` - * [ ] add at least one example * [ ] describe time and/or memory complexity
non_process
improve docs seq analysis gc add at least one example describe time and or memory complexity
0
20,431
27,098,151,954
IssuesEvent
2023-02-15 05:56:49
anobaka/InsideWorld
https://api.github.com/repos/anobaka/InsideWorld
closed
优化文件处理器部分操作交互
enhancement file-processor
+ [x] 当上方文件数量发生变化时,可能会导致误点击【提取】等按钮,提取按钮现调整至文件名右侧,并增加确认弹窗(支持回车提交) ![image](https://user-images.githubusercontent.com/2888789/207590578-fa8f417f-c014-43bc-87b6-22ee409958b0.png) ![image](https://user-images.githubusercontent.com/2888789/207590732-c138e556-44bf-410e-9157-5f6b6de6e7cf.png) + [x] 文件操作按钮可能需要尽可能贴近文件名 + [x] 根目录变更时显示加载中图标
1.0
优化文件处理器部分操作交互 - + [x] 当上方文件数量发生变化时,可能会导致误点击【提取】等按钮,提取按钮现调整至文件名右侧,并增加确认弹窗(支持回车提交) ![image](https://user-images.githubusercontent.com/2888789/207590578-fa8f417f-c014-43bc-87b6-22ee409958b0.png) ![image](https://user-images.githubusercontent.com/2888789/207590732-c138e556-44bf-410e-9157-5f6b6de6e7cf.png) + [x] 文件操作按钮可能需要尽可能贴近文件名 + [x] 根目录变更时显示加载中图标
process
优化文件处理器部分操作交互 当上方文件数量发生变化时,可能会导致误点击【提取】等按钮,提取按钮现调整至文件名右侧,并增加确认弹窗(支持回车提交) 文件操作按钮可能需要尽可能贴近文件名 根目录变更时显示加载中图标
1
262,590
19,822,793,084
IssuesEvent
2022-01-20 00:44:08
vercel/next.js
https://api.github.com/repos/vercel/next.js
opened
Docs: Tailwind scaffolding in Next.js Projects for separation of concerns
template: documentation
### What is the improvement or update you wish to see? I would like to see a template on how we should organize scaffolding of CSS files when using tailwind. ### Is there any context that might help us understand? I know we can use global.css file for global variables, but what is the proper way of organizing CSS files for the sake of separation of concerns, route folders, and/or general organization? Are there any gotchas/caveats to separating CSS files? Checked the with-tailwind example, and unfortunately it only provided the simplest example with global.css file(initial setup) and inline modifications. And thanks! It really has been great getting to know Next.js btw. Still got loads to learn but having fun. ### Does the docs page already exist? Please link to it. https://github.com/vercel/next.js/tree/canary/examples/with-tailwindcss
1.0
Docs: Tailwind scaffolding in Next.js Projects for separation of concerns - ### What is the improvement or update you wish to see? I would like to see a template on how we should organize scaffolding of CSS files when using tailwind. ### Is there any context that might help us understand? I know we can use global.css file for global variables, but what is the proper way of organizing CSS files for the sake of separation of concerns, route folders, and/or general organization? Are there any gotchas/caveats to separating CSS files? Checked the with-tailwind example, and unfortunately it only provided the simplest example with global.css file(initial setup) and inline modifications. And thanks! It really has been great getting to know Next.js btw. Still got loads to learn but having fun. ### Does the docs page already exist? Please link to it. https://github.com/vercel/next.js/tree/canary/examples/with-tailwindcss
non_process
docs tailwind scaffolding in next js projects for separation of concerns what is the improvement or update you wish to see i would like to see a template on how we should organize scaffolding of css files when using tailwind is there any context that might help us understand i know we can use global css file for global variables but what is the proper way of organizing css files for the sake of separation of concerns route folders and or general organization are there any gotchas caveats to separating css files checked the with tailwind example and unfortunately it only provided the simplest example with global css file initial setup and inline modifications and thanks it really has been great getting to know next js btw still got loads to learn but having fun does the docs page already exist please link to it
0
284,561
21,444,711,111
IssuesEvent
2022-04-25 04:13:04
cengage/react-magma
https://api.github.com/repos/cengage/react-magma
closed
Documentation> Checkbox link> "Not Found" error message is appearing upon clicking the link
documentation
**Describe the bug** "Not Found" error message is appearing upon clicking the link "Checkbox". **To Reproduce** Steps to reproduce the behavior: 1. Go to https://react-magma.cengage.com/version/2.5.7/api/radio/ 2. Click on "Checkbox" link 3. Verify the error **Expected behavior** Appropriate page should appear upon clicking the link "Checkbox". **Screenshots** https://somup.com/c3nXXZZRQO **Desktop:** - OS: [Win 10] - Browser [chrome] - Version [Chrome version: Version 98.0.4758.81 (Official Build) (64-bit)]
1.0
Documentation> Checkbox link> "Not Found" error message is appearing upon clicking the link - **Describe the bug** "Not Found" error message is appearing upon clicking the link "Checkbox". **To Reproduce** Steps to reproduce the behavior: 1. Go to https://react-magma.cengage.com/version/2.5.7/api/radio/ 2. Click on "Checkbox" link 3. Verify the error **Expected behavior** Appropriate page should appear upon clicking the link "Checkbox". **Screenshots** https://somup.com/c3nXXZZRQO **Desktop:** - OS: [Win 10] - Browser [chrome] - Version [Chrome version: Version 98.0.4758.81 (Official Build) (64-bit)]
non_process
documentation checkbox link not found error message is appearing upon clicking the link describe the bug not found error message is appearing upon clicking the link checkbox to reproduce steps to reproduce the behavior go to click on checkbox link verify the error expected behavior appropriate page should appear upon clicking the link checkbox screenshots desktop os browser version
0
10,346
13,172,404,656
IssuesEvent
2020-08-11 18:22:41
openenclave/openenclave
https://api.github.com/repos/openenclave/openenclave
closed
Unable to distinguish API versions
attestation bug engineering process triaged
The master branch saw a change to the host-side quote verification library: `oe_verify_remote_report` now expects 5 arguments instead of 3 arguments in v0.7.0. We now check for `#if defined(OE_CLAIM_ID_VERSION)`, which isn't defined in v0.7.0, to tell the difference. It would be nice if the version number were updated when the API changes but as far as I can tell, `OE_API_VERSION` is always the same (2). Perhaps it makes sense to add separate versioning for the host-side quote verification library too?
1.0
Unable to distinguish API versions - The master branch saw a change to the host-side quote verification library: `oe_verify_remote_report` now expects 5 arguments instead of 3 arguments in v0.7.0. We now check for `#if defined(OE_CLAIM_ID_VERSION)`, which isn't defined in v0.7.0, to tell the difference. It would be nice if the version number were updated when the API changes but as far as I can tell, `OE_API_VERSION` is always the same (2). Perhaps it makes sense to add separate versioning for the host-side quote verification library too?
process
unable to distinguish api versions the master branch saw a change to the host side quote verification library oe verify remote report now expects arguments instead of arguments in we now check for if defined oe claim id version which isn t defined in to tell the difference it would be nice if the version number were updated when the api changes but as far as i can tell oe api version is always the same perhaps it makes sense to add separate versioning for the host side quote verification library too
1
10,265
13,112,556,253
IssuesEvent
2020-08-05 02:34:22
kubeflow/manifests
https://api.github.com/repos/kubeflow/manifests
closed
Owners file for kubeflow/metadata needs to be updated
area/metadata kind/process lifecycle/stale priority/p2
The OWNERs file for metadata needs to be updated it looks like the current set of approvers is no longer up to date and responsive. See #928
1.0
Owners file for kubeflow/metadata needs to be updated - The OWNERs file for metadata needs to be updated it looks like the current set of approvers is no longer up to date and responsive. See #928
process
owners file for kubeflow metadata needs to be updated the owners file for metadata needs to be updated it looks like the current set of approvers is no longer up to date and responsive see
1
10,826
13,609,564,470
IssuesEvent
2020-09-23 05:36:14
GoogleCloudPlatform/getting-started-java
https://api.github.com/repos/GoogleCloudPlatform/getting-started-java
closed
Dependency Dashboard
type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-compiler-plugin-3.x -->chore(deps): update dependency org.apache.maven.plugins:maven-compiler-plugin to v3.8.1 - [ ] <!-- rebase-branch=renovate/javax.servlet-javax.servlet-api-4.x -->chore(deps): update dependency javax.servlet:javax.servlet-api to v4 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-compiler-plugin-3.x -->chore(deps): update dependency org.apache.maven.plugins:maven-compiler-plugin to v3.8.1 - [ ] <!-- rebase-branch=renovate/javax.servlet-javax.servlet-api-4.x -->chore(deps): update dependency javax.servlet:javax.servlet-api to v4 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency org apache maven plugins maven compiler plugin to chore deps update dependency javax servlet javax servlet api to check this box to trigger a request for renovate to run again on this repository
1
355,314
25,175,898,206
IssuesEvent
2022-11-11 09:13:58
Devanshshah1309/pe
https://api.github.com/repos/Devanshshah1309/pe
opened
Link to FAQ Section not working
severity.Low type.DocumentationBug
The link to the FAQ section here is broken (or there's no link at all). ![Screenshot 2022-11-11 at 5.12.34 PM.png](https://raw.githubusercontent.com/Devanshshah1309/pe/main/files/4cc1cad7-63a1-4ec1-971a-c141515a3ae3.png) This prevents me from being able to go to the FAQ section if I'm only interested in that - it is not purely a cosmetic bug. It hinders my reading. Hence, low severity. <!--session: 1668147090315-757e20e0-6c8f-46c9-a0a7-2575a40077ea--> <!--Version: Web v3.4.4-->
1.0
Link to FAQ Section not working - The link to the FAQ section here is broken (or there's no link at all). ![Screenshot 2022-11-11 at 5.12.34 PM.png](https://raw.githubusercontent.com/Devanshshah1309/pe/main/files/4cc1cad7-63a1-4ec1-971a-c141515a3ae3.png) This prevents me from being able to go to the FAQ section if I'm only interested in that - it is not purely a cosmetic bug. It hinders my reading. Hence, low severity. <!--session: 1668147090315-757e20e0-6c8f-46c9-a0a7-2575a40077ea--> <!--Version: Web v3.4.4-->
non_process
link to faq section not working the link to the faq section here is broken or there s no link at all this prevents me from being able to go to the faq section if i m only interested in that it is not purely a cosmetic bug it hinders my reading hence low severity
0
11,531
14,403,749,597
IssuesEvent
2020-12-03 16:25:55
LOVDnl/LOVD3
https://api.github.com/repos/LOVDnl/LOVD3
opened
Curators and up can create variant submissions without classification.
bug minor cat: submission process
Curators and up can create variant submissions without classification, but only when a gene is selected. VOG entries still require at least the reported classification, while VOT entries do not. Submitters are always required to fill in the reported classification.
1.0
Curators and up can create variant submissions without classification. - Curators and up can create variant submissions without classification, but only when a gene is selected. VOG entries still require at least the reported classification, while VOT entries do not. Submitters are always required to fill in the reported classification.
process
curators and up can create variant submissions without classification curators and up can create variant submissions without classification but only when a gene is selected vog entries still require at least the reported classification while vot entries do not submitters are always required to fill in the reported classification
1
9,175
12,226,438,488
IssuesEvent
2020-05-03 10:55:40
labnote-ant/labnote
https://api.github.com/repos/labnote-ant/labnote
closed
Add description box
chemical-view process-view
In chemical view or process view, it needs a description box where the user can input details of chemicals or process.
1.0
Add description box - In chemical view or process view, it needs a description box where the user can input details of chemicals or process.
process
add description box in chemical view or process view it needs a description box where the user can input details of chemicals or process
1
14,960
18,445,033,755
IssuesEvent
2021-10-15 00:01:02
cloudfoundry/cf-k8s-api
https://api.github.com/repos/cloudfoundry/cf-k8s-api
closed
[Feature]: API Client can List Processes for an App via `GET /v3/apps/:guid/processes`
Processes
### Blockers/Dependencies _No response_ ### Background **As a** client of the API Shim **I want** to be able to list all Processes for my App **So that** I can discover information about my Processes The CF CLI hits this endpoint during `cf push` so that it can get a list of Process guids to make future API calls (such as fetching stats for a Process). ### Acceptance Criteria ## Scenarios ### Happy Path (App with Processes) **GIVEN** I have a CFApp and CFProcesses are associated with it **WHEN** I make the following API request: ```bash curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \ -X GET \ -H "Authorization: bearer <placeholder-bearer-token>" ``` **THEN** I see a response that reflects the information on the CFProcesses ```json HTTP/1.1 200 OK Content-Type: application/json { "pagination": { "total_results": 2, "total_pages": 1, "first": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=1&per_page=2" }, "last": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2" }, "next": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2" }, "previous": null }, "resources": [ { "guid": "6a901b7c-9417-4dc1-8189-d3234aa0ab82", "type": "web", "command": "[PRIVATE DATA HIDDEN IN LISTS]", "instances": 5, "memory_in_mb": 256, "disk_in_mb": 1024, "health_check": { "type": "port", "data": { "timeout": null, "invocation_timeout": null } }, "relationships": { "app": { "data": { "guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" } } }, "metadata": { "labels": {}, "annotations": {} }, "created_at": "2016-03-23T18:48:22Z", "updated_at": "2016-03-23T18:48:42Z", "links": { "self": { "href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82" }, "scale": { "href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/actions/scale", "method": "POST" }, "app": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" }, "space": { "href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576" }, "stats": { "href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/stats" } } }, { "guid": "3fccacd9-4b02-4b96-8d02-8e865865e9eb", "type": "worker", "command": "[PRIVATE DATA HIDDEN IN LISTS]", "instances": 1, "memory_in_mb": 256, "disk_in_mb": 1024, "health_check": { "type": "process", "data": { "timeout": null, "invocation_timeout": null } }, "relationships": { "app": { "data": { "guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" } }, }, "metadata": { "labels": {}, "annotations": {} }, "created_at": "2016-03-23T18:48:22Z", "updated_at": "2016-03-23T18:48:42Z", "links": { "self": { "href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb" }, "scale": { "href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/actions/scale", "method": "POST" }, "app": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" }, "space": { "href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576" }, "stats": { "href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/stats" } } } ] } ``` *Note*: we're omitting the `revision` key entirely. The `metadata` key will always contain empty hashes, as in other stories. --- ### App with No Processes **GIVEN** I have a CFApp and **no** CFProcesses are associated with it **WHEN** I make the following API request: ```bash curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \ -X GET \ -H "Authorization: bearer <placeholder-bearer-token>" ``` **THEN** I get back a response with an empty resources array ```json HTTP/1.1 200 OK Content-Type: application/json { "pagination": { "total_results": 0, "total_pages": 1, "first": { "href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50" }, "last": { "href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50" }, "next": null, "previous": null }, "resources": [ ] } ``` --- ### App doesn't exist **GIVEN** I have do not have a CFApp with the guid below **WHEN** I make the following API request: ```bash curl "https://api-shim.example.org/v3/apps/<non-existant-app-guid>/processes" \ -X GET \ -H "Authorization: bearer <placeholder-bearer-token>" ``` **THEN** I get back a 404 response ```json HTTP/1.1 404 Not Found { "errors": [ { "detail": "App not found", "title": "CF-ResourceNotFound", "code": 10010 } ] } ``` ### Dev Notes * V3 API Docs: https://v3-apidocs.cloudfoundry.org/version/3.107.0/index.html#list-processes * Pagination: Always return all results for now (as we have been doing on other stories) * Query parameters: Ignore filter parameters for now. We can add them in later * Be sure to add the necessary RBAC annotations in the new Process repository. Otherwise the app will error when deployed to a real cluster (but will work locally)
1.0
[Feature]: API Client can List Processes for an App via `GET /v3/apps/:guid/processes` - ### Blockers/Dependencies _No response_ ### Background **As a** client of the API Shim **I want** to be able to list all Processes for my App **So that** I can discover information about my Processes The CF CLI hits this endpoint during `cf push` so that it can get a list of Process guids to make future API calls (such as fetching stats for a Process). ### Acceptance Criteria ## Scenarios ### Happy Path (App with Processes) **GIVEN** I have a CFApp and CFProcesses are associated with it **WHEN** I make the following API request: ```bash curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \ -X GET \ -H "Authorization: bearer <placeholder-bearer-token>" ``` **THEN** I see a response that reflects the information on the CFProcesses ```json HTTP/1.1 200 OK Content-Type: application/json { "pagination": { "total_results": 2, "total_pages": 1, "first": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=1&per_page=2" }, "last": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2" }, "next": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2" }, "previous": null }, "resources": [ { "guid": "6a901b7c-9417-4dc1-8189-d3234aa0ab82", "type": "web", "command": "[PRIVATE DATA HIDDEN IN LISTS]", "instances": 5, "memory_in_mb": 256, "disk_in_mb": 1024, "health_check": { "type": "port", "data": { "timeout": null, "invocation_timeout": null } }, "relationships": { "app": { "data": { "guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" } } }, "metadata": { "labels": {}, "annotations": {} }, "created_at": "2016-03-23T18:48:22Z", "updated_at": "2016-03-23T18:48:42Z", "links": { "self": { "href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82" }, "scale": { "href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/actions/scale", "method": "POST" }, "app": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" }, "space": { "href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576" }, "stats": { "href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/stats" } } }, { "guid": "3fccacd9-4b02-4b96-8d02-8e865865e9eb", "type": "worker", "command": "[PRIVATE DATA HIDDEN IN LISTS]", "instances": 1, "memory_in_mb": 256, "disk_in_mb": 1024, "health_check": { "type": "process", "data": { "timeout": null, "invocation_timeout": null } }, "relationships": { "app": { "data": { "guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" } }, }, "metadata": { "labels": {}, "annotations": {} }, "created_at": "2016-03-23T18:48:22Z", "updated_at": "2016-03-23T18:48:42Z", "links": { "self": { "href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb" }, "scale": { "href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/actions/scale", "method": "POST" }, "app": { "href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5" }, "space": { "href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576" }, "stats": { "href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/stats" } } } ] } ``` *Note*: we're omitting the `revision` key entirely. The `metadata` key will always contain empty hashes, as in other stories. --- ### App with No Processes **GIVEN** I have a CFApp and **no** CFProcesses are associated with it **WHEN** I make the following API request: ```bash curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \ -X GET \ -H "Authorization: bearer <placeholder-bearer-token>" ``` **THEN** I get back a response with an empty resources array ```json HTTP/1.1 200 OK Content-Type: application/json { "pagination": { "total_results": 0, "total_pages": 1, "first": { "href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50" }, "last": { "href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50" }, "next": null, "previous": null }, "resources": [ ] } ``` --- ### App doesn't exist **GIVEN** I have do not have a CFApp with the guid below **WHEN** I make the following API request: ```bash curl "https://api-shim.example.org/v3/apps/<non-existant-app-guid>/processes" \ -X GET \ -H "Authorization: bearer <placeholder-bearer-token>" ``` **THEN** I get back a 404 response ```json HTTP/1.1 404 Not Found { "errors": [ { "detail": "App not found", "title": "CF-ResourceNotFound", "code": 10010 } ] } ``` ### Dev Notes * V3 API Docs: https://v3-apidocs.cloudfoundry.org/version/3.107.0/index.html#list-processes * Pagination: Always return all results for now (as we have been doing on other stories) * Query parameters: Ignore filter parameters for now. We can add them in later * Be sure to add the necessary RBAC annotations in the new Process repository. Otherwise the app will error when deployed to a real cluster (but will work locally)
process
api client can list processes for an app via get apps guid processes blockers dependencies no response background as a client of the api shim i want to be able to list all processes for my app so that i can discover information about my processes the cf cli hits this endpoint during cf push so that it can get a list of process guids to make future api calls such as fetching stats for a process acceptance criteria scenarios happy path app with processes given i have a cfapp and cfprocesses are associated with it when i make the following api request bash curl x get h authorization bearer then i see a response that reflects the information on the cfprocesses json http ok content type application json pagination total results total pages first href last href next href previous null resources guid type web command instances memory in mb disk in mb health check type port data timeout null invocation timeout null relationships app data guid metadata labels annotations created at updated at links self href scale href method post app href space href stats href guid type worker command instances memory in mb disk in mb health check type process data timeout null invocation timeout null relationships app data guid metadata labels annotations created at updated at links self href scale href method post app href space href stats href note we re omitting the revision key entirely the metadata key will always contain empty hashes as in other stories app with no processes given i have a cfapp and no cfprocesses are associated with it when i make the following api request bash curl x get h authorization bearer then i get back a response with an empty resources array json http ok content type application json pagination total results total pages first href last href next null previous null resources app doesn t exist given i have do not have a cfapp with the guid below when i make the following api request bash curl x get h authorization bearer then i get back a response json http not found errors detail app not found title cf resourcenotfound code dev notes api docs pagination always return all results for now as we have been doing on other stories query parameters ignore filter parameters for now we can add them in later be sure to add the necessary rbac annotations in the new process repository otherwise the app will error when deployed to a real cluster but will work locally
1
271,136
29,299,168,536
IssuesEvent
2023-05-25 01:06:58
hshivhare67/kernel_v4.19.72_CVE-2023-0461
https://api.github.com/repos/hshivhare67/kernel_v4.19.72_CVE-2023-0461
opened
CVE-2023-33203 (Medium) detected in linuxlinux-4.19.282
Mend: dependency security vulnerability
## CVE-2023-33203 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72_CVE-2023-0461/commit/20984407a51d9f25ee9889e4b1304489f480d36e">20984407a51d9f25ee9889e4b1304489f480d36e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The Linux kernel before 6.2.9 has a race condition and resultant use-after-free in drivers/net/ethernet/qualcomm/emac/emac.c if a physically proximate attacker unplugs an emac based device. <p>Publish Date: 2023-05-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-33203>CVE-2023-33203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-33203">https://www.linuxkernelcves.com/cves/CVE-2023-33203</a></p> <p>Release Date: 2023-05-18</p> <p>Fix Resolution: v4.14.312,v4.19.280,v5.4.240,v5.10.177,v5.15.105,v6.1.22,v6.2.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-33203 (Medium) detected in linuxlinux-4.19.282 - ## CVE-2023-33203 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72_CVE-2023-0461/commit/20984407a51d9f25ee9889e4b1304489f480d36e">20984407a51d9f25ee9889e4b1304489f480d36e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The Linux kernel before 6.2.9 has a race condition and resultant use-after-free in drivers/net/ethernet/qualcomm/emac/emac.c if a physically proximate attacker unplugs an emac based device. <p>Publish Date: 2023-05-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-33203>CVE-2023-33203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-33203">https://www.linuxkernelcves.com/cves/CVE-2023-33203</a></p> <p>Release Date: 2023-05-18</p> <p>Fix Resolution: v4.14.312,v4.19.280,v5.4.240,v5.10.177,v5.15.105,v6.1.22,v6.2.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net ethernet qualcomm emac emac c drivers net ethernet qualcomm emac emac c vulnerability details the linux kernel before has a race condition and resultant use after free in drivers net ethernet qualcomm emac emac c if a physically proximate attacker unplugs an emac based device publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,858
5,824,282,508
IssuesEvent
2017-05-07 11:31:07
QCoDeS/Qcodes
https://api.github.com/repos/QCoDeS/Qcodes
closed
Leaking sockets?
bug mulitprocessing p2
### Steps to reproduce 1. Run code for some time on windows including possible shutdown and restart of notebook and instrument communication. ### Expected behaviour Things should keep working ### Actual behaviour Notebook fails with socket related connection issues from Tornado and ZMQ. ``` File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\multikernelmanager.py", line 33, in wrapped r = method(*args, **kwargs) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\ioloop\manager.py", line 33, in wrapped socket = f(self, *args, **kwargs) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\connect.py", line 492, in connect_shell return self._create_connected_socket('shell', identity=identity) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\connect.py", line 476, in _create_connected_socket sock = self.context.socket(socket_type) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\z mq\sugar\context.py", line 146, in socket s = self._socket_class(self, socket_type, **kwargs) File "zmq\backend\cython\socket.pyx", line 285, in zmq.backend.cython.sock et.Socket.__cinit__ (zmq\backend\cython\socket.c:3861) zmq.error.ZMQError: No buffer space available OSError: [WinError 10055] An operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full ``` Regular internet (browsing etc) is flaky too We suspect that sockets are leaked and not cleaned up. This may be in the VISA driver of network instruments such as ZNMB20 VNA ### System **operating system** Windows **qcodes branch** master **qcodes commit** ?
1.0
Leaking sockets? - ### Steps to reproduce 1. Run code for some time on windows including possible shutdown and restart of notebook and instrument communication. ### Expected behaviour Things should keep working ### Actual behaviour Notebook fails with socket related connection issues from Tornado and ZMQ. ``` File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\multikernelmanager.py", line 33, in wrapped r = method(*args, **kwargs) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\ioloop\manager.py", line 33, in wrapped socket = f(self, *args, **kwargs) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\connect.py", line 492, in connect_shell return self._create_connected_socket('shell', identity=identity) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j upyter_client\connect.py", line 476, in _create_connected_socket sock = self.context.socket(socket_type) File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\z mq\sugar\context.py", line 146, in socket s = self._socket_class(self, socket_type, **kwargs) File "zmq\backend\cython\socket.pyx", line 285, in zmq.backend.cython.sock et.Socket.__cinit__ (zmq\backend\cython\socket.c:3861) zmq.error.ZMQError: No buffer space available OSError: [WinError 10055] An operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full ``` Regular internet (browsing etc) is flaky too We suspect that sockets are leaked and not cleaned up. This may be in the VISA driver of network instruments such as ZNMB20 VNA ### System **operating system** Windows **qcodes branch** master **qcodes commit** ?
process
leaking sockets steps to reproduce run code for some time on windows including possible shutdown and restart of notebook and instrument communication expected behaviour things should keep working actual behaviour notebook fails with socket related connection issues from tornado and zmq file c users envs qcodes master lib site packages j upyter client multikernelmanager py line in wrapped r method args kwargs file c users envs qcodes master lib site packages j upyter client ioloop manager py line in wrapped socket f self args kwargs file c users envs qcodes master lib site packages j upyter client connect py line in connect shell return self create connected socket shell identity identity file c users envs qcodes master lib site packages j upyter client connect py line in create connected socket sock self context socket socket type file c users envs qcodes master lib site packages z mq sugar context py line in socket s self socket class self socket type kwargs file zmq backend cython socket pyx line in zmq backend cython sock et socket cinit zmq backend cython socket c zmq error zmqerror no buffer space available oserror an operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full regular internet browsing etc is flaky too we suspect that sockets are leaked and not cleaned up this may be in the visa driver of network instruments such as vna system operating system windows qcodes branch master qcodes commit
1
117,932
25,216,777,579
IssuesEvent
2022-11-14 09:45:05
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Feature][Custom JS Lib Epic] Delete installed JS package on demand
JS Evaluation Task FE Coders Pod
## Summary Remove installed package from application when user clicks on the delete JS package button.
1.0
[Feature][Custom JS Lib Epic] Delete installed JS package on demand - ## Summary Remove installed package from application when user clicks on the delete JS package button.
non_process
delete installed js package on demand summary remove installed package from application when user clicks on the delete js package button
0
160,013
25,095,864,333
IssuesEvent
2022-11-08 10:07:13
metacpan/metacpan-web
https://api.github.com/repos/metacpan/metacpan-web
closed
multiple levels of indentation not handled
type:Bug design-2022-followup
Multiple levels of `=over 4` are not visible in the site redesign -- everything appears at the same level, so it's impossible to tell what content is meant to be nested. example: https://metacpan.org/pod/Net::IDN::Encode -- there is a list of functions, and inside the function description is a list of options, but the options are shown at the same level as the functions themselves.
1.0
multiple levels of indentation not handled - Multiple levels of `=over 4` are not visible in the site redesign -- everything appears at the same level, so it's impossible to tell what content is meant to be nested. example: https://metacpan.org/pod/Net::IDN::Encode -- there is a list of functions, and inside the function description is a list of options, but the options are shown at the same level as the functions themselves.
non_process
multiple levels of indentation not handled multiple levels of over are not visible in the site redesign everything appears at the same level so it s impossible to tell what content is meant to be nested example there is a list of functions and inside the function description is a list of options but the options are shown at the same level as the functions themselves
0
350,710
31,931,967,598
IssuesEvent
2023-09-19 08:03:21
Convergence-Project/step-backend
https://api.github.com/repos/Convergence-Project/step-backend
opened
[2주차] (문제집) 게시판 좋아요 기능 구현
🎯test ✨feature
✏️Description - 작업사항을 입력해주세요 ✅TODO - - [ ] 테스트 코드 작성 - [ ] 컨트롤러 프론트 연동 🐾ETC -
1.0
[2주차] (문제집) 게시판 좋아요 기능 구현 - ✏️Description - 작업사항을 입력해주세요 ✅TODO - - [ ] 테스트 코드 작성 - [ ] 컨트롤러 프론트 연동 🐾ETC -
non_process
문제집 게시판 좋아요 기능 구현 ✏️description 작업사항을 입력해주세요 ✅todo 테스트 코드 작성 컨트롤러 프론트 연동 🐾etc
0
18,485
24,550,797,339
IssuesEvent
2022-10-12 12:27:50
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] 'View consent' and 'View website' buttons are not getting displayed on the study overview screen
Bug P1 iOS Process: Fixed Process: Tested dev
Steps: 1. Sign up or sign in to the app 2. Click on any study to enroll 3. Click on the participate 4. I enrollment flow, click on the 'Cancel' button 5. Click on 'End task' or 'Click on 'Discard result' and observe AR: 'View consent' and 'View website' buttons are not getting displayed on the study overview screen ER: ''View consent' and 'View website' buttons should get displayed on the study overview screen
2.0
[iOS] 'View consent' and 'View website' buttons are not getting displayed on the study overview screen - Steps: 1. Sign up or sign in to the app 2. Click on any study to enroll 3. Click on the participate 4. I enrollment flow, click on the 'Cancel' button 5. Click on 'End task' or 'Click on 'Discard result' and observe AR: 'View consent' and 'View website' buttons are not getting displayed on the study overview screen ER: ''View consent' and 'View website' buttons should get displayed on the study overview screen
process
view consent and view website buttons are not getting displayed on the study overview screen steps sign up or sign in to the app click on any study to enroll click on the participate i enrollment flow click on the cancel button click on end task or click on discard result and observe ar view consent and view website buttons are not getting displayed on the study overview screen er view consent and view website buttons should get displayed on the study overview screen
1
14,604
17,703,628,988
IssuesEvent
2021-08-25 03:26:03
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - basisOfRecord
Term - change Class - Record-level non-normative Process - complete
## Term change * Submitter: John Wieczorek * Efficacy Justification (why is this change necessary?): completeness * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): Result of recent public review * Stability Justification (what concerns are there that this might affect existing implementations?): An addition, no effect on stability except to promote standardization on a ratified term. * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Current Term definition: https://dwc.tdwg.org/list/#dwc_basisOfRecord Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): basisOfRecord * Organized in Class (e.g., Occurrence, Event, Location, Taxon): None * Definition of the term (normative): The specific nature of the data record. * Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use the standard label of one of the Darwin Core classes. * Examples (not normative): `PreservedSpecimen`, `FossilSpecimen`, `LivingSpecimen`, `MaterialSample`, `Event`, `HumanObservation`, `MachineObservation`, `Taxon`, `Occurrence`, **`MaterialCitation`** * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/basisOfRecord-2017-10-06** * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/RecordBasis This is to accommodate the addition of the MaterialCitation class.
1.0
Change term - basisOfRecord - ## Term change * Submitter: John Wieczorek * Efficacy Justification (why is this change necessary?): completeness * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): Result of recent public review * Stability Justification (what concerns are there that this might affect existing implementations?): An addition, no effect on stability except to promote standardization on a ratified term. * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Current Term definition: https://dwc.tdwg.org/list/#dwc_basisOfRecord Proposed attributes of the new term: * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): basisOfRecord * Organized in Class (e.g., Occurrence, Event, Location, Taxon): None * Definition of the term (normative): The specific nature of the data record. * Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use the standard label of one of the Darwin Core classes. * Examples (not normative): `PreservedSpecimen`, `FossilSpecimen`, `LivingSpecimen`, `MaterialSample`, `Event`, `HumanObservation`, `MachineObservation`, `Taxon`, `Occurrence`, **`MaterialCitation`** * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/basisOfRecord-2017-10-06** * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/RecordBasis This is to accommodate the addition of the MaterialCitation class.
process
change term basisofrecord term change submitter john wieczorek efficacy justification why is this change necessary completeness demand justification if the change is semantic in nature name at least two organizations that independently need this term result of recent public review stability justification what concerns are there that this might affect existing implementations an addition no effect on stability except to promote standardization on a ratified term implications for dwciri namespace does this change affect a dwciri term version none current term definition proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes basisofrecord organized in class e g occurrence event location taxon none definition of the term normative the specific nature of the data record usage comments recommendations regarding content etc not normative recommended best practice is to use the standard label of one of the darwin core classes examples not normative preservedspecimen fossilspecimen livingspecimen materialsample event humanobservation machineobservation taxon occurrence materialcitation refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative abcd xpath of the equivalent term in abcd or efg not normative datasets dataset units unit recordbasis this is to accommodate the addition of the materialcitation class
1
438,189
12,623,665,407
IssuesEvent
2020-06-14 00:30:23
hack4impact-uiuc/kids-save-ocean
https://api.github.com/repos/hack4impact-uiuc/kids-save-ocean
closed
Remove fake links
high priority
Feed: - Saved projects - My projects - Updates - Followers - Following Navbar: - Notifications - Current project Also remove all hardcoded data and comment out features projects from homepage and feed
1.0
Remove fake links - Feed: - Saved projects - My projects - Updates - Followers - Following Navbar: - Notifications - Current project Also remove all hardcoded data and comment out features projects from homepage and feed
non_process
remove fake links feed saved projects my projects updates followers following navbar notifications current project also remove all hardcoded data and comment out features projects from homepage and feed
0
336,706
10,195,758,226
IssuesEvent
2019-08-12 18:57:07
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
opened
Add kaniko image version to version stream
area/tekton area/versions kind/enhancement priority/important-soon
We hardcode the default Kaniko image and version in the code currently. That's kinda silly when we've got this whole version stream thing here. =) So let's add a Kaniko version to the version stream, and then update the logic in the CRD translation to use that.
1.0
Add kaniko image version to version stream - We hardcode the default Kaniko image and version in the code currently. That's kinda silly when we've got this whole version stream thing here. =) So let's add a Kaniko version to the version stream, and then update the logic in the CRD translation to use that.
non_process
add kaniko image version to version stream we hardcode the default kaniko image and version in the code currently that s kinda silly when we ve got this whole version stream thing here so let s add a kaniko version to the version stream and then update the logic in the crd translation to use that
0
7,647
10,738,585,256
IssuesEvent
2019-10-29 15:00:40
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
opened
Move USAJOBS data pull from Apply button to Next Steps "Continue"
Apply Process State Dept.
Who: What: Why: Acceptance Criteria: - Currently the USAJOBS one profile data is pulled for a student when they select "Apply" - Change the data pull to when they click "Continue" on the Next Steps page
1.0
Move USAJOBS data pull from Apply button to Next Steps "Continue" - Who: What: Why: Acceptance Criteria: - Currently the USAJOBS one profile data is pulled for a student when they select "Apply" - Change the data pull to when they click "Continue" on the Next Steps page
process
move usajobs data pull from apply button to next steps continue who what why acceptance criteria currently the usajobs one profile data is pulled for a student when they select apply change the data pull to when they click continue on the next steps page
1
129,603
12,414,793,498
IssuesEvent
2020-05-22 15:07:38
alpheios-project/alpheios-core
https://api.github.com/repos/alpheios-project/alpheios-core
closed
selective enabling of Alpheios on components
components documentation
for alpheios-project/components#129 we disabled alpheios on the panel and popup . sometimes we want to be able to enable it selectively. This requires some thought about the best way to do it but, for example, the ge'ez parser provides short definitions for its words in Latin. We should able to enable Alpheios on those definitions in the popup. One way to do this might be to look for the language code on the text that is displayed in a component and compare that to available languages to determine if Alpheios can be activated. But we might need even finer grained control of that. Probably an issue for after refactoring of component state and data.
1.0
selective enabling of Alpheios on components - for alpheios-project/components#129 we disabled alpheios on the panel and popup . sometimes we want to be able to enable it selectively. This requires some thought about the best way to do it but, for example, the ge'ez parser provides short definitions for its words in Latin. We should able to enable Alpheios on those definitions in the popup. One way to do this might be to look for the language code on the text that is displayed in a component and compare that to available languages to determine if Alpheios can be activated. But we might need even finer grained control of that. Probably an issue for after refactoring of component state and data.
non_process
selective enabling of alpheios on components for alpheios project components we disabled alpheios on the panel and popup sometimes we want to be able to enable it selectively this requires some thought about the best way to do it but for example the ge ez parser provides short definitions for its words in latin we should able to enable alpheios on those definitions in the popup one way to do this might be to look for the language code on the text that is displayed in a component and compare that to available languages to determine if alpheios can be activated but we might need even finer grained control of that probably an issue for after refactoring of component state and data
0
1,659
4,288,680,282
IssuesEvent
2016-07-17 16:31:09
log2timeline/plaso
https://api.github.com/repos/log2timeline/plaso
closed
Preprocessor not working for Windows
bug preprocessing
Preprocessor not working for Windows ``` 2015-12-26 20:53:12,946 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: sysregistry to /WINDOWS/system32/config 2015-12-26 20:53:12,951 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: systemroot to /WINDOWS 2015-12-26 20:53:12,955 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: windir to /WINDOWS 2015-12-26 20:53:12,985 [INFO] (MainProcess) PID:4027 <extraction_frontend> Parser filter expression changed to: win7 ``` Should be: ``` 2015-12-26 20:53:27,192 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: sysregistry to \WINDOWS\system32\config 2015-12-26 20:53:27,196 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: systemroot to \WINDOWS 2015-12-26 20:53:27,200 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: windir to \WINDOWS 2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: code_page to cp1252 2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: hostname to TEST 2015-12-26 20:53:27,521 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: programfiles to \Program Files ``` * [x] fix issue * ~~introduces a complication for preg and requires http://codereview.appspot.com/284880043/~~ * ~~https://codereview.appspot.com/276600043/~~
1.0
Preprocessor not working for Windows - Preprocessor not working for Windows ``` 2015-12-26 20:53:12,946 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: sysregistry to /WINDOWS/system32/config 2015-12-26 20:53:12,951 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: systemroot to /WINDOWS 2015-12-26 20:53:12,955 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: windir to /WINDOWS 2015-12-26 20:53:12,985 [INFO] (MainProcess) PID:4027 <extraction_frontend> Parser filter expression changed to: win7 ``` Should be: ``` 2015-12-26 20:53:27,192 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: sysregistry to \WINDOWS\system32\config 2015-12-26 20:53:27,196 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: systemroot to \WINDOWS 2015-12-26 20:53:27,200 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: windir to \WINDOWS 2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: code_page to cp1252 2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: hostname to TEST 2015-12-26 20:53:27,521 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: programfiles to \Program Files ``` * [x] fix issue * ~~introduces a complication for preg and requires http://codereview.appspot.com/284880043/~~ * ~~https://codereview.appspot.com/276600043/~~
process
preprocessor not working for windows preprocessor not working for windows mainprocess pid set attribute sysregistry to windows config mainprocess pid set attribute systemroot to windows mainprocess pid set attribute windir to windows mainprocess pid parser filter expression changed to should be mainprocess pid set attribute sysregistry to windows config mainprocess pid set attribute systemroot to windows mainprocess pid set attribute windir to windows mainprocess pid set attribute code page to mainprocess pid set attribute hostname to test mainprocess pid set attribute programfiles to program files fix issue introduces a complication for preg and requires
1
681,397
23,309,660,230
IssuesEvent
2022-08-08 06:58:49
oasis-engine/engine
https://api.github.com/repos/oasis-engine/engine
closed
Improve the text system
feature 2D high priority
Design: @GuoLei1990 , @singlecoder PR: @singlecoder @cptbtptpbcptdtptp PR reviewers: @GuoLei1990 , @cptbtptpbcptdtptp , @gz65555
1.0
Improve the text system - Design: @GuoLei1990 , @singlecoder PR: @singlecoder @cptbtptpbcptdtptp PR reviewers: @GuoLei1990 , @cptbtptpbcptdtptp , @gz65555
non_process
improve the text system design singlecoder pr singlecoder cptbtptpbcptdtptp pr reviewers cptbtptpbcptdtptp
0
9,376
12,374,399,803
IssuesEvent
2020-05-19 01:27:35
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
change triage party meeeting to google meet
kind/process priority/important-soon
for better experience with most of the maintainers
1.0
change triage party meeeting to google meet - for better experience with most of the maintainers
process
change triage party meeeting to google meet for better experience with most of the maintainers
1
3,052
6,044,561,726
IssuesEvent
2017-06-12 06:22:52
javabird25/long-hour-and-a-half
https://api.github.com/repos/javabird25/long-hour-and-a-half
closed
"THIS IS A BUG"
bug will be processed soon
When character is peeing during a class without underwear or/and outerwear, "THIS IS A BUG" will appear. ![image](https://user-images.githubusercontent.com/12773022/27000502-4c981818-4dcd-11e7-8e28-1f8d2ea527d0.png) Because there is no special wear handling in `ASK_TO_PEE` stage.
1.0
"THIS IS A BUG" - When character is peeing during a class without underwear or/and outerwear, "THIS IS A BUG" will appear. ![image](https://user-images.githubusercontent.com/12773022/27000502-4c981818-4dcd-11e7-8e28-1f8d2ea527d0.png) Because there is no special wear handling in `ASK_TO_PEE` stage.
process
this is a bug when character is peeing during a class without underwear or and outerwear this is a bug will appear because there is no special wear handling in ask to pee stage
1
12,114
14,740,543,303
IssuesEvent
2021-01-07 09:15:25
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Bogus Email Address
anc-process anp-2 ant-enhancement has attachment
In GitLab by @kdjstudios on Nov 19, 2018, 16:01 Hello Team, I just recently noticed one of the errors we are receiving: [SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg](/uploads/67c74a6f96dfbf330c7c90eba35a7cc1/SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg) It would appear that we are having an issue with being able to send to "none@none.com"; which is completely valid. My thoughts on this would be the following to resolve this. I did a quick internal search and we have over 15 customers that use this address, and over 600 accounts that use the address. The email is not a required field correct? So we should be able to remove them. 1. Notify operations that on a certain date we will be removing all "bogus" email addresses from the system. 2. If any accounts are setup for email invoicing that have a bogus email address. Notify Operations to correct those accounts accordingly. 3. On that date remove all bogus email addresses. 4. Update the validation on all email address fields to check for bogus email addresses before allowing to save. This will then in turn any accounts that did not get updated to a valid email and are setup for email invoicing will be displayed with an error to the site managers on the next billing cycle right?
1.0
Bogus Email Address - In GitLab by @kdjstudios on Nov 19, 2018, 16:01 Hello Team, I just recently noticed one of the errors we are receiving: [SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg](/uploads/67c74a6f96dfbf330c7c90eba35a7cc1/SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg) It would appear that we are having an issue with being able to send to "none@none.com"; which is completely valid. My thoughts on this would be the following to resolve this. I did a quick internal search and we have over 15 customers that use this address, and over 600 accounts that use the address. The email is not a required field correct? So we should be able to remove them. 1. Notify operations that on a certain date we will be removing all "bogus" email addresses from the system. 2. If any accounts are setup for email invoicing that have a bogus email address. Notify Operations to correct those accounts accordingly. 3. On that date remove all bogus email addresses. 4. Update the validation on all email address fields to check for bogus email addresses before allowing to save. This will then in turn any accounts that did not get updated to a valid email and are setup for email invoicing will be displayed with an error to the site managers on the next billing cycle right?
process
bogus email address in gitlab by kdjstudios on nov hello team i just recently noticed one of the errors we are receiving uploads sa billing error report customers update netsmtpfatalerror none none com recipient address rejected msg it would appear that we are having an issue with being able to send to none none com which is completely valid my thoughts on this would be the following to resolve this i did a quick internal search and we have over customers that use this address and over accounts that use the address the email is not a required field correct so we should be able to remove them notify operations that on a certain date we will be removing all bogus email addresses from the system if any accounts are setup for email invoicing that have a bogus email address notify operations to correct those accounts accordingly on that date remove all bogus email addresses update the validation on all email address fields to check for bogus email addresses before allowing to save this will then in turn any accounts that did not get updated to a valid email and are setup for email invoicing will be displayed with an error to the site managers on the next billing cycle right
1
193,146
22,216,072,041
IssuesEvent
2022-06-08 01:53:00
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2017-18552 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2017-18552 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in net/rds/af_rds.c in the Linux kernel before 4.11. There is an out of bounds write and read in the function rds_recv_track_latency. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18552>CVE-2017-18552</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v4.11-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-18552 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-18552 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in net/rds/af_rds.c in the Linux kernel before 4.11. There is an out of bounds write and read in the function rds_recv_track_latency. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18552>CVE-2017-18552</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v4.11-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net rds af rds c net rds af rds c vulnerability details an issue was discovered in net rds af rds c in the linux kernel before there is an out of bounds write and read in the function rds recv track latency publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
587
3,062,247,959
IssuesEvent
2015-08-16 11:45:30
yeoman/yeoman-app
https://api.github.com/repos/yeoman/yeoman-app
opened
Align with the Inquirer.js API
help-wanted renderer process
[Inquirer.js](https://github.com/SBoudrias/Inquirer.js) can handle either an Array of `objects` or an Array of `strings` for the attribute `choices`. Our current implementation supports only an Array of `objects`. To be compatible with the Inquirer.js API we should support this format too.
1.0
Align with the Inquirer.js API - [Inquirer.js](https://github.com/SBoudrias/Inquirer.js) can handle either an Array of `objects` or an Array of `strings` for the attribute `choices`. Our current implementation supports only an Array of `objects`. To be compatible with the Inquirer.js API we should support this format too.
process
align with the inquirer js api can handle either an array of objects or an array of strings for the attribute choices our current implementation supports only an array of objects to be compatible with the inquirer js api we should support this format too
1
13,582
16,131,672,904
IssuesEvent
2021-04-29 06:18:00
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Need to Support multiple Email Providers
Feature request Process: Fixed Process: Tested dev
Currently, we are supporting only Gmail but we need to support other email providers as well with MailGun API Support **Priority** High **Labels** Feature request
2.0
Need to Support multiple Email Providers - Currently, we are supporting only Gmail but we need to support other email providers as well with MailGun API Support **Priority** High **Labels** Feature request
process
need to support multiple email providers currently we are supporting only gmail but we need to support other email providers as well with mailgun api support priority high labels feature request
1
4,177
7,111,594,565
IssuesEvent
2018-01-17 14:41:29
mono/mono
https://api.github.com/repos/mono/mono
opened
Process.WaitForExit() gets slower the larger your open file limit (ulimit -n) is on Linux
area-Runtime: Process os-linux
## Steps to Reproduce 1. Compile this: `csc test.cs`: ```csharp public static class Program { public static int Main () { for (int i = 0; i < 100; i++) { var p = System.Diagnostics.Process.Start ("echo", "hello"); if (!p.WaitForExit (10000)) return 1; } return 0; } } ``` 2. `$ (ulimit -n 1000; time mono test.exe)` ``` ... real 0m0.402s user 0m0.028s sys 0m0.184s ``` 3. `$ (ulimit -n 100000; time mono test.exe)` ``` ... real 0m1.277s user 0m0.380s sys 0m0.592s ``` 4. `$ (ulimit -n 1000000; time mono test.exe)` ``` ... real 0m8.758s user 0m3.424s sys 0m4.424s ``` ## Current Behavior It gets slower the higher your ulimit -n is. Note: you may need to change your system settings (`/etc/security/limits.conf`) to allow higher limits. <!-- What is the current behavior? --> ## Expected Behavior Not getting slower. <!-- Please describe the behavior you are expecting --> ### On which platforms did you notice this [ ] macOS [ X ] Linux, Ubuntu 14.04/16.04 [ ] Windows **Version Used**: master, 5.8.088 and 4.2.1 (so either a very old bug or something outside our control) This was the root cause behind an issue (https://github.com/mono/mono/pull/6537) we had on Jenkins because the Azure Linux VM builders had ulimit -n set to 1048576.
1.0
Process.WaitForExit() gets slower the larger your open file limit (ulimit -n) is on Linux - ## Steps to Reproduce 1. Compile this: `csc test.cs`: ```csharp public static class Program { public static int Main () { for (int i = 0; i < 100; i++) { var p = System.Diagnostics.Process.Start ("echo", "hello"); if (!p.WaitForExit (10000)) return 1; } return 0; } } ``` 2. `$ (ulimit -n 1000; time mono test.exe)` ``` ... real 0m0.402s user 0m0.028s sys 0m0.184s ``` 3. `$ (ulimit -n 100000; time mono test.exe)` ``` ... real 0m1.277s user 0m0.380s sys 0m0.592s ``` 4. `$ (ulimit -n 1000000; time mono test.exe)` ``` ... real 0m8.758s user 0m3.424s sys 0m4.424s ``` ## Current Behavior It gets slower the higher your ulimit -n is. Note: you may need to change your system settings (`/etc/security/limits.conf`) to allow higher limits. <!-- What is the current behavior? --> ## Expected Behavior Not getting slower. <!-- Please describe the behavior you are expecting --> ### On which platforms did you notice this [ ] macOS [ X ] Linux, Ubuntu 14.04/16.04 [ ] Windows **Version Used**: master, 5.8.088 and 4.2.1 (so either a very old bug or something outside our control) This was the root cause behind an issue (https://github.com/mono/mono/pull/6537) we had on Jenkins because the Azure Linux VM builders had ulimit -n set to 1048576.
process
process waitforexit gets slower the larger your open file limit ulimit n is on linux steps to reproduce compile this csc test cs csharp public static class program public static int main for int i i i var p system diagnostics process start echo hello if p waitforexit return return ulimit n time mono test exe real user sys ulimit n time mono test exe real user sys ulimit n time mono test exe real user sys current behavior it gets slower the higher your ulimit n is note you may need to change your system settings etc security limits conf to allow higher limits what is the current behavior expected behavior not getting slower please describe the behavior you are expecting on which platforms did you notice this macos linux ubuntu windows version used master and so either a very old bug or something outside our control this was the root cause behind an issue we had on jenkins because the azure linux vm builders had ulimit n set to
1
14,611
17,754,602,078
IssuesEvent
2021-08-28 13:56:33
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
While resolving toolchains for target @org_tensorflow//tensorflow/tools/git:gen_git_source: No matching toolchains found for types @bazel_tools//tools/cpp:toolchain_type.
type: support / not a bug (process) team-Configurability untriaged
> ATTENTION! Please read and follow: > - if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to bazel-discuss@googlegroups.com > - if this is a _bug_ or _feature request_, fill the form below as best as you can. ### Description of the problem / feature request: This is a problem of toolchain resolution, and my platform is m1 apple mbp. I'm using org_tensorflow to build my project, and fail with the message above. after some research, I found that it was the @local_execution_config_platform that cause the problem. According to the BUILD file here: ` platform( constraint_values = [ "@bazel_tools//platforms:arm", "@bazel_tools//platforms:osx", ], ` platform should use @bazel_tools//platforms:arm to specify the cpu type, which in @bazel_tools//platform , is like: `alias( name = "arm", actual = "@platforms//cpu:arm", ) ` which in @platform , is like: ` constraint_value( name = "arm", constraint_setting = ":cpu", ) ` which leads to cpu type mismatch. According to the @platform, this value in local_execution_config_platform should be @bazel_tools//platforms:aarch64 instead of @bazel_tools//platforms:arm Now I have to manually change it to make it works. ### Feature requests: what underlying problem are you trying to solve with this feature? According to the similar problem here: https://github.com/tensorflow/tensorflow/issues/38840 this could be the problem of $MACHTYPE, it's `x86_64` in /bin/zsh and `arm64-apple-darwin20` in /bin/bash but when I change this to `arm64-apple-darwin20`, still no luck. Is it a bug or what? Would it be fixed in the future? ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. on a M1 apple mbp device, load @org_tensorflow, should reproduce the problem. ### What operating system are you running Bazel on? M1 apple mbp ### What's the output of `bazel info release`? this try to download 3.4.0 bazel but I already install 4.2.0rc3, but this problem can reproduce in 4.1.0
1.0
While resolving toolchains for target @org_tensorflow//tensorflow/tools/git:gen_git_source: No matching toolchains found for types @bazel_tools//tools/cpp:toolchain_type. - > ATTENTION! Please read and follow: > - if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to bazel-discuss@googlegroups.com > - if this is a _bug_ or _feature request_, fill the form below as best as you can. ### Description of the problem / feature request: This is a problem of toolchain resolution, and my platform is m1 apple mbp. I'm using org_tensorflow to build my project, and fail with the message above. after some research, I found that it was the @local_execution_config_platform that cause the problem. According to the BUILD file here: ` platform( constraint_values = [ "@bazel_tools//platforms:arm", "@bazel_tools//platforms:osx", ], ` platform should use @bazel_tools//platforms:arm to specify the cpu type, which in @bazel_tools//platform , is like: `alias( name = "arm", actual = "@platforms//cpu:arm", ) ` which in @platform , is like: ` constraint_value( name = "arm", constraint_setting = ":cpu", ) ` which leads to cpu type mismatch. According to the @platform, this value in local_execution_config_platform should be @bazel_tools//platforms:aarch64 instead of @bazel_tools//platforms:arm Now I have to manually change it to make it works. ### Feature requests: what underlying problem are you trying to solve with this feature? According to the similar problem here: https://github.com/tensorflow/tensorflow/issues/38840 this could be the problem of $MACHTYPE, it's `x86_64` in /bin/zsh and `arm64-apple-darwin20` in /bin/bash but when I change this to `arm64-apple-darwin20`, still no luck. Is it a bug or what? Would it be fixed in the future? ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. on a M1 apple mbp device, load @org_tensorflow, should reproduce the problem. ### What operating system are you running Bazel on? M1 apple mbp ### What's the output of `bazel info release`? this try to download 3.4.0 bazel but I already install 4.2.0rc3, but this problem can reproduce in 4.1.0
process
while resolving toolchains for target org tensorflow tensorflow tools git gen git source no matching toolchains found for types bazel tools tools cpp toolchain type attention please read and follow if this is a question about how to build test query deploy using bazel or a discussion starter send it to bazel discuss googlegroups com if this is a bug or feature request fill the form below as best as you can description of the problem feature request this is a problem of toolchain resolution and my platform is apple mbp i m using org tensorflow to build my project and fail with the message above after some research i found that it was the local execution config platform that cause the problem according to the build file here platform constraint values bazel tools platforms arm bazel tools platforms osx platform should use bazel tools platforms arm to specify the cpu type which in bazel tools platform is like alias name arm actual platforms cpu arm which in platform is like constraint value name arm constraint setting cpu which leads to cpu type mismatch according to the platform this value in local execution config platform should be bazel tools platforms instead of bazel tools platforms arm now i have to manually change it to make it works feature requests what underlying problem are you trying to solve with this feature according to the similar problem here this could be the problem of machtype it s in bin zsh and apple in bin bash but when i change this to apple still no luck is it a bug or what would it be fixed in the future bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible on a apple mbp device load org tensorflow should reproduce the problem what operating system are you running bazel on apple mbp what s the output of bazel info release this try to download bazel but i already install but this problem can reproduce in
1
319,170
23,759,642,214
IssuesEvent
2022-09-01 07:47:05
Unity-Technologies/com.unity.multiplayer.docs
https://api.github.com/repos/Unity-Technologies/com.unity.multiplayer.docs
closed
Feedback for /versioned_docs/version-1.0.0/migration/upgrade-guide.md
documentation IN JIRA
I've just upgraded from MLAPI to the new library and: - there's no mention of the replacements/alternatives for NetworkManager.Singleton.StopHost() and NetworkManager.Singleton.StopClient() in the documentation - messaging needs updating to say that CustomMessagingManager.RegisterNamedMessageHandler no longer works stand-alone but instead needs to be prefixed with "NetworkManager.Singleton." - string arrays can no longer be used for RPC parameters. The page titled "C# Primitives" still lists "string" even though the migration page says we're now supposed to use Unity.Collections.FixedStringxxxxBytes structs instead
1.0
Feedback for /versioned_docs/version-1.0.0/migration/upgrade-guide.md - I've just upgraded from MLAPI to the new library and: - there's no mention of the replacements/alternatives for NetworkManager.Singleton.StopHost() and NetworkManager.Singleton.StopClient() in the documentation - messaging needs updating to say that CustomMessagingManager.RegisterNamedMessageHandler no longer works stand-alone but instead needs to be prefixed with "NetworkManager.Singleton." - string arrays can no longer be used for RPC parameters. The page titled "C# Primitives" still lists "string" even though the migration page says we're now supposed to use Unity.Collections.FixedStringxxxxBytes structs instead
non_process
feedback for versioned docs version migration upgrade guide md i ve just upgraded from mlapi to the new library and there s no mention of the replacements alternatives for networkmanager singleton stophost and networkmanager singleton stopclient in the documentation messaging needs updating to say that custommessagingmanager registernamedmessagehandler no longer works stand alone but instead needs to be prefixed with networkmanager singleton string arrays can no longer be used for rpc parameters the page titled c primitives still lists string even though the migration page says we re now supposed to use unity collections fixedstringxxxxbytes structs instead
0
5,309
8,125,526,762
IssuesEvent
2018-08-16 21:14:54
MetaMask/metamask-extension
https://api.github.com/repos/MetaMask/metamask-extension
closed
Process: Reduce number of head branches
L09-process P2-sooner T02-refactor
we have a lot of "head" branches and we should reduce the number to reduce complexity: - master - uat - mascara-prod - newUI-flat i suggest: - master - uat <- mascara-prod + newUI-flat
1.0
Process: Reduce number of head branches - we have a lot of "head" branches and we should reduce the number to reduce complexity: - master - uat - mascara-prod - newUI-flat i suggest: - master - uat <- mascara-prod + newUI-flat
process
process reduce number of head branches we have a lot of head branches and we should reduce the number to reduce complexity master uat mascara prod newui flat i suggest master uat mascara prod newui flat
1
289,752
31,997,403,853
IssuesEvent
2023-09-21 10:02:54
Mend-developer-platform-load/8945136_33
https://api.github.com/repos/Mend-developer-platform-load/8945136_33
opened
CVE-2022-38750 (Medium) detected in snakeyaml-1.19.jar
Mend: dependency security vulnerability
## CVE-2022-38750 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.19.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - rocketmq-broker-4.6.0.jar (Root Library) - rocketmq-acl-4.6.0.jar - :x: **snakeyaml-1.19.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Mend-developer-platform-load/8945136_33/commit/3b8eebc04951436067314b2f84da6260dad2fbda">3b8eebc04951436067314b2f84da6260dad2fbda</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38750>CVE-2022-38750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution (org.yaml:snakeyaml): 1.31</p> <p>Direct dependency fix Resolution (org.apache.rocketmq:rocketmq-broker): 5.1.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation will be attempted for this issue.
True
CVE-2022-38750 (Medium) detected in snakeyaml-1.19.jar - ## CVE-2022-38750 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.19.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - rocketmq-broker-4.6.0.jar (Root Library) - rocketmq-acl-4.6.0.jar - :x: **snakeyaml-1.19.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Mend-developer-platform-load/8945136_33/commit/3b8eebc04951436067314b2f84da6260dad2fbda">3b8eebc04951436067314b2f84da6260dad2fbda</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38750>CVE-2022-38750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution (org.yaml:snakeyaml): 1.31</p> <p>Direct dependency fix Resolution (org.apache.rocketmq:rocketmq-broker): 5.1.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation will be attempted for this issue.
non_process
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy rocketmq broker jar root library rocketmq acl jar x snakeyaml jar vulnerable library found in head commit a href found in base branch main vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution org apache rocketmq rocketmq broker rescue worker helmet automatic remediation will be attempted for this issue
0
105,797
4,242,024,952
IssuesEvent
2016-07-06 18:09:18
easydigitaldownloads/easy-digital-downloads
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
closed
PayPal: Negative fees getting included twice
Bug High Priority Payments PayPal
When an order has negative fees (such as through Discounts Pro), the discounts are taken into account twice once the customer reaches PayPal. See screenshots below. ![mgcgbnhfodjfhfea](https://cloud.githubusercontent.com/assets/1034109/16620134/e0ae2d4e-4356-11e6-98df-ac10387c027c.png) ![digcpgldelgfnijd](https://cloud.githubusercontent.com/assets/1034109/16620289/7d11e0cc-4357-11e6-8ea4-bf2bf823a7bd.png) This is almost certainly due to changes introduced in #4632 and #4658 The issue affects PayPal Standard for sure but I believe it also affects at minimum the PayPal Pro / Express and 2Checkout gateways. We may have to fix it in the gateways in order to prevent undoing our bug fixes with negative fees.
1.0
PayPal: Negative fees getting included twice - When an order has negative fees (such as through Discounts Pro), the discounts are taken into account twice once the customer reaches PayPal. See screenshots below. ![mgcgbnhfodjfhfea](https://cloud.githubusercontent.com/assets/1034109/16620134/e0ae2d4e-4356-11e6-98df-ac10387c027c.png) ![digcpgldelgfnijd](https://cloud.githubusercontent.com/assets/1034109/16620289/7d11e0cc-4357-11e6-8ea4-bf2bf823a7bd.png) This is almost certainly due to changes introduced in #4632 and #4658 The issue affects PayPal Standard for sure but I believe it also affects at minimum the PayPal Pro / Express and 2Checkout gateways. We may have to fix it in the gateways in order to prevent undoing our bug fixes with negative fees.
non_process
paypal negative fees getting included twice when an order has negative fees such as through discounts pro the discounts are taken into account twice once the customer reaches paypal see screenshots below this is almost certainly due to changes introduced in and the issue affects paypal standard for sure but i believe it also affects at minimum the paypal pro express and gateways we may have to fix it in the gateways in order to prevent undoing our bug fixes with negative fees
0
9,517
12,498,729,083
IssuesEvent
2020-06-01 18:49:52
KasiaCieslik/sentiment-analysis-of-tweets-using-emoticons
https://api.github.com/repos/KasiaCieslik/sentiment-analysis-of-tweets-using-emoticons
opened
Write script that uses spacy to preprocess the tweet
data pre-processing high priority medium task
To Do: - use spacy to transform a tweet into a list of grammatical parts of a sentence (e.g., verb, adjective etc.) and find out other elementary aspects of a tweet (length etc.)
1.0
Write script that uses spacy to preprocess the tweet - To Do: - use spacy to transform a tweet into a list of grammatical parts of a sentence (e.g., verb, adjective etc.) and find out other elementary aspects of a tweet (length etc.)
process
write script that uses spacy to preprocess the tweet to do use spacy to transform a tweet into a list of grammatical parts of a sentence e g verb adjective etc and find out other elementary aspects of a tweet length etc
1
274,207
8,558,509,785
IssuesEvent
2018-11-08 18:28:39
Iridescent-CM/technovation-app
https://api.github.com/repos/Iridescent-CM/technovation-app
opened
As an RA I cannot save teams or judges that I invite to my event
Bug Priority - 1 bug
**Describe the bug** As an RA I can select a team to add to my event, but I can't save them **To Reproduce** Steps to reproduce the behavior: 1. Log in as Ra@ra.com 2. Click on Events 3. Find an event and click on 'manage teams' 4. Invite a team and try to click 'save teams' 5. Nothing happens 6. When the page is refreshed the teams have not been added **Expected behavior** I expect to be able to add the teams and judges. **Screenshots** ![screen shot 2018-11-08 at 10 19 32 am](https://user-images.githubusercontent.com/11083917/48219259-e689c200-e340-11e8-9c72-5ea5c1ba0202.png) **Desktop (please complete the following information):** - OS: MacOS Mojave - Browser: Chrome **Label** You MUST use the label menu to add 'bug' and '0 - To discuss' as a label (do *not* choose a bug priority label, devs will do that according to the Defect Priority matrix)
1.0
As an RA I cannot save teams or judges that I invite to my event - **Describe the bug** As an RA I can select a team to add to my event, but I can't save them **To Reproduce** Steps to reproduce the behavior: 1. Log in as Ra@ra.com 2. Click on Events 3. Find an event and click on 'manage teams' 4. Invite a team and try to click 'save teams' 5. Nothing happens 6. When the page is refreshed the teams have not been added **Expected behavior** I expect to be able to add the teams and judges. **Screenshots** ![screen shot 2018-11-08 at 10 19 32 am](https://user-images.githubusercontent.com/11083917/48219259-e689c200-e340-11e8-9c72-5ea5c1ba0202.png) **Desktop (please complete the following information):** - OS: MacOS Mojave - Browser: Chrome **Label** You MUST use the label menu to add 'bug' and '0 - To discuss' as a label (do *not* choose a bug priority label, devs will do that according to the Defect Priority matrix)
non_process
as an ra i cannot save teams or judges that i invite to my event describe the bug as an ra i can select a team to add to my event but i can t save them to reproduce steps to reproduce the behavior log in as ra ra com click on events find an event and click on manage teams invite a team and try to click save teams nothing happens when the page is refreshed the teams have not been added expected behavior i expect to be able to add the teams and judges screenshots desktop please complete the following information os macos mojave browser chrome label you must use the label menu to add bug and to discuss as a label do not choose a bug priority label devs will do that according to the defect priority matrix
0
387,951
11,472,636,300
IssuesEvent
2020-02-09 18:25:42
islos-efe-eme/auto-news
https://api.github.com/repos/islos-efe-eme/auto-news
opened
Implement "ping" endpoint in bots
good first issue new-feature priority:low slack telegram webhook
It's a useful thing to have a simple and fast way to check if a service is online or not, this doesn't means it's working properly but at least replying to external requests through the network. - [ ] Ping endpoint for Slack bot - [ ] Ping endpoint for Telegram bot (**blocked**) - [ ] Unit test for Slack bot - [ ] Unit test for Telegram bot (**blocked**) Expected request: ```js GET /ping ``` Expected response: ```js // status 200 { message: "Pong" } ```
1.0
Implement "ping" endpoint in bots - It's a useful thing to have a simple and fast way to check if a service is online or not, this doesn't means it's working properly but at least replying to external requests through the network. - [ ] Ping endpoint for Slack bot - [ ] Ping endpoint for Telegram bot (**blocked**) - [ ] Unit test for Slack bot - [ ] Unit test for Telegram bot (**blocked**) Expected request: ```js GET /ping ``` Expected response: ```js // status 200 { message: "Pong" } ```
non_process
implement ping endpoint in bots it s a useful thing to have a simple and fast way to check if a service is online or not this doesn t means it s working properly but at least replying to external requests through the network ping endpoint for slack bot ping endpoint for telegram bot blocked unit test for slack bot unit test for telegram bot blocked expected request js get ping expected response js status message pong
0
323
2,772,000,275
IssuesEvent
2015-05-02 07:10:12
FG-Team/HCJ-Website-Builder
https://api.github.com/repos/FG-Team/HCJ-Website-Builder
closed
[DEPRECATED] Tree structure
Feature No Processing
-Klasse TreeView: Graphik für Baumstruktur, beinhaltet einen javax.swing.JTree -Klasse TreeModel: Datenmodell ->Inhalt in Vector<TreeNode> gespeichert. -Klasse TreeNode: Repräsentation eines Baumastes/-blattes, beinhaltet Namen und Ebene des Blattes. -Listener für Tree/Model -Klasse TreePresenter: presenter alle im core Package
1.0
[DEPRECATED] Tree structure - -Klasse TreeView: Graphik für Baumstruktur, beinhaltet einen javax.swing.JTree -Klasse TreeModel: Datenmodell ->Inhalt in Vector<TreeNode> gespeichert. -Klasse TreeNode: Repräsentation eines Baumastes/-blattes, beinhaltet Namen und Ebene des Blattes. -Listener für Tree/Model -Klasse TreePresenter: presenter alle im core Package
process
tree structure klasse treeview graphik für baumstruktur beinhaltet einen javax swing jtree klasse treemodel datenmodell inhalt in vector gespeichert klasse treenode repräsentation eines baumastes blattes beinhaltet namen und ebene des blattes listener für tree model klasse treepresenter presenter alle im core package
1
13,536
16,067,667,023
IssuesEvent
2021-04-23 22:17:12
Azure/azure-event-hubs-java
https://api.github.com/repos/Azure/azure-event-hubs-java
closed
Unable to pass RetryPolicy to EventProcessorHost constructor with AzureStorageCheckpoint
EventProcessorHost enhancement
## Actual Behavior EventProcessorHost has only one constructor which accepts RetryPolicy: ``` EventProcessorHost( final String hostName, final String eventHubPath, final String consumerGroupName, final String eventHubConnectionString, ICheckpointManager checkpointManager, ILeaseManager leaseManager, ScheduledExecutorService executorService, RetryPolicy retryPolicy) ``` As you can see I need to pass `ICheckpointManager` and `ILeaseManager`, but in case I want to reuse built-in implementation `AzureStorageCheckpointLeaseManager` I can't do it, because it is non-public class with package-level constructor. ## Expected Behavior `EventProcessorHost` should have additional constructor overload where it is possible to pass RetryPolicy, ScheduledExecutorService and use Azure Storage checkpointing/leasing ## Versions - OS platform and version: MacOs - Maven package version or commit ID: azure-eventhubs-eph: 2.0.1
1.0
Unable to pass RetryPolicy to EventProcessorHost constructor with AzureStorageCheckpoint - ## Actual Behavior EventProcessorHost has only one constructor which accepts RetryPolicy: ``` EventProcessorHost( final String hostName, final String eventHubPath, final String consumerGroupName, final String eventHubConnectionString, ICheckpointManager checkpointManager, ILeaseManager leaseManager, ScheduledExecutorService executorService, RetryPolicy retryPolicy) ``` As you can see I need to pass `ICheckpointManager` and `ILeaseManager`, but in case I want to reuse built-in implementation `AzureStorageCheckpointLeaseManager` I can't do it, because it is non-public class with package-level constructor. ## Expected Behavior `EventProcessorHost` should have additional constructor overload where it is possible to pass RetryPolicy, ScheduledExecutorService and use Azure Storage checkpointing/leasing ## Versions - OS platform and version: MacOs - Maven package version or commit ID: azure-eventhubs-eph: 2.0.1
process
unable to pass retrypolicy to eventprocessorhost constructor with azurestoragecheckpoint actual behavior eventprocessorhost has only one constructor which accepts retrypolicy eventprocessorhost final string hostname final string eventhubpath final string consumergroupname final string eventhubconnectionstring icheckpointmanager checkpointmanager ileasemanager leasemanager scheduledexecutorservice executorservice retrypolicy retrypolicy as you can see i need to pass icheckpointmanager and ileasemanager but in case i want to reuse built in implementation azurestoragecheckpointleasemanager i can t do it because it is non public class with package level constructor expected behavior eventprocessorhost should have additional constructor overload where it is possible to pass retrypolicy scheduledexecutorservice and use azure storage checkpointing leasing versions os platform and version macos maven package version or commit id azure eventhubs eph
1
151,470
12,037,258,214
IssuesEvent
2020-04-13 21:26:43
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
l3_interfaces of junos_facts gather_network_resources returns error
affects_2.9 bug module needs_triage networking python3 support:core support:network system test traceback
<!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Explain the problem briefly below --> - l3_interfaces submodule of junos_facts: gather_network_resources: throws a AttributeError: 'NoneType' object has no attribute 'text' error. These submodules of junos_facts work just fine in the same playbook: junos_facts: gather_network_resources: - interfaces - l2_interfaces - lag_interfaces ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> junos_facts: gather_network_resources: - l3_interfaces ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ``' username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible --version ansible 2.9.6 config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible-config dump --only-changed ANSIBLE_PIPELINING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = True DEFAULT_GATHERING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = explicit DEFAULT_HOST_LIST(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = [u'/opt/adi/nsg_hosts.py'] DEFAULT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260 DEFAULT_TRANSPORT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = local ENABLE_TASK_DEBUGGER(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False HOST_KEY_CHECKING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False PERSISTENT_COMMAND_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260 PERSISTENT_CONNECT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260 RETRY_FILES_ENABLED(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ cat /etc/centos-release CentOS Linux release 7.6.1810 (Core) Running ansible 2.9.6, python 3. Also tested in ansible 2.9.1 ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> Run the playbook, without -l3_interfaces called the module will run and print out gathered facts successfully. <!--- Paste example playbooks or commands between quotes below --> ```yaml --- - name: Gathers juniper 1a switch facts gather_facts: false hosts: juniperhost serial: 10 vars_prompt: - name: username prompt: "Enter Username" private: no - name: password private: yes prompt: "Enter Password" confirm: yes vars: cli: host: "{{ inventory_hostname }}" username: "{{ username }}" password: "{{ password }}" pre_tasks: - name: gather facts junos_facts: gather_network_resources: - interfaces - l2_interfaces - lag_interfaces - l3_interfaces provider: "{{ cli }}" - name: dubug debug: msg: "{{ ansible_facts }}" ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> Expected that data will be gathered, printed to the screen ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ rpb pb.gatherJunosFacts.yaml -vvvv ansible-playbook 2.9.6 config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)] Using /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg as config file setting up inventory plugins host_list declined parsing /opt/adi/nsg_hosts.py as it did not pass its verify_file() method Parsed /opt/adi/nsg_hosts.py inventory source with script plugin Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/site-packages/ansible/plugins/callback/default.pyc PLAYBOOK: pb.gatherJunosFacts.yaml ******************************************************************************************************************************************************************************* Positional arguments: pb.gatherJunosFacts.yaml become_method: sudo inventory: (u'/opt/adi/nsg_hosts.py',) forks: 5 tags: (u'all',) verbosity: 4 connection: local timeout: 260 1 plays in pb.gatherJunosFacts.yaml Enter Username: Enter Password: confirm Enter Password: PLAY [Gathers juniper 1a switch facts] *************************************************************************************************************************************************************************** TASK [gather facts] ********************************************************************************************************************************************************************************************** task path: /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/pb.gatherJunosFacts.yaml:24 <js922-1a> using connection plugin netconf (was local) <js922-1a> starting connection from persistent connection plugin <js922-1a> local domain socket does not exist, starting it <js922-1a> control socket path is /home/username/.ansible/pc/314f4ddcc8 <js922-1a> local domain socket listeners started successfully <js922-1a> loaded netconf plugin junos from path /usr/lib/python2.7/site-packages/ansible/plugins/netconf/junos.py for network_os junos <js922-1a> <js922-1a> local domain socket path is /home/username/.ansible/pc/314f4ddcc8 <js922-1a> socket_path: /home/username/.ansible/pc/314f4ddcc8 <js922-1a> Attempting python interpreter discovery <js922-1a> ESTABLISH LOCAL CONNECTION FOR USER: username <js922-1a> EXEC /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'/usr/bin/python'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'python3.6'"'"'; command -v '"'"'python3.5'"'"'; command -v '"'"'python2.7'"'"'; command -v '"'"'python2.6'"'"'; command -v '"'"'/usr/libexec/platform-python'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'python'"'"'; echo ENDFOUND && sleep 0' <js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0' Using module file /usr/lib/python2.7/site-packages/ansible/modules/network/junos/junos_facts.py Pipelining is enabled. <js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0' The full traceback is: Traceback (most recent call last): File "<stdin>", line 102, in <module> File "<stdin>", line 94, in _ansiballz_main File "<stdin>", line 40, in invoke_module File "/usr/lib64/python2.7/runpy.py", line 176, in run_module fname, loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 82, in _run_module_code mod_name, mod_fname, mod_loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 135, in <module> File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 126, in main File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py", line 62, in get_facts File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py", line 105, in get_network_resources_facts File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 72, in populate_facts File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 93, in parse_l3_if_resources AttributeError: 'NoneType' object has no attribute 'text' fatal: [js922-1a]: FAILED! => { "ansible_facts": { "discovered_interpreter_python": "/usr/bin/python" }, "changed": false, "module_stderr": "Traceback (most recent call last):\n File \"<stdin>\", line 102, in <module>\n File \"<stdin>\", line 94, in _ansiballz_main\n File \"<stdin>\", line 40, in invoke_module\n File \"/usr/lib64/python2.7/runpy.py\", line 176, in run_module\n fname, loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 82, in _run_module_code\n mod_name, mod_fname, mod_loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 72, in _run_code\n exec code in run_globals\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 135, in <module>\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 126, in main\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py\", line 62, in get_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py\", line 105, in get_network_resources_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 72, in populate_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 93, in parse_l3_if_resources\nAttributeError: 'NoneType' object has no attribute 'text'\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1 } PLAY RECAP ******************************************************************************************************************************************************************************************************* js922-1a : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 <!--- Paste verbatim command output between quotes --> ```paste below ansible-playbook pb.gatherJunosFacts.yaml -vvvv ```
1.0
l3_interfaces of junos_facts gather_network_resources returns error - <!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Explain the problem briefly below --> - l3_interfaces submodule of junos_facts: gather_network_resources: throws a AttributeError: 'NoneType' object has no attribute 'text' error. These submodules of junos_facts work just fine in the same playbook: junos_facts: gather_network_resources: - interfaces - l2_interfaces - lag_interfaces ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> junos_facts: gather_network_resources: - l3_interfaces ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ``' username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible --version ansible 2.9.6 config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible-config dump --only-changed ANSIBLE_PIPELINING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = True DEFAULT_GATHERING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = explicit DEFAULT_HOST_LIST(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = [u'/opt/adi/nsg_hosts.py'] DEFAULT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260 DEFAULT_TRANSPORT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = local ENABLE_TASK_DEBUGGER(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False HOST_KEY_CHECKING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False PERSISTENT_COMMAND_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260 PERSISTENT_CONNECT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260 RETRY_FILES_ENABLED(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ cat /etc/centos-release CentOS Linux release 7.6.1810 (Core) Running ansible 2.9.6, python 3. Also tested in ansible 2.9.1 ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> Run the playbook, without -l3_interfaces called the module will run and print out gathered facts successfully. <!--- Paste example playbooks or commands between quotes below --> ```yaml --- - name: Gathers juniper 1a switch facts gather_facts: false hosts: juniperhost serial: 10 vars_prompt: - name: username prompt: "Enter Username" private: no - name: password private: yes prompt: "Enter Password" confirm: yes vars: cli: host: "{{ inventory_hostname }}" username: "{{ username }}" password: "{{ password }}" pre_tasks: - name: gather facts junos_facts: gather_network_resources: - interfaces - l2_interfaces - lag_interfaces - l3_interfaces provider: "{{ cli }}" - name: dubug debug: msg: "{{ ansible_facts }}" ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> Expected that data will be gathered, printed to the screen ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ rpb pb.gatherJunosFacts.yaml -vvvv ansible-playbook 2.9.6 config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)] Using /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg as config file setting up inventory plugins host_list declined parsing /opt/adi/nsg_hosts.py as it did not pass its verify_file() method Parsed /opt/adi/nsg_hosts.py inventory source with script plugin Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/site-packages/ansible/plugins/callback/default.pyc PLAYBOOK: pb.gatherJunosFacts.yaml ******************************************************************************************************************************************************************************* Positional arguments: pb.gatherJunosFacts.yaml become_method: sudo inventory: (u'/opt/adi/nsg_hosts.py',) forks: 5 tags: (u'all',) verbosity: 4 connection: local timeout: 260 1 plays in pb.gatherJunosFacts.yaml Enter Username: Enter Password: confirm Enter Password: PLAY [Gathers juniper 1a switch facts] *************************************************************************************************************************************************************************** TASK [gather facts] ********************************************************************************************************************************************************************************************** task path: /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/pb.gatherJunosFacts.yaml:24 <js922-1a> using connection plugin netconf (was local) <js922-1a> starting connection from persistent connection plugin <js922-1a> local domain socket does not exist, starting it <js922-1a> control socket path is /home/username/.ansible/pc/314f4ddcc8 <js922-1a> local domain socket listeners started successfully <js922-1a> loaded netconf plugin junos from path /usr/lib/python2.7/site-packages/ansible/plugins/netconf/junos.py for network_os junos <js922-1a> <js922-1a> local domain socket path is /home/username/.ansible/pc/314f4ddcc8 <js922-1a> socket_path: /home/username/.ansible/pc/314f4ddcc8 <js922-1a> Attempting python interpreter discovery <js922-1a> ESTABLISH LOCAL CONNECTION FOR USER: username <js922-1a> EXEC /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'/usr/bin/python'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'python3.6'"'"'; command -v '"'"'python3.5'"'"'; command -v '"'"'python2.7'"'"'; command -v '"'"'python2.6'"'"'; command -v '"'"'/usr/libexec/platform-python'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'python'"'"'; echo ENDFOUND && sleep 0' <js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0' Using module file /usr/lib/python2.7/site-packages/ansible/modules/network/junos/junos_facts.py Pipelining is enabled. <js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0' The full traceback is: Traceback (most recent call last): File "<stdin>", line 102, in <module> File "<stdin>", line 94, in _ansiballz_main File "<stdin>", line 40, in invoke_module File "/usr/lib64/python2.7/runpy.py", line 176, in run_module fname, loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 82, in _run_module_code mod_name, mod_fname, mod_loader, pkg_name) File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 135, in <module> File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 126, in main File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py", line 62, in get_facts File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py", line 105, in get_network_resources_facts File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 72, in populate_facts File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 93, in parse_l3_if_resources AttributeError: 'NoneType' object has no attribute 'text' fatal: [js922-1a]: FAILED! => { "ansible_facts": { "discovered_interpreter_python": "/usr/bin/python" }, "changed": false, "module_stderr": "Traceback (most recent call last):\n File \"<stdin>\", line 102, in <module>\n File \"<stdin>\", line 94, in _ansiballz_main\n File \"<stdin>\", line 40, in invoke_module\n File \"/usr/lib64/python2.7/runpy.py\", line 176, in run_module\n fname, loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 82, in _run_module_code\n mod_name, mod_fname, mod_loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 72, in _run_code\n exec code in run_globals\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 135, in <module>\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 126, in main\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py\", line 62, in get_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py\", line 105, in get_network_resources_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 72, in populate_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 93, in parse_l3_if_resources\nAttributeError: 'NoneType' object has no attribute 'text'\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1 } PLAY RECAP ******************************************************************************************************************************************************************************************************* js922-1a : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0 <!--- Paste verbatim command output between quotes --> ```paste below ansible-playbook pb.gatherJunosFacts.yaml -vvvv ```
non_process
interfaces of junos facts gather network resources returns error summary interfaces submodule of junos facts gather network resources throws a attributeerror nonetype object has no attribute text error these submodules of junos facts work just fine in the same playbook junos facts gather network resources interfaces interfaces lag interfaces issue type bug report component name junos facts gather network resources interfaces ansible version username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible version ansible config file opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default oct configuration paste below username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible config dump only changed ansible pipelining opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg true default gathering opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg explicit default host list opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg default timeout opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg default transport opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg local enable task debugger opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg false host key checking opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg false persistent command timeout opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg persistent connect timeout opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg retry files enabled opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg false username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier os environment username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier cat etc centos release centos linux release core running ansible python also tested in ansible steps to reproduce run the playbook without interfaces called the module will run and print out gathered facts successfully yaml name gathers juniper switch facts gather facts false hosts juniperhost serial vars prompt name username prompt enter username private no name password private yes prompt enter password confirm yes vars cli host inventory hostname username username password password pre tasks name gather facts junos facts gather network resources interfaces interfaces lag interfaces interfaces provider cli name dubug debug msg ansible facts expected results expected that data will be gathered printed to the screen actual results username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier rpb pb gatherjunosfacts yaml vvvv ansible playbook config file opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible playbook python version default oct using opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg as config file setting up inventory plugins host list declined parsing opt adi nsg hosts py as it did not pass its verify file method parsed opt adi nsg hosts py inventory source with script plugin loading callback plugin default of type stdout from usr lib site packages ansible plugins callback default pyc playbook pb gatherjunosfacts yaml positional arguments pb gatherjunosfacts yaml become method sudo inventory u opt adi nsg hosts py forks tags u all verbosity connection local timeout plays in pb gatherjunosfacts yaml enter username enter password confirm enter password play task task path opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier pb gatherjunosfacts yaml using connection plugin netconf was local starting connection from persistent connection plugin local domain socket does not exist starting it control socket path is home username ansible pc local domain socket listeners started successfully loaded netconf plugin junos from path usr lib site packages ansible plugins netconf junos py for network os junos local domain socket path is home username ansible pc socket path home username ansible pc attempting python interpreter discovery establish local connection for user username exec bin sh c echo platform uname echo found command v usr bin python command v command v command v command v command v command v usr libexec platform python command v usr bin command v python echo endfound sleep exec bin sh c usr bin python sleep using module file usr lib site packages ansible modules network junos junos facts py pipelining is enabled exec bin sh c usr bin python sleep the full traceback is traceback most recent call last file line in file line in ansiballz main file line in invoke module file usr runpy py line in run module fname loader pkg name file usr runpy py line in run module code mod name mod fname mod loader pkg name file usr runpy py line in run code exec code in run globals file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in main file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts facts py line in get facts file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network common facts facts py line in get network resources facts file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in populate facts file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in parse if resources attributeerror nonetype object has no attribute text fatal failed ansible facts discovered interpreter python usr bin python changed false module stderr traceback most recent call last n file line in n file line in ansiballz main n file line in invoke module n file usr runpy py line in run module n fname loader pkg name n file usr runpy py line in run module code n mod name mod fname mod loader pkg name n file usr runpy py line in run code n exec code in run globals n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in main n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts facts py line in get facts n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network common facts facts py line in get network resources facts n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in populate facts n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in parse if resources nattributeerror nonetype object has no attribute text n module stdout msg module failure nsee stdout stderr for the exact error rc play recap ok changed unreachable failed skipped rescued ignored paste below ansible playbook pb gatherjunosfacts yaml vvvv
0
51,493
12,740,528,507
IssuesEvent
2020-06-26 02:53:32
carla-simulator/carla
https://api.github.com/repos/carla-simulator/carla
closed
Docker image for building Carla on Linux
backlog build system feature request help wanted
Make a Dockerfile that allows us building an image with a Linux environment set up for building Carla. The main complication is having Unreal built without redistributing the code/binaries.
1.0
Docker image for building Carla on Linux - Make a Dockerfile that allows us building an image with a Linux environment set up for building Carla. The main complication is having Unreal built without redistributing the code/binaries.
non_process
docker image for building carla on linux make a dockerfile that allows us building an image with a linux environment set up for building carla the main complication is having unreal built without redistributing the code binaries
0
34,806
16,685,373,121
IssuesEvent
2021-06-08 07:28:59
PrestaShop/PrestaShop
https://api.github.com/repos/PrestaShop/PrestaShop
closed
Update stock via webserivice with memcache exception and memory leak
1.7.7.2 Bug NMI Performance Stocks WS
#### Describe the bug When i am updating stock via web service with memcached enable i am getting exception and also i have memory leak (ram 128GB) When memcached is disabled everything works as expected. #### Expected behavior update stock #### Steps to Reproduce Steps to reproduce the behavior: 1. Enable memcached cache 2. Try to update stock via web service 3. HttpStatusCode: InternalServerError 4. Invalid argument supplied for foreach() (.../httpdocs/ps1772/classes/cache/Cache.php line 567 **Screenshots** Request: ``` <?xml version="1.0" encoding="utf-8"?> <prestashop> <stock_available> <id>52920</id> <id_product>1024345</id_product> <id_product_attribute>0</id_product_attribute> <id_shop>1</id_shop> <id_shop_group>0</id_shop_group> <quantity>1</quantity> <depends_on_stock>0</depends_on_stock> <out_of_stock>2</out_of_stock> </stock_available> </prestashop> ``` Response: ``` <?xml version="1.0" encoding="UTF-8"?> <prestashop xmlns:xlink="http://www.w3.org/1999/xlink"> <errors> <error> <code><![CDATA[3]]></code> <message><![CDATA[[PHP Warning #2] Invalid argument supplied for foreach() (/var/www/vhosts/xxx/httpdocs/ps1772/classes/cache/Cache.php, line 567)]]></message> </error> </errors> </prestashop> HttpStatusCode: InternalServerError ``` #### Additional information - Web server Apache + NGINX Reverse proxy + php fpm on apache. - PrestaShop version: 1.7.7.2 - Server information Linux #1 SMP Debian 4.19.171-2 (2021-01-30) x86_64 - Server software version: Apache - PHP version: 7.3.28 - Memory limit: 512M - Max execution time: 300 - Upload Max File size: 24M - MySQL version: 10.3.27-MariaDB-0+deb10u1
True
Update stock via webserivice with memcache exception and memory leak - #### Describe the bug When i am updating stock via web service with memcached enable i am getting exception and also i have memory leak (ram 128GB) When memcached is disabled everything works as expected. #### Expected behavior update stock #### Steps to Reproduce Steps to reproduce the behavior: 1. Enable memcached cache 2. Try to update stock via web service 3. HttpStatusCode: InternalServerError 4. Invalid argument supplied for foreach() (.../httpdocs/ps1772/classes/cache/Cache.php line 567 **Screenshots** Request: ``` <?xml version="1.0" encoding="utf-8"?> <prestashop> <stock_available> <id>52920</id> <id_product>1024345</id_product> <id_product_attribute>0</id_product_attribute> <id_shop>1</id_shop> <id_shop_group>0</id_shop_group> <quantity>1</quantity> <depends_on_stock>0</depends_on_stock> <out_of_stock>2</out_of_stock> </stock_available> </prestashop> ``` Response: ``` <?xml version="1.0" encoding="UTF-8"?> <prestashop xmlns:xlink="http://www.w3.org/1999/xlink"> <errors> <error> <code><![CDATA[3]]></code> <message><![CDATA[[PHP Warning #2] Invalid argument supplied for foreach() (/var/www/vhosts/xxx/httpdocs/ps1772/classes/cache/Cache.php, line 567)]]></message> </error> </errors> </prestashop> HttpStatusCode: InternalServerError ``` #### Additional information - Web server Apache + NGINX Reverse proxy + php fpm on apache. - PrestaShop version: 1.7.7.2 - Server information Linux #1 SMP Debian 4.19.171-2 (2021-01-30) x86_64 - Server software version: Apache - PHP version: 7.3.28 - Memory limit: 512M - Max execution time: 300 - Upload Max File size: 24M - MySQL version: 10.3.27-MariaDB-0+deb10u1
non_process
update stock via webserivice with memcache exception and memory leak describe the bug when i am updating stock via web service with memcached enable i am getting exception and also i have memory leak ram when memcached is disabled everything works as expected expected behavior update stock steps to reproduce steps to reproduce the behavior enable memcached cache try to update stock via web service httpstatuscode internalservererror invalid argument supplied for foreach httpdocs classes cache cache php line screenshots request response prestashop xmlns xlink httpstatuscode internalservererror additional information web server apache nginx reverse proxy php fpm on apache prestashop version server information linux smp debian server software version apache php version memory limit max execution time upload max file size mysql version mariadb
0
65,328
12,553,756,046
IssuesEvent
2020-06-06 23:17:45
Pokecube-Development/Pokecube-Issues-and-Wiki
https://api.github.com/repos/Pokecube-Development/Pokecube-Issues-and-Wiki
closed
Metagross dupped
1.14.x 1.15.2 Bug - Code Fixed
#### Issue Description: Metagross ressurrect and dupped #### What happens: my metagross has a bug, when it died to some golbats when I was upgrading I ran to cure it so I wouldn't die, when I realized there was another metagross following me that had resurrected #### What you expected to happen: the metagross no duppe on die #### Steps to reproduce: 1.the pokemon die 2.pokebolla returns to your inventory 3.the body that does not disappear resurrects and the pokemon duppe ... ____ #### Affected Versions (Do *not* use "latest"): Replace with a list of all mods you have in. - Pokecube AIO: - Minecraft: 1.15.2 - Forge: 31.1.18
1.0
Metagross dupped - #### Issue Description: Metagross ressurrect and dupped #### What happens: my metagross has a bug, when it died to some golbats when I was upgrading I ran to cure it so I wouldn't die, when I realized there was another metagross following me that had resurrected #### What you expected to happen: the metagross no duppe on die #### Steps to reproduce: 1.the pokemon die 2.pokebolla returns to your inventory 3.the body that does not disappear resurrects and the pokemon duppe ... ____ #### Affected Versions (Do *not* use "latest"): Replace with a list of all mods you have in. - Pokecube AIO: - Minecraft: 1.15.2 - Forge: 31.1.18
non_process
metagross dupped issue description metagross ressurrect and dupped what happens my metagross has a bug when it died to some golbats when i was upgrading i ran to cure it so i wouldn t die when i realized there was another metagross following me that had resurrected what you expected to happen the metagross no duppe on die steps to reproduce the pokemon die pokebolla returns to your inventory the body that does not disappear resurrects and the pokemon duppe affected versions do not use latest replace with a list of all mods you have in pokecube aio minecraft forge
0
476
2,911,385,686
IssuesEvent
2015-06-22 09:13:47
haskell-distributed/distributed-process-simplelocalnet
https://api.github.com/repos/haskell-distributed/distributed-process-simplelocalnet
opened
SimpleLocalnet should remove peers that no longer exist
distributed-process-simplelocalnet Feature Request
_From @edsko on November 7, 2012 15:33_ This is not a *major* issue, because even if we do, there is still no guarantee that a peer still exists after it has been returned by `findPeers`. _Copied from original issue: haskell-distributed/distributed-process#66_
1.0
SimpleLocalnet should remove peers that no longer exist - _From @edsko on November 7, 2012 15:33_ This is not a *major* issue, because even if we do, there is still no guarantee that a peer still exists after it has been returned by `findPeers`. _Copied from original issue: haskell-distributed/distributed-process#66_
process
simplelocalnet should remove peers that no longer exist from edsko on november this is not a major issue because even if we do there is still no guarantee that a peer still exists after it has been returned by findpeers copied from original issue haskell distributed distributed process
1
10,304
4,038,834,705
IssuesEvent
2016-05-20 00:14:41
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
opened
Object+Barcode to Container
Function-ContainerOrBarcode
We are starting a large project to add barcodes to existing tissue vials. At least initially the upload of barcodes to specimen records with be done using the Object+Barcode to Container tool in Object Tracking. However, there a some issues that I would like to see revised for ease of use and error control. 1) "Parent Cont Type" should be changed to read "Move to Container Type" or "Barcode Container Type". My student was adding in cryovial barcodes and misunderstood this. She assumed the "parent" of a cryovial was a freezer box - this resulted in all the cryovials being converted to freezer boxes inside the freezer box positions. 2) The "Move to Container Type" field above should have a null default value so as not to change the type of an existing container unless the user specifically intends to do so (it certainly shouldn't be set on "Nalgene" as the automatic default requiring manual change to a different value.) 3) Please if at all possible add a live link to the specimen record from this page, so that the specimen record can be accessed easily for verification.
1.0
Object+Barcode to Container - We are starting a large project to add barcodes to existing tissue vials. At least initially the upload of barcodes to specimen records with be done using the Object+Barcode to Container tool in Object Tracking. However, there a some issues that I would like to see revised for ease of use and error control. 1) "Parent Cont Type" should be changed to read "Move to Container Type" or "Barcode Container Type". My student was adding in cryovial barcodes and misunderstood this. She assumed the "parent" of a cryovial was a freezer box - this resulted in all the cryovials being converted to freezer boxes inside the freezer box positions. 2) The "Move to Container Type" field above should have a null default value so as not to change the type of an existing container unless the user specifically intends to do so (it certainly shouldn't be set on "Nalgene" as the automatic default requiring manual change to a different value.) 3) Please if at all possible add a live link to the specimen record from this page, so that the specimen record can be accessed easily for verification.
non_process
object barcode to container we are starting a large project to add barcodes to existing tissue vials at least initially the upload of barcodes to specimen records with be done using the object barcode to container tool in object tracking however there a some issues that i would like to see revised for ease of use and error control parent cont type should be changed to read move to container type or barcode container type my student was adding in cryovial barcodes and misunderstood this she assumed the parent of a cryovial was a freezer box this resulted in all the cryovials being converted to freezer boxes inside the freezer box positions the move to container type field above should have a null default value so as not to change the type of an existing container unless the user specifically intends to do so it certainly shouldn t be set on nalgene as the automatic default requiring manual change to a different value please if at all possible add a live link to the specimen record from this page so that the specimen record can be accessed easily for verification
0
62,500
15,278,925,586
IssuesEvent
2021-02-23 02:41:13
Elaniobro/MMM-nyc-transit
https://api.github.com/repos/Elaniobro/MMM-nyc-transit
closed
[🐛 BUG] - Please put your title here.
:building_construction: dependencies :white_medium_square: minor
**Describe the bug** Bump version of station pkg once npm published.
1.0
[🐛 BUG] - Please put your title here. - **Describe the bug** Bump version of station pkg once npm published.
non_process
please put your title here describe the bug bump version of station pkg once npm published
0
8,231
11,415,575,212
IssuesEvent
2020-02-02 12:02:10
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
Changing `scss`/`css` file breaks style on livereload
:bug: Bug CSS Preprocessing HMR Stale
**Choose one:** is this a 🐛 bug report or 🙋 feature request? 🐛 bug report ### 🎛 Configuration (.babelrc, package.json, cli command) The project I'm working is available [here](https://github.com/wescoder/tavern), you can checkout and reproduce the issue on it. <!--- If describing a bug, tell us what your babel configuration looks like --> .babelrc: ```js { "presets": ["env", "preact"], "plugins": [ "transform-function-bind", ["transform-react-jsx", { "pragma": "h" }], ["module-resolver", { "root": ["."], "alias": { "@app": "./src" } }], "transform-object-rest-spread" ] } ``` .postcssrc: ``` { "modules": true, "plugins": { "autoprefixer": { "grid": true } } } ``` ### 🤔 Expected Behavior Changing a `scss` file with css modules enabled that is "named-imported" (`import s from './style.scss'`) in js and used in `className` by `preact`/`react`/etc should update the js that requires it providing the new `className` variable to it, without needing a manual full page reload. ### 😯 Current Behavior Changing the `scss` updates the stylesheet causing the network fetch for the new asset but not for the js file, so all styles changed breaks on the page since the old classes doesn't exist anymore, to see the changes a manual full page reload is needed. ### 💁 Possible Solution I really don't have a clue about this... Sorry... ### 🔦 Context For example going into `src/home/home.scss` and changing `.Title` color will reproduce the bug (as of commit [451f3a2](https://github.com/wescoder/tavern/commit/451f3a2a5c23188f08419e33c8d2689e0c83e765)) ### 💻 Code Sample You can see the open sourced project here ([https://github.com/wescoder/tavern](https://github.com/wescoder/tavern)) ### 🌍 Your Environment | Software | Version(s) | | ---------------- | ---------- | | Parcel | ^1.4.1 | | Node | v9.2.0 | | npm/Yarn | 1.3.2 | Operating System | Linux Deepin 15.5 (Ubuntu based) |
1.0
Changing `scss`/`css` file breaks style on livereload - **Choose one:** is this a 🐛 bug report or 🙋 feature request? 🐛 bug report ### 🎛 Configuration (.babelrc, package.json, cli command) The project I'm working is available [here](https://github.com/wescoder/tavern), you can checkout and reproduce the issue on it. <!--- If describing a bug, tell us what your babel configuration looks like --> .babelrc: ```js { "presets": ["env", "preact"], "plugins": [ "transform-function-bind", ["transform-react-jsx", { "pragma": "h" }], ["module-resolver", { "root": ["."], "alias": { "@app": "./src" } }], "transform-object-rest-spread" ] } ``` .postcssrc: ``` { "modules": true, "plugins": { "autoprefixer": { "grid": true } } } ``` ### 🤔 Expected Behavior Changing a `scss` file with css modules enabled that is "named-imported" (`import s from './style.scss'`) in js and used in `className` by `preact`/`react`/etc should update the js that requires it providing the new `className` variable to it, without needing a manual full page reload. ### 😯 Current Behavior Changing the `scss` updates the stylesheet causing the network fetch for the new asset but not for the js file, so all styles changed breaks on the page since the old classes doesn't exist anymore, to see the changes a manual full page reload is needed. ### 💁 Possible Solution I really don't have a clue about this... Sorry... ### 🔦 Context For example going into `src/home/home.scss` and changing `.Title` color will reproduce the bug (as of commit [451f3a2](https://github.com/wescoder/tavern/commit/451f3a2a5c23188f08419e33c8d2689e0c83e765)) ### 💻 Code Sample You can see the open sourced project here ([https://github.com/wescoder/tavern](https://github.com/wescoder/tavern)) ### 🌍 Your Environment | Software | Version(s) | | ---------------- | ---------- | | Parcel | ^1.4.1 | | Node | v9.2.0 | | npm/Yarn | 1.3.2 | Operating System | Linux Deepin 15.5 (Ubuntu based) |
process
changing scss css file breaks style on livereload choose one is this a 🐛 bug report or 🙋 feature request 🐛 bug report 🎛 configuration babelrc package json cli command the project i m working is available you can checkout and reproduce the issue on it babelrc js presets plugins transform function bind module resolver root alias app src transform object rest spread postcssrc modules true plugins autoprefixer grid true 🤔 expected behavior changing a scss file with css modules enabled that is named imported import s from style scss in js and used in classname by preact react etc should update the js that requires it providing the new classname variable to it without needing a manual full page reload 😯 current behavior changing the scss updates the stylesheet causing the network fetch for the new asset but not for the js file so all styles changed breaks on the page since the old classes doesn t exist anymore to see the changes a manual full page reload is needed 💁 possible solution i really don t have a clue about this sorry 🔦 context for example going into src home home scss and changing title color will reproduce the bug as of commit 💻 code sample you can see the open sourced project here 🌍 your environment software version s parcel node npm yarn operating system linux deepin ubuntu based
1
4,640
7,482,390,589
IssuesEvent
2018-04-05 01:04:27
UnbFeelings/unb-feelings-GQA
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
opened
Definir critérios de avalição
document help wanted organization process wiki
- [ ] Criar documento "Objetos de Auditoria" - [ ] Listar todos os objetos a serem auditados - [ ] Definir critérios de avaliação para cada objeto auditado
1.0
Definir critérios de avalição - - [ ] Criar documento "Objetos de Auditoria" - [ ] Listar todos os objetos a serem auditados - [ ] Definir critérios de avaliação para cada objeto auditado
process
definir critérios de avalição criar documento objetos de auditoria listar todos os objetos a serem auditados definir critérios de avaliação para cada objeto auditado
1
17,895
23,872,230,172
IssuesEvent
2022-09-07 15:42:10
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[SQL Connector] send avro message to topic and consume using SQL connector with avro format has EOF exception
compute/data-processing
``` Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:150) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: java.io.IOException: java.io.IOException: Failed to deserialize Avro record. at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:141) at org.apache.flink.connector.pulsar.source.reader.split.PulsarOrderedPartitionSplitReader.fetch(PulsarOrderedPartitionSplitReader.java:50) at org.apache.flink.connector.base.source.reader.fetcher.FetchTask.run(FetchTask.java:58) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:142) ... 6 more Caused by: java.io.IOException: Failed to deserialize Avro record. at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:106) at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:46) at org.apache.flink.api.common.serialization.DeserializationSchema.deserialize(DeserializationSchema.java:82) at org.apache.flink.connector.pulsar.table.source.PulsarTableDeserializationSchema.deserialize(PulsarTableDeserializationSchema.java:88) at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:123) ... 9 more Caused by: java.io.EOFException at org.apache.avro.io.BinaryDecoder$InputStreamByteSource.readRaw(BinaryDecoder.java:851) at org.apache.avro.io.BinaryDecoder.doReadBytes(BinaryDecoder.java:373) at org.apache.avro.io.BinaryDecoder.readString(BinaryDecoder.java:290) at org.apache.avro.io.ResolvingDecoder.readString(ResolvingDecoder.java:208) at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:469) at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:459) at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:191) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160) at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:187) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160) at org.apache.avro.generic.GenericDatumReader.readField(GenericDatumReader.java:259) at org.apache.avro.generic.GenericDatumReader.readRecord(GenericDatumReader.java:247) at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:179) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:153) at org.apache.flink.formats.avro.AvroDeserializationSchema.deserialize(AvroDeserializationSchema.java:142) at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:103) ... 13 more ``` It encounted this error in the sendMessageToTopicAndReadUsingAvroBasedSchema() method. We know that such use cases succeeded before. Need to figure out why it's failing now.
1.0
[SQL Connector] send avro message to topic and consume using SQL connector with avro format has EOF exception - ``` Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:150) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: java.io.IOException: java.io.IOException: Failed to deserialize Avro record. at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:141) at org.apache.flink.connector.pulsar.source.reader.split.PulsarOrderedPartitionSplitReader.fetch(PulsarOrderedPartitionSplitReader.java:50) at org.apache.flink.connector.base.source.reader.fetcher.FetchTask.run(FetchTask.java:58) at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:142) ... 6 more Caused by: java.io.IOException: Failed to deserialize Avro record. at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:106) at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:46) at org.apache.flink.api.common.serialization.DeserializationSchema.deserialize(DeserializationSchema.java:82) at org.apache.flink.connector.pulsar.table.source.PulsarTableDeserializationSchema.deserialize(PulsarTableDeserializationSchema.java:88) at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:123) ... 9 more Caused by: java.io.EOFException at org.apache.avro.io.BinaryDecoder$InputStreamByteSource.readRaw(BinaryDecoder.java:851) at org.apache.avro.io.BinaryDecoder.doReadBytes(BinaryDecoder.java:373) at org.apache.avro.io.BinaryDecoder.readString(BinaryDecoder.java:290) at org.apache.avro.io.ResolvingDecoder.readString(ResolvingDecoder.java:208) at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:469) at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:459) at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:191) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160) at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:187) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160) at org.apache.avro.generic.GenericDatumReader.readField(GenericDatumReader.java:259) at org.apache.avro.generic.GenericDatumReader.readRecord(GenericDatumReader.java:247) at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:179) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160) at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:153) at org.apache.flink.formats.avro.AvroDeserializationSchema.deserialize(AvroDeserializationSchema.java:142) at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:103) ... 13 more ``` It encounted this error in the sendMessageToTopicAndReadUsingAvroBasedSchema() method. We know that such use cases succeeded before. Need to figure out why it's failing now.
process
send avro message to topic and consume using sql connector with avro format has eof exception caused by java lang runtimeexception splitfetcher thread received unexpected exception while polling the records at org apache flink connector base source reader fetcher splitfetcher runonce splitfetcher java at org apache flink connector base source reader fetcher splitfetcher run splitfetcher java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java more caused by java io ioexception java io ioexception failed to deserialize avro record at org apache flink connector pulsar source reader split pulsarpartitionsplitreaderbase fetch pulsarpartitionsplitreaderbase java at org apache flink connector pulsar source reader split pulsarorderedpartitionsplitreader fetch pulsarorderedpartitionsplitreader java at org apache flink connector base source reader fetcher fetchtask run fetchtask java at org apache flink connector base source reader fetcher splitfetcher runonce splitfetcher java more caused by java io ioexception failed to deserialize avro record at org apache flink formats avro avrorowdatadeserializationschema deserialize avrorowdatadeserializationschema java at org apache flink formats avro avrorowdatadeserializationschema deserialize avrorowdatadeserializationschema java at org apache flink api common serialization deserializationschema deserialize deserializationschema java at org apache flink connector pulsar table source pulsartabledeserializationschema deserialize pulsartabledeserializationschema java at org apache flink connector pulsar source reader split pulsarpartitionsplitreaderbase fetch pulsarpartitionsplitreaderbase java more caused by java io eofexception at org apache avro io binarydecoder inputstreambytesource readraw binarydecoder java at org apache avro io binarydecoder doreadbytes binarydecoder java at org apache avro io binarydecoder readstring binarydecoder java at org apache avro io resolvingdecoder readstring resolvingdecoder java at org apache avro generic genericdatumreader readstring genericdatumreader java at org apache avro generic genericdatumreader readstring genericdatumreader java at org apache avro generic genericdatumreader readwithoutconversion genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache avro generic genericdatumreader readwithoutconversion genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache avro generic genericdatumreader readfield genericdatumreader java at org apache avro generic genericdatumreader readrecord genericdatumreader java at org apache avro generic genericdatumreader readwithoutconversion genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache flink formats avro avrodeserializationschema deserialize avrodeserializationschema java at org apache flink formats avro avrorowdatadeserializationschema deserialize avrorowdatadeserializationschema java more it encounted this error in the sendmessagetotopicandreadusingavrobasedschema method we know that such use cases succeeded before need to figure out why it s failing now
1
17,770
23,698,732,518
IssuesEvent
2022-08-29 16:53:34
cloudfoundry/korifi
https://api.github.com/repos/cloudfoundry/korifi
opened
[Feature]: Developer can push apps using the top-level `disk_quota` field in the manifest
Top-level process config
### Blockers/Dependencies _No response_ ### Background **As a** developer **I want** top-level process configuration in manifests to be supported **So that** I can use shortcut `cf push` flags like `-c`, `-i`, `-m` etc. ### Acceptance Criteria **GIVEN** I have the sources of an application (e.g. `tests/smoke/assets/test-node-app`) **WHEN I** push it with the following command: ```sh cf push test -k 1GB ``` **THEN I** see the push succeeds with an output similar to this: ``` name: test requested state: started routes: test.vcap.me last uploaded: Mon 29 Aug 16:28:36 UTC 2022 stack: cflinuxfs3 buildpacks: name version detect output buildpack name nodejs_buildpack 1.7.61 nodejs nodejs type: web sidecars: instances: 1/1 memory usage: 256M start command: npm start state since cpu memory disk details #0 running 2022-08-29T16:28:54Z 1.6% 42.3M of 1G 115.7M of 1G ``` ### Dev Notes The CLI will use the top-level `disk_quota` field on app objects for this, but [only if](https://github.com/cloudfoundry/cli/blob/68d897a19792bdffe20c39589148216b23c950d1/actor/v7pushaction/handle_disk_override.go#L14-L20) there's no `manifest.yml`, or if `manifest.yml` doesn't already have a `web` process definition.
1.0
[Feature]: Developer can push apps using the top-level `disk_quota` field in the manifest - ### Blockers/Dependencies _No response_ ### Background **As a** developer **I want** top-level process configuration in manifests to be supported **So that** I can use shortcut `cf push` flags like `-c`, `-i`, `-m` etc. ### Acceptance Criteria **GIVEN** I have the sources of an application (e.g. `tests/smoke/assets/test-node-app`) **WHEN I** push it with the following command: ```sh cf push test -k 1GB ``` **THEN I** see the push succeeds with an output similar to this: ``` name: test requested state: started routes: test.vcap.me last uploaded: Mon 29 Aug 16:28:36 UTC 2022 stack: cflinuxfs3 buildpacks: name version detect output buildpack name nodejs_buildpack 1.7.61 nodejs nodejs type: web sidecars: instances: 1/1 memory usage: 256M start command: npm start state since cpu memory disk details #0 running 2022-08-29T16:28:54Z 1.6% 42.3M of 1G 115.7M of 1G ``` ### Dev Notes The CLI will use the top-level `disk_quota` field on app objects for this, but [only if](https://github.com/cloudfoundry/cli/blob/68d897a19792bdffe20c39589148216b23c950d1/actor/v7pushaction/handle_disk_override.go#L14-L20) there's no `manifest.yml`, or if `manifest.yml` doesn't already have a `web` process definition.
process
developer can push apps using the top level disk quota field in the manifest blockers dependencies no response background as a developer i want top level process configuration in manifests to be supported so that i can use shortcut cf push flags like c i m etc acceptance criteria given i have the sources of an application e g tests smoke assets test node app when i push it with the following command sh cf push test k then i see the push succeeds with an output similar to this name test requested state started routes test vcap me last uploaded mon aug utc stack buildpacks name version detect output buildpack name nodejs buildpack nodejs nodejs type web sidecars instances memory usage start command npm start state since cpu memory disk details running of of dev notes the cli will use the top level disk quota field on app objects for this but there s no manifest yml or if manifest yml doesn t already have a web process definition
1
5,196
7,974,012,858
IssuesEvent
2018-07-17 02:43:30
pelias/pelias
https://api.github.com/repos/pelias/pelias
closed
Harden dependency on libpostal
processed
Currently, if pelias-config specifies `parser: addressit` instead of the expected `parser: libpostal`, exceptions are thrown upon incoming requests. Since we have no intention of supporting Pelias without `libpostal` we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly. Startup of API should also fail if libpostal is not properly installed or could not be loaded for some reason.
1.0
Harden dependency on libpostal - Currently, if pelias-config specifies `parser: addressit` instead of the expected `parser: libpostal`, exceptions are thrown upon incoming requests. Since we have no intention of supporting Pelias without `libpostal` we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly. Startup of API should also fail if libpostal is not properly installed or could not be loaded for some reason.
process
harden dependency on libpostal currently if pelias config specifies parser addressit instead of the expected parser libpostal exceptions are thrown upon incoming requests since we have no intention of supporting pelias without libpostal we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly startup of api should also fail if libpostal is not properly installed or could not be loaded for some reason
1
249,104
7,953,807,638
IssuesEvent
2018-07-12 03:58:35
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: second layer items not visible
Medium Priority
**Version:** 0.7.2.3 beta **Steps to Reproduce:** drop 1st layer of items or wall down drop 2nd layer down 2nd layer isnt visable disconnect reconnect 2nd layer visable **Expected behavior:** 2nd layer of items be visable after placement items taken from carry inventory **Actual behavior:** 2nd layer of items invisable items taken from carry inventory
1.0
USER ISSUE: second layer items not visible - **Version:** 0.7.2.3 beta **Steps to Reproduce:** drop 1st layer of items or wall down drop 2nd layer down 2nd layer isnt visable disconnect reconnect 2nd layer visable **Expected behavior:** 2nd layer of items be visable after placement items taken from carry inventory **Actual behavior:** 2nd layer of items invisable items taken from carry inventory
non_process
user issue second layer items not visible version beta steps to reproduce drop layer of items or wall down drop layer down layer isnt visable disconnect reconnect layer visable expected behavior layer of items be visable after placement items taken from carry inventory actual behavior layer of items invisable items taken from carry inventory
0
177,192
13,686,148,637
IssuesEvent
2020-09-30 08:16:39
photoprism/photoprism
https://api.github.com/repos/photoprism/photoprism
closed
Fixed, Success Message in German
bug please-test
In the German language, when changing an album, after pressing the Save button, the message comes in English (Album saved). <img width="774" alt="Bildschirmfoto 2020-09-01 um 21 36 48" src="https://user-images.githubusercontent.com/49251031/91898594-41ba4080-ec9c-11ea-923a-1370c6955855.png"> <img width="525" alt="Bildschirmfoto 2020-09-01 um 21 41 37" src="https://user-images.githubusercontent.com/49251031/91898610-454dc780-ec9c-11ea-9c27-696d18ebdbbf.png">
1.0
Fixed, Success Message in German - In the German language, when changing an album, after pressing the Save button, the message comes in English (Album saved). <img width="774" alt="Bildschirmfoto 2020-09-01 um 21 36 48" src="https://user-images.githubusercontent.com/49251031/91898594-41ba4080-ec9c-11ea-923a-1370c6955855.png"> <img width="525" alt="Bildschirmfoto 2020-09-01 um 21 41 37" src="https://user-images.githubusercontent.com/49251031/91898610-454dc780-ec9c-11ea-9c27-696d18ebdbbf.png">
non_process
fixed success message in german in the german language when changing an album after pressing the save button the message comes in english album saved img width alt bildschirmfoto um src img width alt bildschirmfoto um src
0
318,225
23,708,577,631
IssuesEvent
2022-08-30 05:21:45
starlite-api/starlite
https://api.github.com/repos/starlite-api/starlite
closed
Bug: Documentation Site doesn't work properly in mobile.
bug documentation help wanted good first issue
The Doc site isn't working well in mobile. You need to press the back arrow to find the menu to see usage and reference.
1.0
Bug: Documentation Site doesn't work properly in mobile. - The Doc site isn't working well in mobile. You need to press the back arrow to find the menu to see usage and reference.
non_process
bug documentation site doesn t work properly in mobile the doc site isn t working well in mobile you need to press the back arrow to find the menu to see usage and reference
0
88,545
8,154,687,915
IssuesEvent
2018-08-23 04:48:45
trashgenerator/dcore
https://api.github.com/repos/trashgenerator/dcore
opened
DC-8-Config сделать Unit test
testing
Unit test на сохранение / загрузку /валидацию значений конфиг-файла
1.0
DC-8-Config сделать Unit test - Unit test на сохранение / загрузку /валидацию значений конфиг-файла
non_process
dc config сделать unit test unit test на сохранение загрузку валидацию значений конфиг файла
0
12,545
14,975,688,580
IssuesEvent
2021-01-28 06:40:54
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Update license and copyright to support rosetta
P3 enhancement process rosetta
**Problem** Licenses and Copyright currently cover all the modules except rosetta. **Solution** Update license and copyright to support go rosetta files **Alternatives** A clear and concise description of any alternative solutions or features you've considered. **Additional Context** Add any other context or screenshots about the feature request here.
1.0
Update license and copyright to support rosetta - **Problem** Licenses and Copyright currently cover all the modules except rosetta. **Solution** Update license and copyright to support go rosetta files **Alternatives** A clear and concise description of any alternative solutions or features you've considered. **Additional Context** Add any other context or screenshots about the feature request here.
process
update license and copyright to support rosetta problem licenses and copyright currently cover all the modules except rosetta solution update license and copyright to support go rosetta files alternatives a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
3,536
6,573,789,459
IssuesEvent
2017-09-11 10:07:26
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Should overwriting child_process.execFile result in changed behaviour of child_process.exec?
child_process
```sh node -e "child_process.execFile = null; child_process.exec('cat package.json')" child_process.js:138 return exports.execFile(opts.file, TypeError: exports.execFile is not a function at Object.exports.exec (child_process.js:138:18) ``` I couldn't find in documentation why child_process behave this way (but it behaves like that for at least [four years](https://github.com/nodejs/node/commit/e8df2676748e944388896dfd767e01906ae2e4eb)). Relevant line https://github.com/nodejs/node/blob/58831b2/lib/child_process.js#L136
1.0
Should overwriting child_process.execFile result in changed behaviour of child_process.exec? - ```sh node -e "child_process.execFile = null; child_process.exec('cat package.json')" child_process.js:138 return exports.execFile(opts.file, TypeError: exports.execFile is not a function at Object.exports.exec (child_process.js:138:18) ``` I couldn't find in documentation why child_process behave this way (but it behaves like that for at least [four years](https://github.com/nodejs/node/commit/e8df2676748e944388896dfd767e01906ae2e4eb)). Relevant line https://github.com/nodejs/node/blob/58831b2/lib/child_process.js#L136
process
should overwriting child process execfile result in changed behaviour of child process exec sh node e child process execfile null child process exec cat package json child process js return exports execfile opts file typeerror exports execfile is not a function at object exports exec child process js i couldn t find in documentation why child process behave this way but it behaves like that for at least relevant line
1
205,805
7,106,120,709
IssuesEvent
2018-01-16 15:43:19
flowup/ngx-swagger-client-generator
https://api.github.com/repos/flowup/ngx-swagger-client-generator
closed
Type Errors
bug priority:HIGH
Hi, I'm having some issues with my generated code. On build I get a bunch of errors like: ``` .../api-client-service.ts(379,59): error TS2345: Argument of type 'HttpOptions' is not assignable to parameter of type '{ headers?: HttpHeaders | { [header: string]: string | string[]; }; observe?: "body"; params?: Ht...'. Types of property 'responseType' are incompatible. Type '"text" | "arraybuffer" | "blob" | "json"' is not assignable to type '"json"'. Type '"text"' is not assignable to type '"json"'. ``` This seem to me like it's probably a versioning issue, e.g. the typescript or angular versions are wrong. I'm using Angular 4.x.x with typescript 2.4.0. I tried angular 5.x.x and TS 2.6.x and has the same issues. Maybe the code you generate targets an older angular version? Additionally `HttpObserve` is not exported in `/common/http`. To get the generated code to build at all I've edited the HttpOptions interface as follows: ``` interface HttpOptions { headers?: HttpHeaders, params?: HttpParams, reportProgress?: boolean, withCredentials?: boolean, } ``` Any ideas?
1.0
Type Errors - Hi, I'm having some issues with my generated code. On build I get a bunch of errors like: ``` .../api-client-service.ts(379,59): error TS2345: Argument of type 'HttpOptions' is not assignable to parameter of type '{ headers?: HttpHeaders | { [header: string]: string | string[]; }; observe?: "body"; params?: Ht...'. Types of property 'responseType' are incompatible. Type '"text" | "arraybuffer" | "blob" | "json"' is not assignable to type '"json"'. Type '"text"' is not assignable to type '"json"'. ``` This seem to me like it's probably a versioning issue, e.g. the typescript or angular versions are wrong. I'm using Angular 4.x.x with typescript 2.4.0. I tried angular 5.x.x and TS 2.6.x and has the same issues. Maybe the code you generate targets an older angular version? Additionally `HttpObserve` is not exported in `/common/http`. To get the generated code to build at all I've edited the HttpOptions interface as follows: ``` interface HttpOptions { headers?: HttpHeaders, params?: HttpParams, reportProgress?: boolean, withCredentials?: boolean, } ``` Any ideas?
non_process
type errors hi i m having some issues with my generated code on build i get a bunch of errors like api client service ts error argument of type httpoptions is not assignable to parameter of type headers httpheaders string string observe body params ht types of property responsetype are incompatible type text arraybuffer blob json is not assignable to type json type text is not assignable to type json this seem to me like it s probably a versioning issue e g the typescript or angular versions are wrong i m using angular x x with typescript i tried angular x x and ts x and has the same issues maybe the code you generate targets an older angular version additionally httpobserve is not exported in common http to get the generated code to build at all i ve edited the httpoptions interface as follows interface httpoptions headers httpheaders params httpparams reportprogress boolean withcredentials boolean any ideas
0
320,019
23,799,458,546
IssuesEvent
2022-09-03 03:42:22
ued123/modeling
https://api.github.com/repos/ued123/modeling
opened
음료수 구매하기 도메인 개선
documentation
## 요약 ### project: cash-api - 자판기 입금/출금 액을 관리할수 있는 도메인: cash - 현금 거래 내역: transaction ### project: purchase-api - cash, drink 로직을 참조해서 사용할수 있습니다 - 주문: purchase
1.0
음료수 구매하기 도메인 개선 - ## 요약 ### project: cash-api - 자판기 입금/출금 액을 관리할수 있는 도메인: cash - 현금 거래 내역: transaction ### project: purchase-api - cash, drink 로직을 참조해서 사용할수 있습니다 - 주문: purchase
non_process
음료수 구매하기 도메인 개선 요약 project cash api 자판기 입금 출금 액을 관리할수 있는 도메인 cash 현금 거래 내역 transaction project purchase api cash drink 로직을 참조해서 사용할수 있습니다 주문 purchase
0
21,724
30,232,574,680
IssuesEvent
2023-07-06 08:02:25
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
closed
Perform every concrete execution in a separate transaction that is rollbacked right after execution
ctg-enhancement comp-instrumented-process comp-spring
**Description** Right now database content is reset between concrete executions with hardcoded SQL statements tuned for `spring-boot-testing-main` project database scheme (e.g. `ALTER TABLE ${repositoryDefinition.tableName} ALTER COLUMN id RESTART WITH 1`). **Expected behavior** Every concrete execution run is performed in a separate transaction that is rollbacked right after execution.
1.0
Perform every concrete execution in a separate transaction that is rollbacked right after execution - **Description** Right now database content is reset between concrete executions with hardcoded SQL statements tuned for `spring-boot-testing-main` project database scheme (e.g. `ALTER TABLE ${repositoryDefinition.tableName} ALTER COLUMN id RESTART WITH 1`). **Expected behavior** Every concrete execution run is performed in a separate transaction that is rollbacked right after execution.
process
perform every concrete execution in a separate transaction that is rollbacked right after execution description right now database content is reset between concrete executions with hardcoded sql statements tuned for spring boot testing main project database scheme e g alter table repositorydefinition tablename alter column id restart with expected behavior every concrete execution run is performed in a separate transaction that is rollbacked right after execution
1
238,693
7,782,180,625
IssuesEvent
2018-06-06 05:03:17
tyarkoni/pliers
https://api.github.com/repos/tyarkoni/pliers
closed
Offset of elements gets added again if ComplexTextStim iterated twice
bug priority: high
Problematic line: `elem.onset = offset if elem.onset is None else offset + elem.onset`
1.0
Offset of elements gets added again if ComplexTextStim iterated twice - Problematic line: `elem.onset = offset if elem.onset is None else offset + elem.onset`
non_process
offset of elements gets added again if complextextstim iterated twice problematic line elem onset offset if elem onset is none else offset elem onset
0
492,874
14,221,967,411
IssuesEvent
2020-11-17 16:19:22
StatCan/daaas
https://api.github.com/repos/StatCan/daaas
closed
I require write access to a private repo
area/security component/grafana priority/blocker
Hello, I need to push a feature branch to the terraform-kubernetes-aks-platform-daaas-private repo. It's specifically for integrating Grafana with Azure AD, using the prometheus operator. Brendan mentioned that either one of you could grant me write access. Thanks in advance! Chuck
1.0
I require write access to a private repo - Hello, I need to push a feature branch to the terraform-kubernetes-aks-platform-daaas-private repo. It's specifically for integrating Grafana with Azure AD, using the prometheus operator. Brendan mentioned that either one of you could grant me write access. Thanks in advance! Chuck
non_process
i require write access to a private repo hello i need to push a feature branch to the terraform kubernetes aks platform daaas private repo it s specifically for integrating grafana with azure ad using the prometheus operator brendan mentioned that either one of you could grant me write access thanks in advance chuck
0
22,734
10,766,017,909
IssuesEvent
2019-11-01 12:43:21
valdisiljuconoks/azfun-demo
https://api.github.com/repos/valdisiljuconoks/azfun-demo
opened
CVE-2019-13173 (High) detected in fstream-0.1.31.tgz
security vulnerability
## CVE-2019-13173 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-0.1.31.tgz</b></p></summary> <p>Advanced file system stream things</p> <p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz">https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz</a></p> <p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/fstream/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - digdug-1.4.0.tgz - decompress-0.2.3.tgz - tar-0.1.20.tgz - :x: **fstream-0.1.31.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/valdisiljuconoks/azfun-demo/commit/d6e21882f73fdf25bf4853c5914f2a50f94b2086">d6e21882f73fdf25bf4853c5914f2a50f94b2086</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable. <p>Publish Date: 2019-07-02 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p> <p>Release Date: 2019-07-02</p> <p>Fix Resolution: 1.0.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-13173 (High) detected in fstream-0.1.31.tgz - ## CVE-2019-13173 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-0.1.31.tgz</b></p></summary> <p>Advanced file system stream things</p> <p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz">https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz</a></p> <p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/fstream/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - digdug-1.4.0.tgz - decompress-0.2.3.tgz - tar-0.1.20.tgz - :x: **fstream-0.1.31.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/valdisiljuconoks/azfun-demo/commit/d6e21882f73fdf25bf4853c5914f2a50f94b2086">d6e21882f73fdf25bf4853c5914f2a50f94b2086</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable. <p>Publish Date: 2019-07-02 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p> <p>Release Date: 2019-07-02</p> <p>Fix Resolution: 1.0.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in fstream tgz cve high severity vulnerability vulnerable library fstream tgz advanced file system stream things library home page a href path to dependency file tmp whitesource archiveextractor ws scm depth azfun demo modules protected shell shell clientresources lib xstyle package json path to vulnerable library tmp whitesource archiveextractor ws scm depth azfun demo modules protected shell shell clientresources lib xstyle node modules fstream package json dependency hierarchy intern geezer tgz root library digdug tgz decompress tgz tar tgz x fstream tgz vulnerable library found in head commit a href vulnerability details fstream before is vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file the fstream dirwriter function is vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
16,201
20,713,571,526
IssuesEvent
2022-03-12 09:11:23
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
PhpExecutableFinder may return directory instead of binary
Bug Process Status: Needs Review
### Symfony version(s) affected 3.4 | 4.4 ### Description We're using \Symfony\Component\Process\PhpExecutableFinder in Drupal's Upgrade Status module to execute PHPStan from a UI. In different environments leveraging Plesk or ISPConfig, the returned path from \Symfony\Component\Process\PhpExecutableFinder::find is a directory. ### How to reproduce I haven't been able to reproduce since I do not have access to a Plesk server. But directories are technically executable. ### Possible Solution The checks for `is_exectuable` may need a `!is_dir` check? ### Additional Context _No response_
1.0
PhpExecutableFinder may return directory instead of binary - ### Symfony version(s) affected 3.4 | 4.4 ### Description We're using \Symfony\Component\Process\PhpExecutableFinder in Drupal's Upgrade Status module to execute PHPStan from a UI. In different environments leveraging Plesk or ISPConfig, the returned path from \Symfony\Component\Process\PhpExecutableFinder::find is a directory. ### How to reproduce I haven't been able to reproduce since I do not have access to a Plesk server. But directories are technically executable. ### Possible Solution The checks for `is_exectuable` may need a `!is_dir` check? ### Additional Context _No response_
process
phpexecutablefinder may return directory instead of binary symfony version s affected description we re using symfony component process phpexecutablefinder in drupal s upgrade status module to execute phpstan from a ui in different environments leveraging plesk or ispconfig the returned path from symfony component process phpexecutablefinder find is a directory how to reproduce i haven t been able to reproduce since i do not have access to a plesk server but directories are technically executable possible solution the checks for is exectuable may need a is dir check additional context no response
1
20,627
27,299,815,214
IssuesEvent
2023-02-24 00:19:25
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
Turn off trusted-contributor bot for release-please
type: process
We expect release-please to keep an open PR with pending changes to be released. We do not need to run samples/system tests until we are ready to publish the packages. Otherwise, we will be running so many unnecessary integration tests when we are not ready to release. When a maintainer is ready to release: 1. approve the PR 2. comment with `/gcbrun` 3. add the `automerge` label
1.0
Turn off trusted-contributor bot for release-please - We expect release-please to keep an open PR with pending changes to be released. We do not need to run samples/system tests until we are ready to publish the packages. Otherwise, we will be running so many unnecessary integration tests when we are not ready to release. When a maintainer is ready to release: 1. approve the PR 2. comment with `/gcbrun` 3. add the `automerge` label
process
turn off trusted contributor bot for release please we expect release please to keep an open pr with pending changes to be released we do not need to run samples system tests until we are ready to publish the packages otherwise we will be running so many unnecessary integration tests when we are not ready to release when a maintainer is ready to release approve the pr comment with gcbrun add the automerge label
1
15,610
19,749,433,830
IssuesEvent
2022-01-15 00:08:31
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
MDBGA: Prevent users from trying to use referential actions on inline m-n's
process/candidate topic: schema validation team/migrations topic: mongodb topic: referential actions kind/subtask
_(This issue is a follow-up to: https://github.com/prisma/prisma/issues/10498)_ Currently, the Prisma Schema allows: ```prisma model A { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId gql String? b_ids String[] @db.Array(ObjectId) bs B[] @relation(fields: [b_ids], onDelete: Cascade) } model B { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId gql String? a_ids String[] @db.Array(ObjectId) as A[] @relation(fields: [a_ids], onDelete: Cascade) } ``` But we don't support referential actions on m-ns for MongoDB. We should: - Invalidate this in prisma format and provide a helpful error. Perhaps link to the documentation to learn how to workaround it. - Document this limitation in [our documentation](https://www.prisma.io/docs/concepts/components/prisma-schema/relations/many-to-many-relations#mongodb). We can tell people to use an explicit join table as a workaround. Later on, we can consider adding referential action support to this type of schema.
1.0
MDBGA: Prevent users from trying to use referential actions on inline m-n's - _(This issue is a follow-up to: https://github.com/prisma/prisma/issues/10498)_ Currently, the Prisma Schema allows: ```prisma model A { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId gql String? b_ids String[] @db.Array(ObjectId) bs B[] @relation(fields: [b_ids], onDelete: Cascade) } model B { id String @id @default(dbgenerated()) @map("_id") @db.ObjectId gql String? a_ids String[] @db.Array(ObjectId) as A[] @relation(fields: [a_ids], onDelete: Cascade) } ``` But we don't support referential actions on m-ns for MongoDB. We should: - Invalidate this in prisma format and provide a helpful error. Perhaps link to the documentation to learn how to workaround it. - Document this limitation in [our documentation](https://www.prisma.io/docs/concepts/components/prisma-schema/relations/many-to-many-relations#mongodb). We can tell people to use an explicit join table as a workaround. Later on, we can consider adding referential action support to this type of schema.
process
mdbga prevent users from trying to use referential actions on inline m n s this issue is a follow up to currently the prisma schema allows prisma model a id string id default dbgenerated map id db objectid gql string b ids string db array objectid bs b relation fields ondelete cascade model b id string id default dbgenerated map id db objectid gql string a ids string db array objectid as a relation fields ondelete cascade but we don t support referential actions on m ns for mongodb we should invalidate this in prisma format and provide a helpful error perhaps link to the documentation to learn how to workaround it document this limitation in we can tell people to use an explicit join table as a workaround later on we can consider adding referential action support to this type of schema
1
672
3,145,591,116
IssuesEvent
2015-09-14 18:44:53
K0zka/kerub
https://api.github.com/repos/K0zka/kerub
opened
Add optional description fields to entities
component:data processing component:ui priority: minimal
Markdown formatted description for the following entities: * Network * VM * Virtual Disk * Expectations (followup of #81 )
1.0
Add optional description fields to entities - Markdown formatted description for the following entities: * Network * VM * Virtual Disk * Expectations (followup of #81 )
process
add optional description fields to entities markdown formatted description for the following entities network vm virtual disk expectations followup of
1
3,502
6,557,593,319
IssuesEvent
2017-09-06 18:01:36
w3c/webauthn
https://api.github.com/repos/w3c/webauthn
closed
The W3C HTML spec is broken, and probably shouldn't be referenced
type:process type:technical
I was pretty confused by https://www.w3.org/TR/html51/browsers.html#relaxing-the-same-origin-restriction step 7.2, as referenced from https://w3c.github.io/webauthn/#dom-webauthentication-makecredential step 3, because it's completely nonsensical: it causes the algorithm to _always_ throw. Then I realized that this is the W3C version of the HTML spec, which has all sorts of known issues like this. For comparison, https://html.spec.whatwg.org/multipage/browsers.html#relaxing-the-same-origin-restriction step 7.2 is correct... Please reference the things that actually have the behavior you want. I guess if you have to reference W3C HTML, that makes it your problem to check that it has the behavior you want and push on its editors to fix it if it does not.
1.0
The W3C HTML spec is broken, and probably shouldn't be referenced - I was pretty confused by https://www.w3.org/TR/html51/browsers.html#relaxing-the-same-origin-restriction step 7.2, as referenced from https://w3c.github.io/webauthn/#dom-webauthentication-makecredential step 3, because it's completely nonsensical: it causes the algorithm to _always_ throw. Then I realized that this is the W3C version of the HTML spec, which has all sorts of known issues like this. For comparison, https://html.spec.whatwg.org/multipage/browsers.html#relaxing-the-same-origin-restriction step 7.2 is correct... Please reference the things that actually have the behavior you want. I guess if you have to reference W3C HTML, that makes it your problem to check that it has the behavior you want and push on its editors to fix it if it does not.
process
the html spec is broken and probably shouldn t be referenced i was pretty confused by step as referenced from step because it s completely nonsensical it causes the algorithm to always throw then i realized that this is the version of the html spec which has all sorts of known issues like this for comparison step is correct please reference the things that actually have the behavior you want i guess if you have to reference html that makes it your problem to check that it has the behavior you want and push on its editors to fix it if it does not
1
11,199
13,957,702,812
IssuesEvent
2020-10-24 08:13:43
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
CZ: Harvesting to thematic viewer
CZ - Czech Republic Geoportal Harvesting process
From: Jitka.Faugnerova@cenia.cz [Jitka.Faugnerova@cenia.cz] Sent: 14 May 2018 14:03 To: TOMAS Robert (JRC-ISPRA) Subject: Harvesting to thematic viewer Dear Robert, could you please check with Angelo or ask him to do the harvesting to your geoportal/viewer? One of our data provider has made some changes in metadata regarding the priority data sets and it seems that last time the metadata was harvested is 09/05/2018. Deadline given by the letter from December 2018 is tomorrow, so we would like to see, whether we have succeeded in making data resources available to the Thematic Viewer. Of course we would like to score better than now :-) Thank you very much in advance. Best regards, Jitka Jitka Faugnerov&aacute; CENIA, ?esk&aacute; informa?n&iacute; agentura ?ivotn&iacute;ho prost?ed&iacute;
1.0
CZ: Harvesting to thematic viewer - From: Jitka.Faugnerova@cenia.cz [Jitka.Faugnerova@cenia.cz] Sent: 14 May 2018 14:03 To: TOMAS Robert (JRC-ISPRA) Subject: Harvesting to thematic viewer Dear Robert, could you please check with Angelo or ask him to do the harvesting to your geoportal/viewer? One of our data provider has made some changes in metadata regarding the priority data sets and it seems that last time the metadata was harvested is 09/05/2018. Deadline given by the letter from December 2018 is tomorrow, so we would like to see, whether we have succeeded in making data resources available to the Thematic Viewer. Of course we would like to score better than now :-) Thank you very much in advance. Best regards, Jitka Jitka Faugnerov&aacute; CENIA, ?esk&aacute; informa?n&iacute; agentura ?ivotn&iacute;ho prost?ed&iacute;
process
cz harvesting to thematic viewer from jitka faugnerova cenia cz sent may to tomas robert jrc ispra subject harvesting to thematic viewer dear robert could you please check with angelo or ask him to do the harvesting to your geoportal viewer one of our data provider has made some changes in metadata regarding the priority data sets and it seems that last time the metadata was harvested is deadline given by the letter from december is tomorrow so we would like to see whether we have succeeded in making data resources available to the thematic viewer of course we would like to score better than now thank you very much in advance best regards jitka jitka faugnerov aacute cenia esk aacute informa n iacute agentura ivotn iacute ho prost ed iacute
1
582,078
17,352,338,381
IssuesEvent
2021-07-29 10:16:49
hedgedoc/hedgedoc
https://api.github.com/repos/hedgedoc/hedgedoc
closed
no link between slides
priority: high scope: frontend status: confirmed type: bug
**Description** When presenting slides, the feature to create links between slides does not work. **To Reproduce** 1. Go to demo slides (of the demo instance) 2. Go to [page 18](https://demo.hedgedoc.org/p/slide-example#/17) 3. Click on link "like this" 4. See that the browser is redirected to home page **Expected behavior** The link should lead to slide `1/3`. **Logs** n/a **Config** n/a **Setup:** - demo instance **Additional context** n/a
1.0
no link between slides - **Description** When presenting slides, the feature to create links between slides does not work. **To Reproduce** 1. Go to demo slides (of the demo instance) 2. Go to [page 18](https://demo.hedgedoc.org/p/slide-example#/17) 3. Click on link "like this" 4. See that the browser is redirected to home page **Expected behavior** The link should lead to slide `1/3`. **Logs** n/a **Config** n/a **Setup:** - demo instance **Additional context** n/a
non_process
no link between slides description when presenting slides the feature to create links between slides does not work to reproduce go to demo slides of the demo instance go to click on link like this see that the browser is redirected to home page expected behavior the link should lead to slide logs n a config n a setup demo instance additional context n a
0
10,139
13,044,162,440
IssuesEvent
2020-07-29 03:47:32
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `JsonArrayAppendSig` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `JsonArrayAppendSig` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `JsonArrayAppendSig` from TiDB - ## Description Port the scalar function `JsonArrayAppendSig` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function jsonarrayappendsig from tidb description port the scalar function jsonarrayappendsig from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
1
14,645
25,331,436,165
IssuesEvent
2022-11-18 13:34:36
AAFC-BICoE/dina-planning
https://api.github.com/repos/AAFC-BICoE/dina-planning
closed
Allow to add material-sample to transaction
type: requirement
Goal: allow to search for material-sample and add them to a transaction. Out-of-scope for this ticket: show that a material-sample is not available since it's part of a transaction.
1.0
Allow to add material-sample to transaction - Goal: allow to search for material-sample and add them to a transaction. Out-of-scope for this ticket: show that a material-sample is not available since it's part of a transaction.
non_process
allow to add material sample to transaction goal allow to search for material sample and add them to a transaction out of scope for this ticket show that a material sample is not available since it s part of a transaction
0
31,668
15,036,210,141
IssuesEvent
2021-02-02 15:00:10
Automattic/wp-calypso
https://api.github.com/repos/Automattic/wp-calypso
closed
PlansGrid: API gets called twice
Performance [Goal] New Onboarding [Pri] High [Type] Bug
Probably after merging https://github.com/Automattic/wp-calypso/pull/48790, when loading PlansGrid in both Gutenboarding and launch flows (when sandboxed), the REST API calls for `/plans` and `/details` are each done 2 times. <img width="731" alt="Screenshot 2021-01-21 at 20 52 03" src="https://user-images.githubusercontent.com/14192054/105399264-0d83f400-5c2c-11eb-8877-bdb0277b1307.png">
True
PlansGrid: API gets called twice - Probably after merging https://github.com/Automattic/wp-calypso/pull/48790, when loading PlansGrid in both Gutenboarding and launch flows (when sandboxed), the REST API calls for `/plans` and `/details` are each done 2 times. <img width="731" alt="Screenshot 2021-01-21 at 20 52 03" src="https://user-images.githubusercontent.com/14192054/105399264-0d83f400-5c2c-11eb-8877-bdb0277b1307.png">
non_process
plansgrid api gets called twice probably after merging when loading plansgrid in both gutenboarding and launch flows when sandboxed the rest api calls for plans and details are each done times img width alt screenshot at src
0
94,120
15,962,337,510
IssuesEvent
2021-04-16 01:05:14
RG4421/nucleus
https://api.github.com/repos/RG4421/nucleus
opened
WS-2017-0266 (Low) detected in http-signature-0.10.1.tgz
security vulnerability
## WS-2017-0266 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-signature-0.10.1.tgz</b></p></summary> <p>Reference implementation of Joyent's HTTP Signature scheme.</p> <p>Library home page: <a href="https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz">https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz</a></p> <p>Path to dependency file: nucleus/packages/@nucleus/package.json</p> <p>Path to vulnerable library: nucleus/packages/@nucleus/node_modules/yui/node_modules/http-signature/package.json</p> <p> Dependency Hierarchy: - ember-cli-addon-docs-0.6.16.tgz (Root Library) - yuidocjs-0.10.2.tgz - yui-3.18.1.tgz - request-2.40.0.tgz - :x: **http-signature-0.10.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks. <p>Publish Date: 2015-01-22 <p>URL: <a href=https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56>WS-2017-0266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/joyent/node-http-signature/pull/36">https://github.com/joyent/node-http-signature/pull/36</a></p> <p>Release Date: 2017-01-31</p> <p>Fix Resolution: 1.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"http-signature","packageVersion":"0.10.1","packageFilePaths":["/packages/@nucleus/package.json"],"isTransitiveDependency":true,"dependencyTree":"ember-cli-addon-docs:0.6.16;yuidocjs:0.10.2;yui:3.18.1;request:2.40.0;http-signature:0.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0266","vulnerabilityDetails":"Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks.","vulnerabilityUrl":"https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56","cvss3Severity":"low","cvss3Score":"3.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Adjacent","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2017-0266 (Low) detected in http-signature-0.10.1.tgz - ## WS-2017-0266 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-signature-0.10.1.tgz</b></p></summary> <p>Reference implementation of Joyent's HTTP Signature scheme.</p> <p>Library home page: <a href="https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz">https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz</a></p> <p>Path to dependency file: nucleus/packages/@nucleus/package.json</p> <p>Path to vulnerable library: nucleus/packages/@nucleus/node_modules/yui/node_modules/http-signature/package.json</p> <p> Dependency Hierarchy: - ember-cli-addon-docs-0.6.16.tgz (Root Library) - yuidocjs-0.10.2.tgz - yui-3.18.1.tgz - request-2.40.0.tgz - :x: **http-signature-0.10.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks. <p>Publish Date: 2015-01-22 <p>URL: <a href=https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56>WS-2017-0266</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/joyent/node-http-signature/pull/36">https://github.com/joyent/node-http-signature/pull/36</a></p> <p>Release Date: 2017-01-31</p> <p>Fix Resolution: 1.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"http-signature","packageVersion":"0.10.1","packageFilePaths":["/packages/@nucleus/package.json"],"isTransitiveDependency":true,"dependencyTree":"ember-cli-addon-docs:0.6.16;yuidocjs:0.10.2;yui:3.18.1;request:2.40.0;http-signature:0.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0266","vulnerabilityDetails":"Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks.","vulnerabilityUrl":"https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56","cvss3Severity":"low","cvss3Score":"3.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Adjacent","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
ws low detected in http signature tgz ws low severity vulnerability vulnerable library http signature tgz reference implementation of joyent s http signature scheme library home page a href path to dependency file nucleus packages nucleus package json path to vulnerable library nucleus packages nucleus node modules yui node modules http signature package json dependency hierarchy ember cli addon docs tgz root library yuidocjs tgz yui tgz request tgz x http signature tgz vulnerable library found in base branch master vulnerability details affected versions before of the http signature package are vulnerable to timing attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree ember cli addon docs yuidocjs yui request http signature isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails affected versions before of the http signature package are vulnerable to timing attacks vulnerabilityurl
0
193,982
15,394,038,524
IssuesEvent
2021-03-03 17:24:52
PawelDabrowski83/genAlgo
https://api.github.com/repos/PawelDabrowski83/genAlgo
closed
Main description
documentation
Main: <problem - w ilu pokoleniach uda się uzyskać gen o wartości 1234> GenePoolService(wszystkie składowe) GenePool(wartość szukana 1234, liczba genów, genePoolService).solve() -> int z rozwiązaniem
1.0
Main description - Main: <problem - w ilu pokoleniach uda się uzyskać gen o wartości 1234> GenePoolService(wszystkie składowe) GenePool(wartość szukana 1234, liczba genów, genePoolService).solve() -> int z rozwiązaniem
non_process
main description main genepoolservice wszystkie składowe genepool wartość szukana liczba genów genepoolservice solve int z rozwiązaniem
0
60,723
17,023,503,209
IssuesEvent
2021-07-03 02:21:37
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
strange routing decisions
Component: gosmore Priority: major Resolution: worksforme Type: defect
**[Submitted to the original trac issue database at 12.35pm, Saturday, 7th November 2009]** Latest builds of Gosmore seem to take some strange routing decisions. The three cases I have seen so far all occur when a roundabout is involved. Attached are screenshots of Gosmore routing decisions together with a text file showing the route as a www.yournavigation.org link (whilst I appreciate yournavigation doesn't use the current Gosmore version, providing these links seemed to me the best way of summarising the route I put into Gosmore).
1.0
strange routing decisions - **[Submitted to the original trac issue database at 12.35pm, Saturday, 7th November 2009]** Latest builds of Gosmore seem to take some strange routing decisions. The three cases I have seen so far all occur when a roundabout is involved. Attached are screenshots of Gosmore routing decisions together with a text file showing the route as a www.yournavigation.org link (whilst I appreciate yournavigation doesn't use the current Gosmore version, providing these links seemed to me the best way of summarising the route I put into Gosmore).
non_process
strange routing decisions latest builds of gosmore seem to take some strange routing decisions the three cases i have seen so far all occur when a roundabout is involved attached are screenshots of gosmore routing decisions together with a text file showing the route as a link whilst i appreciate yournavigation doesn t use the current gosmore version providing these links seemed to me the best way of summarising the route i put into gosmore
0
194,580
22,262,036,522
IssuesEvent
2022-06-10 02:01:08
Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034
https://api.github.com/repos/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034
reopened
CVE-2021-33033 (High) detected in multiple libraries
security vulnerability
## CVE-2021-33033 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Linux kernel before 5.11.14 has a use-after-free in cipso_v4_genopt in net/ipv4/cipso_ipv4.c because the CIPSO and CALIPSO refcounting for the DOI definitions is mishandled, aka CID-ad5d07f4a9cd. This leads to writing an arbitrary value. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33033>CVE-2021-33033</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-33033">https://www.linuxkernelcves.com/cves/CVE-2021-33033</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: v4.14.226, v4.19.181, v5.4.106, v5.10.24, v5.11.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-33033 (High) detected in multiple libraries - ## CVE-2021-33033 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Linux kernel before 5.11.14 has a use-after-free in cipso_v4_genopt in net/ipv4/cipso_ipv4.c because the CIPSO and CALIPSO refcounting for the DOI definitions is mishandled, aka CID-ad5d07f4a9cd. This leads to writing an arbitrary value. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33033>CVE-2021-33033</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-33033">https://www.linuxkernelcves.com/cves/CVE-2021-33033</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: v4.14.226, v4.19.181, v5.4.106, v5.10.24, v5.11.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries linuxlinux linuxlinux linuxlinux vulnerability details the linux kernel before has a use after free in cipso genopt in net cipso c because the cipso and calipso refcounting for the doi definitions is mishandled aka cid this leads to writing an arbitrary value publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
240,662
18,364,263,303
IssuesEvent
2021-10-09 19:30:57
malodetz/Fl-project
https://api.github.com/repos/malodetz/Fl-project
closed
Поправить ConcreteSyntax.md
documentation good first issue
Кажется, что литералы и названия ключевых функций не являются ключевыми словами. Оставить в этой секции только управляющие конструкции и создать новую секцию: "Зарезервированные слова"
1.0
Поправить ConcreteSyntax.md - Кажется, что литералы и названия ключевых функций не являются ключевыми словами. Оставить в этой секции только управляющие конструкции и создать новую секцию: "Зарезервированные слова"
non_process
поправить concretesyntax md кажется что литералы и названия ключевых функций не являются ключевыми словами оставить в этой секции только управляющие конструкции и создать новую секцию зарезервированные слова
0
13,384
15,862,500,366
IssuesEvent
2021-04-08 11:41:09
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Update esbuild
kind/tech process/candidate team/client
Unfortunately, the latest esbuild is not compatible with webpack. That's why we had to downgrade it. We should investigate the problem and make sure we can upgrade to the latest esbuild. https://github.com/prisma/prisma/pull/6260
1.0
Update esbuild - Unfortunately, the latest esbuild is not compatible with webpack. That's why we had to downgrade it. We should investigate the problem and make sure we can upgrade to the latest esbuild. https://github.com/prisma/prisma/pull/6260
process
update esbuild unfortunately the latest esbuild is not compatible with webpack that s why we had to downgrade it we should investigate the problem and make sure we can upgrade to the latest esbuild
1
21,628
30,030,840,395
IssuesEvent
2023-06-27 09:28:51
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
opened
ffi CI job is slow all of a sudden
T-bug C-process
It gets stuck here: ``` ... 2023-06-27T06:21:24.1339208Z cp target/wasm32-unknown-unknown/release/icu_capi_cdylib.wasm icu_capi.wasm 2023-06-27T06:21:24.1355342Z cargo run -p icu_datagen -- \ 2023-06-27T06:21:24.1355873Z --keys-for-bin icu_capi.wasm \ 2023-06-27T06:21:24.1356257Z --locales full \ 2023-06-27T06:21:24.1356610Z --format blob \ 2023-06-27T06:21:24.1357007Z --out data.postcard 2023-06-27T06:21:24.7096518Z Finished dev [unoptimized + debuginfo] target(s) in 0.53s 2023-06-27T06:21:24.7154466Z Running `/home/runner/work/icu4x/icu4x/target/debug/icu4x-datagen --keys-for-bin icu_capi.wasm --locales full --format blob --out data.postcard` 2023-06-27T09:18:34.1380351Z INFO [icu_provider_blob::export::blob_exporter] Serializing blob to output stream... 2023-06-27T09:18:34.2128602Z rm -rf lib; 2023-06-27T09:18:34.2148391Z mkdir lib; 2023-06-27T09:18:34.2167144Z cp ../../include/* lib ... ``` The first commit exhibiting this behavior is https://github.com/unicode-org/icu4x/commit/c394c1c6c121e15b05aa7b0aa98826c0cf965694 but it's unclear why that would be at fault.
1.0
ffi CI job is slow all of a sudden - It gets stuck here: ``` ... 2023-06-27T06:21:24.1339208Z cp target/wasm32-unknown-unknown/release/icu_capi_cdylib.wasm icu_capi.wasm 2023-06-27T06:21:24.1355342Z cargo run -p icu_datagen -- \ 2023-06-27T06:21:24.1355873Z --keys-for-bin icu_capi.wasm \ 2023-06-27T06:21:24.1356257Z --locales full \ 2023-06-27T06:21:24.1356610Z --format blob \ 2023-06-27T06:21:24.1357007Z --out data.postcard 2023-06-27T06:21:24.7096518Z Finished dev [unoptimized + debuginfo] target(s) in 0.53s 2023-06-27T06:21:24.7154466Z Running `/home/runner/work/icu4x/icu4x/target/debug/icu4x-datagen --keys-for-bin icu_capi.wasm --locales full --format blob --out data.postcard` 2023-06-27T09:18:34.1380351Z INFO [icu_provider_blob::export::blob_exporter] Serializing blob to output stream... 2023-06-27T09:18:34.2128602Z rm -rf lib; 2023-06-27T09:18:34.2148391Z mkdir lib; 2023-06-27T09:18:34.2167144Z cp ../../include/* lib ... ``` The first commit exhibiting this behavior is https://github.com/unicode-org/icu4x/commit/c394c1c6c121e15b05aa7b0aa98826c0cf965694 but it's unclear why that would be at fault.
process
ffi ci job is slow all of a sudden it gets stuck here cp target unknown unknown release icu capi cdylib wasm icu capi wasm cargo run p icu datagen keys for bin icu capi wasm locales full format blob out data postcard finished dev target s in running home runner work target debug datagen keys for bin icu capi wasm locales full format blob out data postcard info serializing blob to output stream rm rf lib mkdir lib cp include lib the first commit exhibiting this behavior is but it s unclear why that would be at fault
1
202,767
15,300,100,127
IssuesEvent
2021-02-24 11:51:03
tempesta-tech/tempesta
https://api.github.com/repos/tempesta-tech/tempesta
closed
HTTPtables blocking action ignores block_action directive
bug good to start test
# Scope On configuration ``` listen 192.168.100.4:80; block_action attack reply; block_action error reply; response_body 400 /var/www/tempesta-tech.com/400; srv_group default { server 127.0.0.1:9090; } vhost debian { proxy_pass default; } cache 1; cache_fulfill * *; http_chain l7_rules { hdr "Referer" == "http://badhost.com*" -> block; } http_chain { mark == 1 -> l7_rules; -> debian; } ``` Note that there is no default host as well no default rule for `l7_rules` chain. Configure iptables as: ``` iptables -A INPUT -s 192.168.100.1 -j MARK --set-mark 1 ``` Next try to send 2 requests from `192.168.100.1`: 1. with `Referer: http://goodhost.com` 2. with `Referer: http://badhost.com/foo` In both the cases HTTPtables block the requests, but Tempesta does not send any HTTP response as expected. From browsing the code it seems we just ignore `block` argument returned from HTTPtables. # Testing For the scenario and configuration as the above: - [ ] request with `Referer: http://goodhost.com` - [ ] request with `Referer: http://badhost.com/foo` - [ ] Check that configuration option `response_body` also does the work. In my demo scenario I used ``` block_action attack reply; block_action error reply; response_body 400 /var/www/tempesta-tech.com/400; ``` Please also add tests for the configuration described in https://github.com/tempesta-tech/tempesta/wiki/HTTP-tables#default-rules-and-chains to `./http_rules/test_http_tables.py` - [ ] Drop a request if there is no matching rule in a chain - [ ] implicit creation of the main chain - [ ] Dropping all requests with an empy main chain
1.0
HTTPtables blocking action ignores block_action directive - # Scope On configuration ``` listen 192.168.100.4:80; block_action attack reply; block_action error reply; response_body 400 /var/www/tempesta-tech.com/400; srv_group default { server 127.0.0.1:9090; } vhost debian { proxy_pass default; } cache 1; cache_fulfill * *; http_chain l7_rules { hdr "Referer" == "http://badhost.com*" -> block; } http_chain { mark == 1 -> l7_rules; -> debian; } ``` Note that there is no default host as well no default rule for `l7_rules` chain. Configure iptables as: ``` iptables -A INPUT -s 192.168.100.1 -j MARK --set-mark 1 ``` Next try to send 2 requests from `192.168.100.1`: 1. with `Referer: http://goodhost.com` 2. with `Referer: http://badhost.com/foo` In both the cases HTTPtables block the requests, but Tempesta does not send any HTTP response as expected. From browsing the code it seems we just ignore `block` argument returned from HTTPtables. # Testing For the scenario and configuration as the above: - [ ] request with `Referer: http://goodhost.com` - [ ] request with `Referer: http://badhost.com/foo` - [ ] Check that configuration option `response_body` also does the work. In my demo scenario I used ``` block_action attack reply; block_action error reply; response_body 400 /var/www/tempesta-tech.com/400; ``` Please also add tests for the configuration described in https://github.com/tempesta-tech/tempesta/wiki/HTTP-tables#default-rules-and-chains to `./http_rules/test_http_tables.py` - [ ] Drop a request if there is no matching rule in a chain - [ ] implicit creation of the main chain - [ ] Dropping all requests with an empy main chain
non_process
httptables blocking action ignores block action directive scope on configuration listen block action attack reply block action error reply response body var www tempesta tech com srv group default server vhost debian proxy pass default cache cache fulfill http chain rules hdr referer block http chain mark rules debian note that there is no default host as well no default rule for rules chain configure iptables as iptables a input s j mark set mark next try to send requests from with referer with referer in both the cases httptables block the requests but tempesta does not send any http response as expected from browsing the code it seems we just ignore block argument returned from httptables testing for the scenario and configuration as the above request with referer request with referer check that configuration option response body also does the work in my demo scenario i used block action attack reply block action error reply response body var www tempesta tech com please also add tests for the configuration described in to http rules test http tables py drop a request if there is no matching rule in a chain implicit creation of the main chain dropping all requests with an empy main chain
0
87,162
3,737,636,218
IssuesEvent
2016-03-08 19:58:02
aces/cbrain
https://api.github.com/repos/aces/cbrain
closed
Launch Task of the select tool overlay bug.
High Priority
It redirect to index page, don't know if it should be the case. I think: 1) Should stay on the overlay. 2) Should redirect on the new task page. 3) Should definitively not put the launch at the bottom of the index table. Like this: ![screen shot 2016-02-18 at 3 58 25 pm](https://cloud.githubusercontent.com/assets/777590/13157993/b60bb0d6-d658-11e5-9524-d10584012c7d.png)
1.0
Launch Task of the select tool overlay bug. - It redirect to index page, don't know if it should be the case. I think: 1) Should stay on the overlay. 2) Should redirect on the new task page. 3) Should definitively not put the launch at the bottom of the index table. Like this: ![screen shot 2016-02-18 at 3 58 25 pm](https://cloud.githubusercontent.com/assets/777590/13157993/b60bb0d6-d658-11e5-9524-d10584012c7d.png)
non_process
launch task of the select tool overlay bug it redirect to index page don t know if it should be the case i think should stay on the overlay should redirect on the new task page should definitively not put the launch at the bottom of the index table like this
0
469
2,731,262,838
IssuesEvent
2015-04-16 19:19:24
mozilla/id.webmaker.org
https://api.github.com/repos/mozilla/id.webmaker.org
closed
Implement CSRF
Security
- [ ] find out how this will work with the react router - [ ] should do csrf on per-request basis (rather than per-page-load)? - is this really more secure? - hash per request or secret per request?
True
Implement CSRF - - [ ] find out how this will work with the react router - [ ] should do csrf on per-request basis (rather than per-page-load)? - is this really more secure? - hash per request or secret per request?
non_process
implement csrf find out how this will work with the react router should do csrf on per request basis rather than per page load is this really more secure hash per request or secret per request
0
269,429
23,442,060,951
IssuesEvent
2022-08-15 15:47:30
RPTools/maptool
https://api.github.com/repos/RPTools/maptool
closed
[Bug]: occasional NullPointerException on local WebRTC client connection
bug tested
### Describe the Bug After starting a WebRTC server it failed to start properly. ``` 18:27:52.457 (ClientHandshake.java:145) [Thread-16] INFO - skullman :send: CLIENT_INIT_MSG Exception in thread "Thread-16" java.lang.NullPointerException: Cannot invoke "dev.onvoid.webrtc.RTCPeerConnection.getConnectionState()" because "this.peerConnection" is null at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:185) at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:178) at net.rptools.clientserver.hessian.client.MethodClientDecorator.sendMessage(MethodClientDecorator.java:42) at net.rptools.maptool.server.ClientHandshake.sendMessage(ClientHandshake.java:146) at net.rptools.maptool.server.ClientHandshake.startHandshake(ClientHandshake.java:139) at net.rptools.maptool.client.MapToolConnection.start(MapToolConnection.java:76) at net.rptools.maptool.client.MapTool.createConnection(MapTool.java:1227) at net.rptools.maptool.client.AppActions$62.lambda$executeAction$1(AppActions.java:2326) at net.rptools.maptool.client.AppActions$ClientAction.lambda$runBackground$0(AppActions.java:3496) at java.base/java.lang.Thread.run(Thread.java:831) 18:27:52.546 (WebRTCServerConnection.java:70) [WebSocketConnectReadThread-81] INFO - S WebSocket connected 18:27:52.762 (WebRTCClientConnection.java:136) [WebSocketConnectReadThread-84] INFO - C WebSocket connected 18:27:52.922 (WebRTCClientConnection.java:408) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onRenegotiationNeeded 18:27:52.926 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: HAVE_LOCAL_OFFER 18:27:52.936 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange GATHERING [sentry-pool-1-thread-2] WARN io.sentry.connection.AbstractConnection.lockdown - Initiated a temporary lockdown because of exception: {"detail":"bad sentry DSN public key"} [sentry-pool-1-thread-2] ERROR io.sentry.connection.AsyncConnection - An exception occurred while sending the event to Sentry. io.sentry.connection.ConnectionException: {"detail":"bad sentry DSN public key"} at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:216) at io.sentry.connection.AbstractConnection.send(AbstractConnection.java:88) at io.sentry.connection.AsyncConnection$EventSubmitter.run(AsyncConnection.java:187) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630) at java.base/java.lang.Thread.run(Thread.java:831) Caused by: java.io.IOException: Server returned HTTP response code: 400 for URL: https://sentry.io/api/1404248/store/ at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1985) at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1577) at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getInputStream(HttpsURLConnectionImpl.java:224) at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:172) ... 5 more 18:27:53.094 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: HAVE_REMOTE_OFFER 18:27:53.096 (WebRTCClientConnection.java:73) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S setRemoteDescription success. 18:27:53.098 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: STABLE 18:27:53.100 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange GATHERING 18:27:53.240 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: STABLE 18:27:53.240 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CHECKING 18:27:53.241 (WebRTCClientConnection.java:243) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C setRemoteDescription success. 18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTING 18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTING 18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange COMPLETE 18:27:53.257 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CHECKING 18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange COMPLETE 18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange COMPLETED 18:27:53.261 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C localDataChannel onStateChange OPEN 18:27:53.261 (WebRTCClientConnection.java:395) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onDataChannel 18:27:53.262 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S localDataChannel onStateChange OPEN 18:27:53.264 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_INIT_MSG 18:27:53.266 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: USE_AUTH_TYPE_MSG 18:27:53.267 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: USE_AUTH_TYPE_MSG 18:27:53.271 (ClientHandshake.java:145) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :send: CLIENT_AUTH_MESSAGE 18:27:53.272 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_AUTH_MESSAGE 18:27:53.272 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: CONNECTION_SUCCESSFUL_MSG 18:27:53.277 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: CONNECTION_SUCCESSFUL_MSG ``` ### To Reproduce Start a WebRTC server a bunch of times and get unlucky. ### Expected Behaviour Local client is always able to send messages. ### Screenshots _No response_ ### MapTool Info 1.11.4 with WebRTC patch ### Desktop Linux[Debian] ### Additional Context _No response_
1.0
[Bug]: occasional NullPointerException on local WebRTC client connection - ### Describe the Bug After starting a WebRTC server it failed to start properly. ``` 18:27:52.457 (ClientHandshake.java:145) [Thread-16] INFO - skullman :send: CLIENT_INIT_MSG Exception in thread "Thread-16" java.lang.NullPointerException: Cannot invoke "dev.onvoid.webrtc.RTCPeerConnection.getConnectionState()" because "this.peerConnection" is null at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:185) at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:178) at net.rptools.clientserver.hessian.client.MethodClientDecorator.sendMessage(MethodClientDecorator.java:42) at net.rptools.maptool.server.ClientHandshake.sendMessage(ClientHandshake.java:146) at net.rptools.maptool.server.ClientHandshake.startHandshake(ClientHandshake.java:139) at net.rptools.maptool.client.MapToolConnection.start(MapToolConnection.java:76) at net.rptools.maptool.client.MapTool.createConnection(MapTool.java:1227) at net.rptools.maptool.client.AppActions$62.lambda$executeAction$1(AppActions.java:2326) at net.rptools.maptool.client.AppActions$ClientAction.lambda$runBackground$0(AppActions.java:3496) at java.base/java.lang.Thread.run(Thread.java:831) 18:27:52.546 (WebRTCServerConnection.java:70) [WebSocketConnectReadThread-81] INFO - S WebSocket connected 18:27:52.762 (WebRTCClientConnection.java:136) [WebSocketConnectReadThread-84] INFO - C WebSocket connected 18:27:52.922 (WebRTCClientConnection.java:408) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onRenegotiationNeeded 18:27:52.926 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: HAVE_LOCAL_OFFER 18:27:52.936 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange GATHERING [sentry-pool-1-thread-2] WARN io.sentry.connection.AbstractConnection.lockdown - Initiated a temporary lockdown because of exception: {"detail":"bad sentry DSN public key"} [sentry-pool-1-thread-2] ERROR io.sentry.connection.AsyncConnection - An exception occurred while sending the event to Sentry. io.sentry.connection.ConnectionException: {"detail":"bad sentry DSN public key"} at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:216) at io.sentry.connection.AbstractConnection.send(AbstractConnection.java:88) at io.sentry.connection.AsyncConnection$EventSubmitter.run(AsyncConnection.java:187) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630) at java.base/java.lang.Thread.run(Thread.java:831) Caused by: java.io.IOException: Server returned HTTP response code: 400 for URL: https://sentry.io/api/1404248/store/ at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1985) at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1577) at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getInputStream(HttpsURLConnectionImpl.java:224) at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:172) ... 5 more 18:27:53.094 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: HAVE_REMOTE_OFFER 18:27:53.096 (WebRTCClientConnection.java:73) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S setRemoteDescription success. 18:27:53.098 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: STABLE 18:27:53.100 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange GATHERING 18:27:53.240 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: STABLE 18:27:53.240 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CHECKING 18:27:53.241 (WebRTCClientConnection.java:243) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C setRemoteDescription success. 18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTING 18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTING 18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange COMPLETE 18:27:53.257 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CHECKING 18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange COMPLETE 18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CONNECTED 18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange COMPLETED 18:27:53.261 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C localDataChannel onStateChange OPEN 18:27:53.261 (WebRTCClientConnection.java:395) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onDataChannel 18:27:53.262 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S localDataChannel onStateChange OPEN 18:27:53.264 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_INIT_MSG 18:27:53.266 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: USE_AUTH_TYPE_MSG 18:27:53.267 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: USE_AUTH_TYPE_MSG 18:27:53.271 (ClientHandshake.java:145) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :send: CLIENT_AUTH_MESSAGE 18:27:53.272 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_AUTH_MESSAGE 18:27:53.272 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: CONNECTION_SUCCESSFUL_MSG 18:27:53.277 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: CONNECTION_SUCCESSFUL_MSG ``` ### To Reproduce Start a WebRTC server a bunch of times and get unlucky. ### Expected Behaviour Local client is always able to send messages. ### Screenshots _No response_ ### MapTool Info 1.11.4 with WebRTC patch ### Desktop Linux[Debian] ### Additional Context _No response_
non_process
occasional nullpointerexception on local webrtc client connection describe the bug after starting a webrtc server it failed to start properly clienthandshake java info skullman send client init msg exception in thread thread java lang nullpointerexception cannot invoke dev onvoid webrtc rtcpeerconnection getconnectionstate because this peerconnection is null at net rptools clientserver simple client webrtcclientconnection sendmessage webrtcclientconnection java at net rptools clientserver simple client webrtcclientconnection sendmessage webrtcclientconnection java at net rptools clientserver hessian client methodclientdecorator sendmessage methodclientdecorator java at net rptools maptool server clienthandshake sendmessage clienthandshake java at net rptools maptool server clienthandshake starthandshake clienthandshake java at net rptools maptool client maptoolconnection start maptoolconnection java at net rptools maptool client maptool createconnection maptool java at net rptools maptool client appactions lambda executeaction appactions java at net rptools maptool client appactions clientaction lambda runbackground appactions java at java base java lang thread run thread java webrtcserverconnection java info s websocket connected webrtcclientconnection java info c websocket connected webrtcclientconnection java info c peerconnection onrenegotiationneeded webrtcclientconnection java info c peerconnection onsignalingchange have local offer webrtcclientconnection java info c peerconnection onicegatheringchange gathering warn io sentry connection abstractconnection lockdown initiated a temporary lockdown because of exception detail bad sentry dsn public key error io sentry connection asyncconnection an exception occurred while sending the event to sentry io sentry connection connectionexception detail bad sentry dsn public key at io sentry connection httpconnection dosend httpconnection java at io sentry connection abstractconnection send abstractconnection java at io sentry connection asyncconnection eventsubmitter run asyncconnection java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java io ioexception server returned http response code for url at java base sun net at java base sun net at java base sun net at io sentry connection httpconnection dosend httpconnection java more webrtcclientconnection java info s peerconnection onsignalingchange have remote offer webrtcclientconnection java info s setremotedescription success webrtcclientconnection java info s peerconnection onsignalingchange stable webrtcclientconnection java info s peerconnection onicegatheringchange gathering webrtcclientconnection java info c peerconnection onsignalingchange stable webrtcclientconnection java info c peerconnection oniceconnectionchange checking webrtcclientconnection java info c setremotedescription success webrtcclientconnection java info c peerconnection onconnectionchange connecting webrtcclientconnection java info s peerconnection onconnectionchange connecting webrtcclientconnection java info c peerconnection onicegatheringchange complete webrtcclientconnection java info s peerconnection oniceconnectionchange checking webrtcclientconnection java info s peerconnection onicegatheringchange complete webrtcclientconnection java info c peerconnection onconnectionchange connected webrtcclientconnection java info s peerconnection onconnectionchange connected webrtcclientconnection java info c peerconnection oniceconnectionchange connected webrtcclientconnection java info s peerconnection oniceconnectionchange connected webrtcclientconnection java info c peerconnection oniceconnectionchange completed webrtcclientconnection java info c localdatachannel onstatechange open webrtcclientconnection java info s peerconnection ondatachannel webrtcclientconnection java info s localdatachannel onstatechange open serverhandshake java info skullman draconis got client init msg serverhandshake java info skullman draconis send use auth type msg clienthandshake java info skullman got use auth type msg clienthandshake java info skullman send client auth message serverhandshake java info skullman draconis got client auth message serverhandshake java info skullman draconis send connection successful msg clienthandshake java info skullman got connection successful msg to reproduce start a webrtc server a bunch of times and get unlucky expected behaviour local client is always able to send messages screenshots no response maptool info with webrtc patch desktop linux additional context no response
0
113,657
9,660,262,093
IssuesEvent
2019-05-20 15:08:01
OpenTechFund/opentech.fund
https://api.github.com/repos/OpenTechFund/opentech.fund
closed
Comment editing in submission detail view
RTBC needs tests todd_mini_sprint
Task from #1128 **Acceptance criteria** - [x] Private comments in the detail view allow editing by the author of the comment - [x] Internal comments in the detail view allow editing by the author of the comment - [x] Edit button is visible on comment (styled as per edit button on submission detail page) - [x] Inline editing of comment with 'submit' and 'cancel' button - [x] Updated comment to display on edit submission - [x] Last edited date to be displayed alongside post date - [x] Comment sort order remains as it currently is (by original post date) **QA Criteria** Dev: - [x] checked feature meets acceptance criteria/conforms exactly to the specification. - [x] provided good unit test coverage (if this is non-trivial behaviour). - [x] checked all tests for the project pass with this feature enabled. - [x] checked code conforms to the project coding standards. - [x] had code reviewed by another developer and resolved any issues raised. - [x] tested this feature as an end user of the website/app (Can I get to it? Is it useable? Can I break it? Does it work in an end-to-end context?) - [x] checked that this feature works on the server/s I am deploying it to. QA: - [ ] tested this feature as a front end user and it meets the acceptance criteria/conforms to the specification and design. - [ ] checked that the feature works on the server/s deployed to
1.0
Comment editing in submission detail view - Task from #1128 **Acceptance criteria** - [x] Private comments in the detail view allow editing by the author of the comment - [x] Internal comments in the detail view allow editing by the author of the comment - [x] Edit button is visible on comment (styled as per edit button on submission detail page) - [x] Inline editing of comment with 'submit' and 'cancel' button - [x] Updated comment to display on edit submission - [x] Last edited date to be displayed alongside post date - [x] Comment sort order remains as it currently is (by original post date) **QA Criteria** Dev: - [x] checked feature meets acceptance criteria/conforms exactly to the specification. - [x] provided good unit test coverage (if this is non-trivial behaviour). - [x] checked all tests for the project pass with this feature enabled. - [x] checked code conforms to the project coding standards. - [x] had code reviewed by another developer and resolved any issues raised. - [x] tested this feature as an end user of the website/app (Can I get to it? Is it useable? Can I break it? Does it work in an end-to-end context?) - [x] checked that this feature works on the server/s I am deploying it to. QA: - [ ] tested this feature as a front end user and it meets the acceptance criteria/conforms to the specification and design. - [ ] checked that the feature works on the server/s deployed to
non_process
comment editing in submission detail view task from acceptance criteria private comments in the detail view allow editing by the author of the comment internal comments in the detail view allow editing by the author of the comment edit button is visible on comment styled as per edit button on submission detail page inline editing of comment with submit and cancel button updated comment to display on edit submission last edited date to be displayed alongside post date comment sort order remains as it currently is by original post date qa criteria dev checked feature meets acceptance criteria conforms exactly to the specification provided good unit test coverage if this is non trivial behaviour checked all tests for the project pass with this feature enabled checked code conforms to the project coding standards had code reviewed by another developer and resolved any issues raised tested this feature as an end user of the website app can i get to it is it useable can i break it does it work in an end to end context checked that this feature works on the server s i am deploying it to qa tested this feature as a front end user and it meets the acceptance criteria conforms to the specification and design checked that the feature works on the server s deployed to
0
239,786
26,232,081,030
IssuesEvent
2023-01-05 01:45:10
pustovitDmytro/lalaps
https://api.github.com/repos/pustovitDmytro/lalaps
opened
CVE-2022-24999 (High) detected in qs-6.10.1.tgz
security vulnerability
## CVE-2022-24999 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/superagent/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - supertest-6.1.6.tgz (Root Library) - superagent-6.1.0.tgz - :x: **qs-6.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/pustovitDmytro/lalaps/commit/0888eee49d490c03a2d773287f1270e8a617d8ca">0888eee49d490c03a2d773287f1270e8a617d8ca</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable). <p>Publish Date: 2022-11-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p> <p>Release Date: 2022-11-26</p> <p>Fix Resolution (qs): 6.10.3</p> <p>Direct dependency fix Resolution (supertest): 6.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24999 (High) detected in qs-6.10.1.tgz - ## CVE-2022-24999 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/superagent/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - supertest-6.1.6.tgz (Root Library) - superagent-6.1.0.tgz - :x: **qs-6.10.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/pustovitDmytro/lalaps/commit/0888eee49d490c03a2d773287f1270e8a617d8ca">0888eee49d490c03a2d773287f1270e8a617d8ca</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable). <p>Publish Date: 2022-11-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p> <p>Release Date: 2022-11-26</p> <p>Fix Resolution (qs): 6.10.3</p> <p>Direct dependency fix Resolution (supertest): 6.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules superagent node modules qs package json dependency hierarchy supertest tgz root library superagent tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details qs before as used in express before and other products allows attackers to cause a node process hang for an express application because an proto key can be used in many typical express use cases an unauthenticated remote attacker can place the attack payload in the query string of the url that is used to visit the application such as a b a a the fix was backported to qs and and therefore express which has deps qs in its release description is not vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs direct dependency fix resolution supertest step up your open source security game with mend
0
24,324
12,065,516,652
IssuesEvent
2020-04-16 10:06:04
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Document licenses for third party open source projects that we depend on
planned/3.15 team/core-services
We need something like this: https://neo4j.com/terms/thirdpartylicenses/ The page should be generated automatically. Tagging both @lguychard and @tsenart because the implementation is probably different for TypeScript and Go dependencies. I would like to prioritize this for 3.15 release on April 20 but if you want to do it earlier, I wouldn't complain :) slack thread: https://sourcegraph.slack.com/archives/C0B2RU51Q/p1582918787128100
1.0
Document licenses for third party open source projects that we depend on - We need something like this: https://neo4j.com/terms/thirdpartylicenses/ The page should be generated automatically. Tagging both @lguychard and @tsenart because the implementation is probably different for TypeScript and Go dependencies. I would like to prioritize this for 3.15 release on April 20 but if you want to do it earlier, I wouldn't complain :) slack thread: https://sourcegraph.slack.com/archives/C0B2RU51Q/p1582918787128100
non_process
document licenses for third party open source projects that we depend on we need something like this the page should be generated automatically tagging both lguychard and tsenart because the implementation is probably different for typescript and go dependencies i would like to prioritize this for release on april but if you want to do it earlier i wouldn t complain slack thread
0
6,438
9,540,966,250
IssuesEvent
2019-04-30 20:57:23
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
opened
Firestore: 'test_collection_add' flakes
api: firestore flaky testing type: process
From https://source.cloud.google.com/results/invocations/7deedac3-8b60-447d-bb5e-74a4a6653e57/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Ffirestore/log: ```python _____________________________ test_collection_add ______________________________ client = <google.cloud.firestore_v1.client.Client object at 0x7fb9ca171ad0> cleanup = <built-in method append of list object at 0x7fb9c7c0acb0> def test_collection_add(client, cleanup): collection1 = client.collection("collek") collection2 = client.collection("collek", "shun", "child") collection3 = client.collection("collek", "table", "child") explicit_doc_id = "hula" + unique_resource_id("-") > assert set(collection1.list_documents()) == set() E assert set([<google....fb9c7c06290>]) == set([]) E Extra items in the left set: E <google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290> E Full diff: E - set([<google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290>]) E + set([]) ``` This test should be using a unique / name-mangled collection ID, in order to prevent collision with CI running for other PRs.
1.0
Firestore: 'test_collection_add' flakes - From https://source.cloud.google.com/results/invocations/7deedac3-8b60-447d-bb5e-74a4a6653e57/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Ffirestore/log: ```python _____________________________ test_collection_add ______________________________ client = <google.cloud.firestore_v1.client.Client object at 0x7fb9ca171ad0> cleanup = <built-in method append of list object at 0x7fb9c7c0acb0> def test_collection_add(client, cleanup): collection1 = client.collection("collek") collection2 = client.collection("collek", "shun", "child") collection3 = client.collection("collek", "table", "child") explicit_doc_id = "hula" + unique_resource_id("-") > assert set(collection1.list_documents()) == set() E assert set([<google....fb9c7c06290>]) == set([]) E Extra items in the left set: E <google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290> E Full diff: E - set([<google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290>]) E + set([]) ``` This test should be using a unique / name-mangled collection ID, in order to prevent collision with CI running for other PRs.
process
firestore test collection add flakes from python test collection add client cleanup def test collection add client cleanup client collection collek client collection collek shun child client collection collek table child explicit doc id hula unique resource id assert set list documents set e assert set set e extra items in the left set e e full diff e set e set this test should be using a unique name mangled collection id in order to prevent collision with ci running for other prs
1
101,069
8,773,741,130
IssuesEvent
2018-12-18 17:42:32
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
no error message for bad behavior when unable to connect to sync servers
QA/Test-Plan-Specified QA/Yes feature/sync
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description Follow up to https://github.com/brave/brave-core/pull/1019 The final scenario in the test plan for the above PR did not pass QA. Scenario is `Bad behavior from Sync servers` in PR. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Create a sync chain and ensure you see two devices 2. Quit Brave 3. Launch Brave without internet 4. You should see a `Unable to connect to the Sync servers.` error dialog ## Actual result: No error dialog ## Expected result: Should be error dialog ## Reproduces how often: easily ## Brave version (brave://version info) Brave | 0.58.12 Chromium: 71.0.3578.80 (Official Build) (64-bit) -- | -- Revision | 2ac50e7249fbd55e6f517a28131605c9fb9fe897-refs/branch-heads/3578@{#860} OS | Mac OS X ### Reproducible on current release: - Does it reproduce on brave-browser dev/beta builds? ### Website problems only: - Does the issue resolve itself when disabling Brave Shields? n/a - Is the issue reproducible on the latest version of Chrome? n/a ### Additional Information cc @brave/legacy_qa for verification on other platforms
1.0
no error message for bad behavior when unable to connect to sync servers - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description Follow up to https://github.com/brave/brave-core/pull/1019 The final scenario in the test plan for the above PR did not pass QA. Scenario is `Bad behavior from Sync servers` in PR. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Create a sync chain and ensure you see two devices 2. Quit Brave 3. Launch Brave without internet 4. You should see a `Unable to connect to the Sync servers.` error dialog ## Actual result: No error dialog ## Expected result: Should be error dialog ## Reproduces how often: easily ## Brave version (brave://version info) Brave | 0.58.12 Chromium: 71.0.3578.80 (Official Build) (64-bit) -- | -- Revision | 2ac50e7249fbd55e6f517a28131605c9fb9fe897-refs/branch-heads/3578@{#860} OS | Mac OS X ### Reproducible on current release: - Does it reproduce on brave-browser dev/beta builds? ### Website problems only: - Does the issue resolve itself when disabling Brave Shields? n/a - Is the issue reproducible on the latest version of Chrome? n/a ### Additional Information cc @brave/legacy_qa for verification on other platforms
non_process
no error message for bad behavior when unable to connect to sync servers have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description follow up to the final scenario in the test plan for the above pr did not pass qa scenario is bad behavior from sync servers in pr steps to reproduce create a sync chain and ensure you see two devices quit brave launch brave without internet you should see a unable to connect to the sync servers error dialog actual result no error dialog expected result should be error dialog reproduces how often easily brave version brave version info brave chromium   official build   bit revision refs branch heads os mac os x reproducible on current release does it reproduce on brave browser dev beta builds website problems only does the issue resolve itself when disabling brave shields n a is the issue reproducible on the latest version of chrome n a additional information cc brave legacy qa for verification on other platforms
0
60,171
6,673,557,552
IssuesEvent
2017-10-04 15:25:41
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
UWP: 2 System.IO.Pipes failing after enabling other tests on same project
area-System.IO test-run-uwp-ilc
``` System.IO.Pipes.Tests.NamedPipeTest_Specific.Windows_GetImpersonationUserName_Succeed [FAIL] GetNamedPipeHandleState failed Expected: True Actual: False Stack Trace: at System.IO.Pipes.Tests.NamedPipeTest_Specific.<Windows_GetImpersonationUserName_Succeed>d__8.MoveNext() --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.RunAsClient_Windows [FAIL] Expected delegate to have been invoked Expected: True Actual: False Stack Trace: at System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.<RunAsClient_Windows>d__2.MoveNext() --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 xUnit.net console test runner (64-bit .NET Core) Copyright (C) 2014 Outercurve Foundation. ```
1.0
UWP: 2 System.IO.Pipes failing after enabling other tests on same project - ``` System.IO.Pipes.Tests.NamedPipeTest_Specific.Windows_GetImpersonationUserName_Succeed [FAIL] GetNamedPipeHandleState failed Expected: True Actual: False Stack Trace: at System.IO.Pipes.Tests.NamedPipeTest_Specific.<Windows_GetImpersonationUserName_Succeed>d__8.MoveNext() --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.RunAsClient_Windows [FAIL] Expected delegate to have been invoked Expected: True Actual: False Stack Trace: at System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.<RunAsClient_Windows>d__2.MoveNext() --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 --- End of stack trace from previous location where exception was thrown --- at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21 at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53 xUnit.net console test runner (64-bit .NET Core) Copyright (C) 2014 Outercurve Foundation. ```
non_process
uwp system io pipes failing after enabling other tests on same project system io pipes tests namedpipetest specific windows getimpersonationusername succeed getnamedpipehandlestate failed expected true actual false stack trace at system io pipes tests namedpipetest specific d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task at void xunit sdk testinvoker c b d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task system io pipes tests namedpipetest runasclient runasclient windows expected delegate to have been invoked expected true actual false stack trace at system io pipes tests namedpipetest runasclient d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task at void xunit sdk testinvoker c b d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task xunit net console test runner bit net core copyright c outercurve foundation
0
51,191
7,680,269,405
IssuesEvent
2018-05-16 00:34:39
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Add Objective-C API to disable fail fast mode
disposition/FOR DISCUSSION disposition/Needs Documentation kind/enhancement lang/ObjC priority/P3
As part of the 0.14 release, fail fast is now enabled by default. In order to disable fail fast mode, a flag must be set on metadata requests. https://github.com/grpc/grpc/blob/master/include/grpc/impl/codegen/grpc_types.h#L212 We need to ensure that there is some way a user can disable fail fast mode.
1.0
Add Objective-C API to disable fail fast mode - As part of the 0.14 release, fail fast is now enabled by default. In order to disable fail fast mode, a flag must be set on metadata requests. https://github.com/grpc/grpc/blob/master/include/grpc/impl/codegen/grpc_types.h#L212 We need to ensure that there is some way a user can disable fail fast mode.
non_process
add objective c api to disable fail fast mode as part of the release fail fast is now enabled by default in order to disable fail fast mode a flag must be set on metadata requests we need to ensure that there is some way a user can disable fail fast mode
0
14,607
17,703,636,347
IssuesEvent
2021-08-25 03:27:04
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - scientificName
Term - change Class - Taxon non-normative Process - complete
## Term change * Submitter: Quentin Groom * Efficacy Justification (why is this change necessary?): To improve clarity of the term usage, particularly to distinguish the different terms that can hold a scientific Latin name * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): This is largely for people and organizations publishing Darwin Core files to avoid repeated questions that keep cropping up. The issue #28 highlighted that the definitions of `scientificName`, `acceptedNameUsage `and `originalNameUsage` are all similar to one another, however, their intended usage is quite distinct, even though it is not clearly documented. The intension of this suggested change is to add to the comments of the term to help users understand the use of the terms more easily. The suggested explanations were given by @deepreef in #28, but they are only preliminary. * Stability Justification (what concerns are there that this might affect existing implementations?): The intension is that the comments would reinforce the existing definition and thus improve stability. * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No implication Current Term definition: https://dwc.tdwg.org/list/#dwc_scientificName Proposed attributes of the new term: * Usage comments (recommendations regarding content, etc., not normative): **This term should not contain identification qualifications, which should instead be supplied in the IdentificationQualifier term. When applied to an Organism or Occurrence, this term should be used to represent the scientific name that was applied to the associated Organism in accordance with the Taxon to which it was or is currently identified.** * Examples: `Coleoptera` (order). `Vespertilionidae` (family). `Manis` (genus). `Ctenomys sociabilis` (genus + specificEpithet). `Ambystoma tigrinum diaboli` (genus + specificEpithet + infraspecificEpithet). `Roptrocerus typographi (Györfi, 1952)` (genus + specificEpithet + scientificNameAuthorship), `Quercus agrifolia var. oxyadenia (Torr.) J.T. Howell` (genus + specificEpithet + taxonRank + infraspecificEpithet + scientificNameAuthorship).
1.0
Change term - scientificName - ## Term change * Submitter: Quentin Groom * Efficacy Justification (why is this change necessary?): To improve clarity of the term usage, particularly to distinguish the different terms that can hold a scientific Latin name * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): This is largely for people and organizations publishing Darwin Core files to avoid repeated questions that keep cropping up. The issue #28 highlighted that the definitions of `scientificName`, `acceptedNameUsage `and `originalNameUsage` are all similar to one another, however, their intended usage is quite distinct, even though it is not clearly documented. The intension of this suggested change is to add to the comments of the term to help users understand the use of the terms more easily. The suggested explanations were given by @deepreef in #28, but they are only preliminary. * Stability Justification (what concerns are there that this might affect existing implementations?): The intension is that the comments would reinforce the existing definition and thus improve stability. * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No implication Current Term definition: https://dwc.tdwg.org/list/#dwc_scientificName Proposed attributes of the new term: * Usage comments (recommendations regarding content, etc., not normative): **This term should not contain identification qualifications, which should instead be supplied in the IdentificationQualifier term. When applied to an Organism or Occurrence, this term should be used to represent the scientific name that was applied to the associated Organism in accordance with the Taxon to which it was or is currently identified.** * Examples: `Coleoptera` (order). `Vespertilionidae` (family). `Manis` (genus). `Ctenomys sociabilis` (genus + specificEpithet). `Ambystoma tigrinum diaboli` (genus + specificEpithet + infraspecificEpithet). `Roptrocerus typographi (Györfi, 1952)` (genus + specificEpithet + scientificNameAuthorship), `Quercus agrifolia var. oxyadenia (Torr.) J.T. Howell` (genus + specificEpithet + taxonRank + infraspecificEpithet + scientificNameAuthorship).
process
change term scientificname term change submitter quentin groom efficacy justification why is this change necessary to improve clarity of the term usage particularly to distinguish the different terms that can hold a scientific latin name demand justification if the change is semantic in nature name at least two organizations that independently need this term this is largely for people and organizations publishing darwin core files to avoid repeated questions that keep cropping up the issue highlighted that the definitions of scientificname acceptednameusage and originalnameusage are all similar to one another however their intended usage is quite distinct even though it is not clearly documented the intension of this suggested change is to add to the comments of the term to help users understand the use of the terms more easily the suggested explanations were given by deepreef in but they are only preliminary stability justification what concerns are there that this might affect existing implementations the intension is that the comments would reinforce the existing definition and thus improve stability implications for dwciri namespace does this change affect a dwciri term version no implication current term definition proposed attributes of the new term usage comments recommendations regarding content etc not normative this term should not contain identification qualifications which should instead be supplied in the identificationqualifier term when applied to an organism or occurrence this term should be used to represent the scientific name that was applied to the associated organism in accordance with the taxon to which it was or is currently identified examples coleoptera order vespertilionidae family manis genus ctenomys sociabilis genus specificepithet ambystoma tigrinum diaboli genus specificepithet infraspecificepithet roptrocerus typographi györfi genus specificepithet scientificnameauthorship quercus agrifolia var oxyadenia torr j t howell genus specificepithet taxonrank infraspecificepithet scientificnameauthorship
1
21,044
27,987,347,455
IssuesEvent
2023-03-26 20:49:17
bitfocus/companion-module-requests
https://api.github.com/repos/bitfocus/companion-module-requests
opened
Hive Player
NOT YET PROCESSED
- [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested** The name of the device, hardware, or software you would like to control: Hive Player / What you would like to be able to make it do from Companion: be able to control playback and layer opacity and skip cues etc. Direct links or attachments to the ethernet control protocol or API: I can email a pdf with full UDP commands or contact Hive at www.hive.run
1.0
Hive Player - - [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested** The name of the device, hardware, or software you would like to control: Hive Player / What you would like to be able to make it do from Companion: be able to control playback and layer opacity and skip cues etc. Direct links or attachments to the ethernet control protocol or API: I can email a pdf with full UDP commands or contact Hive at www.hive.run
process
hive player i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control hive player what you would like to be able to make it do from companion be able to control playback and layer opacity and skip cues etc direct links or attachments to the ethernet control protocol or api i can email a pdf with full udp commands or contact hive at
1
22,202
30,758,364,639
IssuesEvent
2023-07-29 11:08:21
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
skypilot-nightly 1.0.0.dev20230729 has 2 GuardDog issues
guarddog exec-base64 silent-process-execution
https://pypi.org/project/skypilot-nightly https://inspector.pypi.io/project/skypilot-nightly ```{ "dependency": "skypilot-nightly", "version": "1.0.0.dev20230729", "result": { "issues": 2, "errors": {}, "results": { "exec-base64": [ { "location": "skypilot-nightly-1.0.0.dev20230729/sky/cloud_stores.py:113", "code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "silent-process-execution": [ { "location": "skypilot-nightly-1.0.0.dev20230729/sky/skylet/log_lib.py:219", "code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpwfkzzrpo/skypilot-nightly" } }```
1.0
skypilot-nightly 1.0.0.dev20230729 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly https://inspector.pypi.io/project/skypilot-nightly ```{ "dependency": "skypilot-nightly", "version": "1.0.0.dev20230729", "result": { "issues": 2, "errors": {}, "results": { "exec-base64": [ { "location": "skypilot-nightly-1.0.0.dev20230729/sky/cloud_stores.py:113", "code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')", "message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n" } ], "silent-process-execution": [ { "location": "skypilot-nightly-1.0.0.dev20230729/sky/skylet/log_lib.py:219", "code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpwfkzzrpo/skypilot-nightly" } }```
process
skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpwfkzzrpo skypilot nightly
1
637,702
20,675,805,666
IssuesEvent
2022-03-10 09:09:38
geosolutions-it/MapStore2-C027
https://api.github.com/repos/geosolutions-it/MapStore2-C027
closed
Requested changes to GeoNetwork
Task Priority: Medium C027-COMUNE_FI-2021-SUPPORT
The client is requesting few changes to the new GN 3 RNDT installed in the new SDI. From the client: >Per quanto riguarda l'interfaccia, invece avremmo la necessità di due migliorie che sono: >- Nella pagina di ricerca dei dataset, tra i filtraggi aggiungere un filtro che consente di estrarre i risultati in base ad un range di "Data di pubblicazione" (filtrare tutti i dataset con Data di pubblicazione da gg/mm/aaaa a gg/mm/aaaa) ![image.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/c5674621-bc26-41ed-bf88-93d63505bcf9) >- Nel pdf che è possibile esportare dopo aver selezionato i dataset eliminare i seguenti campi: > - Update frequency > - Record updated > - Risorse > - More onlineinformation >Quindi gli unici campi visibili dovranno esssere: > - Descrizione > - Estensione > - Parole chiave > - Access and useconstraints > - Data type > - Format > - Identificatoreunivoco ![image](https://user-images.githubusercontent.com/1280027/155680936-3c48df42-c541-4c2a-99be-a32bc5f18728.png)
1.0
Requested changes to GeoNetwork - The client is requesting few changes to the new GN 3 RNDT installed in the new SDI. From the client: >Per quanto riguarda l'interfaccia, invece avremmo la necessità di due migliorie che sono: >- Nella pagina di ricerca dei dataset, tra i filtraggi aggiungere un filtro che consente di estrarre i risultati in base ad un range di "Data di pubblicazione" (filtrare tutti i dataset con Data di pubblicazione da gg/mm/aaaa a gg/mm/aaaa) ![image.png](https://images.zenhubusercontent.com/5c9ca89d6607b6793c66ec35/c5674621-bc26-41ed-bf88-93d63505bcf9) >- Nel pdf che è possibile esportare dopo aver selezionato i dataset eliminare i seguenti campi: > - Update frequency > - Record updated > - Risorse > - More onlineinformation >Quindi gli unici campi visibili dovranno esssere: > - Descrizione > - Estensione > - Parole chiave > - Access and useconstraints > - Data type > - Format > - Identificatoreunivoco ![image](https://user-images.githubusercontent.com/1280027/155680936-3c48df42-c541-4c2a-99be-a32bc5f18728.png)
non_process
requested changes to geonetwork the client is requesting few changes to the new gn rndt installed in the new sdi from the client per quanto riguarda l interfaccia invece avremmo la necessità di due migliorie che sono nella pagina di ricerca dei dataset tra i filtraggi aggiungere un filtro che consente di estrarre i risultati in base ad un range di data di pubblicazione filtrare tutti i dataset con data di pubblicazione da gg mm aaaa a gg mm aaaa nel pdf che è possibile esportare dopo aver selezionato i dataset eliminare i seguenti campi update frequency record updated risorse more onlineinformation quindi gli unici campi visibili dovranno esssere descrizione estensione parole chiave access and useconstraints data type format identificatoreunivoco
0
235,552
18,051,853,073
IssuesEvent
2021-09-19 21:55:03
EdwarMalpica/AsignacionOficinas
https://api.github.com/repos/EdwarMalpica/AsignacionOficinas
closed
Implementacion de la configuracion necesaria para la capa de persistencia
documentation enhancement
Codificar la configuracion necesaria para la creacion de la base de datos
1.0
Implementacion de la configuracion necesaria para la capa de persistencia - Codificar la configuracion necesaria para la creacion de la base de datos
non_process
implementacion de la configuracion necesaria para la capa de persistencia codificar la configuracion necesaria para la creacion de la base de datos
0
60,683
7,373,620,840
IssuesEvent
2018-03-13 17:46:58
ParabolInc/action
https://api.github.com/repos/ParabolInc/action
opened
Design: retrospective free→pro upgrade path
design enhancement groom-ready
## Issue - Enhancement Let's make some money. Our bet is that adding a team handbook and retrospective meeting type will be enough differentiated value to drive adoption and conversion. Today, we're largely driving conversion through top-down sales. However, we should always have a experience we are testing to drive bottoms-up conversions. ### Some history When we originally launched Parabol's payment features, Parabol was trial-ware. Teams could try us free for 30 days. They could get another 30 days free if they gave us their credit card number. When the trial expired, their users would lose access to all of their organization's teams until they paid. *What we learned:* our value prop wasn't thick enough to drive self-adoption, so both free usage _and_ conversions were low. *What we didn't learn:* trialing for premium features is bad. In fact, trialing for premium features might still be good. ### Goals & Acceptance Criteria (optional) An designed, lo-fidelity experience where users can: - Get to know all of the features of Parabol and get themselves and their teams hooked, meaning: _(this is how we'll know this design has worked)_ - See clear: - Messages for which features are pro and are subjected to being limited - Calls to action for how to convert - At least one team is running at least one meeting (action/retro) regularly (at least 1/mo) - At the end of the evaluation period, the team converts at least 10% of the time Validation criteria: - The lo-fi design is checked with at least two of our partner/advisors Once the design is completed, the following are created - A list of implementation Issues/stories - A list of supporting marketing workflows (for example, emails to send) ### Implementation considerations #### Trial-ware In addition to `personal` and `pro`, we create a new tier called `pro-trial` (or, extend the `pro` tier to have an expiration date). When a user creates their first org, their org is in the `pro-trial` tier. We assume that it takes a quarter (3 months/3x4-week periods) to convert. After 7 weeks, we give them a banner that their trial is about to run out, but they can extend it for 4 more weeks free if they give us their cc now. After a maximal period of 12 weeks, if they do not convert, we offer them the a workflow to demote their account to the `personal` free tier. We send marketing communications to support conversion. When the account is in the `personal` tier, we disallow access to things such as Retrospective Meetings and access to the Team Handbook. #### Limit-ware All accounts start in the `personal` tier. Some features are limited, just as we limit access to a team's Task archive today. We begin by adding a limit to the number of Retrospective meetings that can be run. For example, we offer a limit of 3 per org. We limit the number of Handbook items that can be created. For example, 12 per org. When one of these limits is exceeded, we decorate the interface with affordances to convert and remove these limits. We send marketing communications to support conversion. - **Estimated effort:** 15 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
1.0
Design: retrospective free→pro upgrade path - ## Issue - Enhancement Let's make some money. Our bet is that adding a team handbook and retrospective meeting type will be enough differentiated value to drive adoption and conversion. Today, we're largely driving conversion through top-down sales. However, we should always have a experience we are testing to drive bottoms-up conversions. ### Some history When we originally launched Parabol's payment features, Parabol was trial-ware. Teams could try us free for 30 days. They could get another 30 days free if they gave us their credit card number. When the trial expired, their users would lose access to all of their organization's teams until they paid. *What we learned:* our value prop wasn't thick enough to drive self-adoption, so both free usage _and_ conversions were low. *What we didn't learn:* trialing for premium features is bad. In fact, trialing for premium features might still be good. ### Goals & Acceptance Criteria (optional) An designed, lo-fidelity experience where users can: - Get to know all of the features of Parabol and get themselves and their teams hooked, meaning: _(this is how we'll know this design has worked)_ - See clear: - Messages for which features are pro and are subjected to being limited - Calls to action for how to convert - At least one team is running at least one meeting (action/retro) regularly (at least 1/mo) - At the end of the evaluation period, the team converts at least 10% of the time Validation criteria: - The lo-fi design is checked with at least two of our partner/advisors Once the design is completed, the following are created - A list of implementation Issues/stories - A list of supporting marketing workflows (for example, emails to send) ### Implementation considerations #### Trial-ware In addition to `personal` and `pro`, we create a new tier called `pro-trial` (or, extend the `pro` tier to have an expiration date). When a user creates their first org, their org is in the `pro-trial` tier. We assume that it takes a quarter (3 months/3x4-week periods) to convert. After 7 weeks, we give them a banner that their trial is about to run out, but they can extend it for 4 more weeks free if they give us their cc now. After a maximal period of 12 weeks, if they do not convert, we offer them the a workflow to demote their account to the `personal` free tier. We send marketing communications to support conversion. When the account is in the `personal` tier, we disallow access to things such as Retrospective Meetings and access to the Team Handbook. #### Limit-ware All accounts start in the `personal` tier. Some features are limited, just as we limit access to a team's Task archive today. We begin by adding a limit to the number of Retrospective meetings that can be run. For example, we offer a limit of 3 per org. We limit the number of Handbook items that can be created. For example, 12 per org. When one of these limits is exceeded, we decorate the interface with affordances to convert and remove these limits. We send marketing communications to support conversion. - **Estimated effort:** 15 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
non_process
design retrospective free→pro upgrade path issue enhancement let s make some money our bet is that adding a team handbook and retrospective meeting type will be enough differentiated value to drive adoption and conversion today we re largely driving conversion through top down sales however we should always have a experience we are testing to drive bottoms up conversions some history when we originally launched parabol s payment features parabol was trial ware teams could try us free for days they could get another days free if they gave us their credit card number when the trial expired their users would lose access to all of their organization s teams until they paid what we learned our value prop wasn t thick enough to drive self adoption so both free usage and conversions were low what we didn t learn trialing for premium features is bad in fact trialing for premium features might still be good goals acceptance criteria optional an designed lo fidelity experience where users can get to know all of the features of parabol and get themselves and their teams hooked meaning this is how we ll know this design has worked see clear messages for which features are pro and are subjected to being limited calls to action for how to convert at least one team is running at least one meeting action retro regularly at least mo at the end of the evaluation period the team converts at least of the time validation criteria the lo fi design is checked with at least two of our partner advisors once the design is completed the following are created a list of implementation issues stories a list of supporting marketing workflows for example emails to send implementation considerations trial ware in addition to personal and pro we create a new tier called pro trial or extend the pro tier to have an expiration date when a user creates their first org their org is in the pro trial tier we assume that it takes a quarter months week periods to convert after weeks we give them a banner that their trial is about to run out but they can extend it for more weeks free if they give us their cc now after a maximal period of weeks if they do not convert we offer them the a workflow to demote their account to the personal free tier we send marketing communications to support conversion when the account is in the personal tier we disallow access to things such as retrospective meetings and access to the team handbook limit ware all accounts start in the personal tier some features are limited just as we limit access to a team s task archive today we begin by adding a limit to the number of retrospective meetings that can be run for example we offer a limit of per org we limit the number of handbook items that can be created for example per org when one of these limits is exceeded we decorate the interface with affordances to convert and remove these limits we send marketing communications to support conversion estimated effort points
0