Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
19,081
| 25,127,606,887
|
IssuesEvent
|
2022-11-09 12:58:17
|
inmanta/inmanta-core
|
https://api.github.com/repos/inmanta/inmanta-core
|
reopened
|
Enable timestamp on the log messages produced by the test cases
|
process tiny
|
In order to debug a test case that times out on a wait condition, it's handy to see the timestamps in the log messages. That way it easier to see in which stage a lot of time is being lost.
Adding the timestamps can be achieved by providing the following options to pytest: `--log-format="%(asctime)s.%(msecs)03d %(levelname)s %(message)s"`.
If `pytest.ini` supports it, add it there.
|
1.0
|
Enable timestamp on the log messages produced by the test cases - In order to debug a test case that times out on a wait condition, it's handy to see the timestamps in the log messages. That way it easier to see in which stage a lot of time is being lost.
Adding the timestamps can be achieved by providing the following options to pytest: `--log-format="%(asctime)s.%(msecs)03d %(levelname)s %(message)s"`.
If `pytest.ini` supports it, add it there.
|
process
|
enable timestamp on the log messages produced by the test cases in order to debug a test case that times out on a wait condition it s handy to see the timestamps in the log messages that way it easier to see in which stage a lot of time is being lost adding the timestamps can be achieved by providing the following options to pytest log format asctime s msecs levelname s message s if pytest ini supports it add it there
| 1
|
139,101
| 12,837,281,236
|
IssuesEvent
|
2020-07-07 15:33:10
|
rust-bio/rust-bio
|
https://api.github.com/repos/rust-bio/rust-bio
|
closed
|
Improve docs: `seq_analysis::gc`
|
documentation
|
* [ ] add at least one example
* [ ] describe time and/or memory complexity
|
1.0
|
Improve docs: `seq_analysis::gc` - * [ ] add at least one example
* [ ] describe time and/or memory complexity
|
non_process
|
improve docs seq analysis gc add at least one example describe time and or memory complexity
| 0
|
20,431
| 27,098,151,954
|
IssuesEvent
|
2023-02-15 05:56:49
|
anobaka/InsideWorld
|
https://api.github.com/repos/anobaka/InsideWorld
|
closed
|
优化文件处理器部分操作交互
|
enhancement file-processor
|
+ [x] 当上方文件数量发生变化时,可能会导致误点击【提取】等按钮,提取按钮现调整至文件名右侧,并增加确认弹窗(支持回车提交)


+ [x] 文件操作按钮可能需要尽可能贴近文件名
+ [x] 根目录变更时显示加载中图标
|
1.0
|
优化文件处理器部分操作交互 - + [x] 当上方文件数量发生变化时,可能会导致误点击【提取】等按钮,提取按钮现调整至文件名右侧,并增加确认弹窗(支持回车提交)


+ [x] 文件操作按钮可能需要尽可能贴近文件名
+ [x] 根目录变更时显示加载中图标
|
process
|
优化文件处理器部分操作交互 当上方文件数量发生变化时,可能会导致误点击【提取】等按钮,提取按钮现调整至文件名右侧,并增加确认弹窗(支持回车提交) 文件操作按钮可能需要尽可能贴近文件名 根目录变更时显示加载中图标
| 1
|
262,590
| 19,822,793,084
|
IssuesEvent
|
2022-01-20 00:44:08
|
vercel/next.js
|
https://api.github.com/repos/vercel/next.js
|
opened
|
Docs: Tailwind scaffolding in Next.js Projects for separation of concerns
|
template: documentation
|
### What is the improvement or update you wish to see?
I would like to see a template on how we should organize scaffolding of CSS files when using tailwind.
### Is there any context that might help us understand?
I know we can use global.css file for global variables, but what is the proper way of organizing CSS files for the sake of separation of concerns, route folders, and/or general organization?
Are there any gotchas/caveats to separating CSS files?
Checked the with-tailwind example, and unfortunately it only provided the simplest example with global.css file(initial setup) and inline modifications.
And thanks! It really has been great getting to know Next.js btw. Still got loads to learn but having fun.
### Does the docs page already exist? Please link to it.
https://github.com/vercel/next.js/tree/canary/examples/with-tailwindcss
|
1.0
|
Docs: Tailwind scaffolding in Next.js Projects for separation of concerns - ### What is the improvement or update you wish to see?
I would like to see a template on how we should organize scaffolding of CSS files when using tailwind.
### Is there any context that might help us understand?
I know we can use global.css file for global variables, but what is the proper way of organizing CSS files for the sake of separation of concerns, route folders, and/or general organization?
Are there any gotchas/caveats to separating CSS files?
Checked the with-tailwind example, and unfortunately it only provided the simplest example with global.css file(initial setup) and inline modifications.
And thanks! It really has been great getting to know Next.js btw. Still got loads to learn but having fun.
### Does the docs page already exist? Please link to it.
https://github.com/vercel/next.js/tree/canary/examples/with-tailwindcss
|
non_process
|
docs tailwind scaffolding in next js projects for separation of concerns what is the improvement or update you wish to see i would like to see a template on how we should organize scaffolding of css files when using tailwind is there any context that might help us understand i know we can use global css file for global variables but what is the proper way of organizing css files for the sake of separation of concerns route folders and or general organization are there any gotchas caveats to separating css files checked the with tailwind example and unfortunately it only provided the simplest example with global css file initial setup and inline modifications and thanks it really has been great getting to know next js btw still got loads to learn but having fun does the docs page already exist please link to it
| 0
|
284,561
| 21,444,711,111
|
IssuesEvent
|
2022-04-25 04:13:04
|
cengage/react-magma
|
https://api.github.com/repos/cengage/react-magma
|
closed
|
Documentation> Checkbox link> "Not Found" error message is appearing upon clicking the link
|
documentation
|
**Describe the bug**
"Not Found" error message is appearing upon clicking the link "Checkbox".
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://react-magma.cengage.com/version/2.5.7/api/radio/
2. Click on "Checkbox" link
3. Verify the error
**Expected behavior**
Appropriate page should appear upon clicking the link "Checkbox".
**Screenshots**
https://somup.com/c3nXXZZRQO
**Desktop:**
- OS: [Win 10]
- Browser [chrome]
- Version [Chrome version: Version 98.0.4758.81 (Official Build) (64-bit)]
|
1.0
|
Documentation> Checkbox link> "Not Found" error message is appearing upon clicking the link - **Describe the bug**
"Not Found" error message is appearing upon clicking the link "Checkbox".
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://react-magma.cengage.com/version/2.5.7/api/radio/
2. Click on "Checkbox" link
3. Verify the error
**Expected behavior**
Appropriate page should appear upon clicking the link "Checkbox".
**Screenshots**
https://somup.com/c3nXXZZRQO
**Desktop:**
- OS: [Win 10]
- Browser [chrome]
- Version [Chrome version: Version 98.0.4758.81 (Official Build) (64-bit)]
|
non_process
|
documentation checkbox link not found error message is appearing upon clicking the link describe the bug not found error message is appearing upon clicking the link checkbox to reproduce steps to reproduce the behavior go to click on checkbox link verify the error expected behavior appropriate page should appear upon clicking the link checkbox screenshots desktop os browser version
| 0
|
10,346
| 13,172,404,656
|
IssuesEvent
|
2020-08-11 18:22:41
|
openenclave/openenclave
|
https://api.github.com/repos/openenclave/openenclave
|
closed
|
Unable to distinguish API versions
|
attestation bug engineering process triaged
|
The master branch saw a change to the host-side quote verification library: `oe_verify_remote_report` now expects 5 arguments instead of 3 arguments in v0.7.0. We now check for `#if defined(OE_CLAIM_ID_VERSION)`, which isn't defined in v0.7.0, to tell the difference.
It would be nice if the version number were updated when the API changes but as far as I can tell, `OE_API_VERSION` is always the same (2). Perhaps it makes sense to add separate versioning for the host-side quote verification library too?
|
1.0
|
Unable to distinguish API versions - The master branch saw a change to the host-side quote verification library: `oe_verify_remote_report` now expects 5 arguments instead of 3 arguments in v0.7.0. We now check for `#if defined(OE_CLAIM_ID_VERSION)`, which isn't defined in v0.7.0, to tell the difference.
It would be nice if the version number were updated when the API changes but as far as I can tell, `OE_API_VERSION` is always the same (2). Perhaps it makes sense to add separate versioning for the host-side quote verification library too?
|
process
|
unable to distinguish api versions the master branch saw a change to the host side quote verification library oe verify remote report now expects arguments instead of arguments in we now check for if defined oe claim id version which isn t defined in to tell the difference it would be nice if the version number were updated when the api changes but as far as i can tell oe api version is always the same perhaps it makes sense to add separate versioning for the host side quote verification library too
| 1
|
10,265
| 13,112,556,253
|
IssuesEvent
|
2020-08-05 02:34:22
|
kubeflow/manifests
|
https://api.github.com/repos/kubeflow/manifests
|
closed
|
Owners file for kubeflow/metadata needs to be updated
|
area/metadata kind/process lifecycle/stale priority/p2
|
The OWNERs file for metadata needs to be updated it looks like the current set of approvers is no longer up to date and responsive.
See #928
|
1.0
|
Owners file for kubeflow/metadata needs to be updated - The OWNERs file for metadata needs to be updated it looks like the current set of approvers is no longer up to date and responsive.
See #928
|
process
|
owners file for kubeflow metadata needs to be updated the owners file for metadata needs to be updated it looks like the current set of approvers is no longer up to date and responsive see
| 1
|
10,826
| 13,609,564,470
|
IssuesEvent
|
2020-09-23 05:36:14
|
GoogleCloudPlatform/getting-started-java
|
https://api.github.com/repos/GoogleCloudPlatform/getting-started-java
|
closed
|
Dependency Dashboard
|
type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-compiler-plugin-3.x -->chore(deps): update dependency org.apache.maven.plugins:maven-compiler-plugin to v3.8.1
- [ ] <!-- rebase-branch=renovate/javax.servlet-javax.servlet-api-4.x -->chore(deps): update dependency javax.servlet:javax.servlet-api to v4
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/org.apache.maven.plugins-maven-compiler-plugin-3.x -->chore(deps): update dependency org.apache.maven.plugins:maven-compiler-plugin to v3.8.1
- [ ] <!-- rebase-branch=renovate/javax.servlet-javax.servlet-api-4.x -->chore(deps): update dependency javax.servlet:javax.servlet-api to v4
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency org apache maven plugins maven compiler plugin to chore deps update dependency javax servlet javax servlet api to check this box to trigger a request for renovate to run again on this repository
| 1
|
355,314
| 25,175,898,206
|
IssuesEvent
|
2022-11-11 09:13:58
|
Devanshshah1309/pe
|
https://api.github.com/repos/Devanshshah1309/pe
|
opened
|
Link to FAQ Section not working
|
severity.Low type.DocumentationBug
|
The link to the FAQ section here is broken (or there's no link at all).

This prevents me from being able to go to the FAQ section if I'm only interested in that - it is not purely a cosmetic bug. It hinders my reading. Hence, low severity.
<!--session: 1668147090315-757e20e0-6c8f-46c9-a0a7-2575a40077ea-->
<!--Version: Web v3.4.4-->
|
1.0
|
Link to FAQ Section not working - The link to the FAQ section here is broken (or there's no link at all).

This prevents me from being able to go to the FAQ section if I'm only interested in that - it is not purely a cosmetic bug. It hinders my reading. Hence, low severity.
<!--session: 1668147090315-757e20e0-6c8f-46c9-a0a7-2575a40077ea-->
<!--Version: Web v3.4.4-->
|
non_process
|
link to faq section not working the link to the faq section here is broken or there s no link at all this prevents me from being able to go to the faq section if i m only interested in that it is not purely a cosmetic bug it hinders my reading hence low severity
| 0
|
11,531
| 14,403,749,597
|
IssuesEvent
|
2020-12-03 16:25:55
|
LOVDnl/LOVD3
|
https://api.github.com/repos/LOVDnl/LOVD3
|
opened
|
Curators and up can create variant submissions without classification.
|
bug minor cat: submission process
|
Curators and up can create variant submissions without classification, but only when a gene is selected. VOG entries still require at least the reported classification, while VOT entries do not. Submitters are always required to fill in the reported classification.
|
1.0
|
Curators and up can create variant submissions without classification. - Curators and up can create variant submissions without classification, but only when a gene is selected. VOG entries still require at least the reported classification, while VOT entries do not. Submitters are always required to fill in the reported classification.
|
process
|
curators and up can create variant submissions without classification curators and up can create variant submissions without classification but only when a gene is selected vog entries still require at least the reported classification while vot entries do not submitters are always required to fill in the reported classification
| 1
|
9,175
| 12,226,438,488
|
IssuesEvent
|
2020-05-03 10:55:40
|
labnote-ant/labnote
|
https://api.github.com/repos/labnote-ant/labnote
|
closed
|
Add description box
|
chemical-view process-view
|
In chemical view or process view, it needs a description box where the user can input details of chemicals or process.
|
1.0
|
Add description box - In chemical view or process view, it needs a description box where the user can input details of chemicals or process.
|
process
|
add description box in chemical view or process view it needs a description box where the user can input details of chemicals or process
| 1
|
14,960
| 18,445,033,755
|
IssuesEvent
|
2021-10-15 00:01:02
|
cloudfoundry/cf-k8s-api
|
https://api.github.com/repos/cloudfoundry/cf-k8s-api
|
closed
|
[Feature]: API Client can List Processes for an App via `GET /v3/apps/:guid/processes`
|
Processes
|
### Blockers/Dependencies
_No response_
### Background
**As a** client of the API Shim
**I want** to be able to list all Processes for my App
**So that** I can discover information about my Processes
The CF CLI hits this endpoint during `cf push` so that it can get a list of Process guids to make future API calls (such as fetching stats for a Process).
### Acceptance Criteria
## Scenarios
### Happy Path (App with Processes)
**GIVEN** I have a CFApp and CFProcesses are associated with it
**WHEN** I make the following API request:
```bash
curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \
-X GET \
-H "Authorization: bearer <placeholder-bearer-token>"
```
**THEN** I see a response that reflects the information on the CFProcesses
```json
HTTP/1.1 200 OK
Content-Type: application/json
{
"pagination": {
"total_results": 2,
"total_pages": 1,
"first": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=1&per_page=2"
},
"last": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2"
},
"next": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2"
},
"previous": null
},
"resources": [
{
"guid": "6a901b7c-9417-4dc1-8189-d3234aa0ab82",
"type": "web",
"command": "[PRIVATE DATA HIDDEN IN LISTS]",
"instances": 5,
"memory_in_mb": 256,
"disk_in_mb": 1024,
"health_check": {
"type": "port",
"data": {
"timeout": null,
"invocation_timeout": null
}
},
"relationships": {
"app": {
"data": {
"guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
}
}
},
"metadata": {
"labels": {},
"annotations": {}
},
"created_at": "2016-03-23T18:48:22Z",
"updated_at": "2016-03-23T18:48:42Z",
"links": {
"self": {
"href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82"
},
"scale": {
"href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/actions/scale",
"method": "POST"
},
"app": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
},
"space": {
"href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576"
},
"stats": {
"href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/stats"
}
}
},
{
"guid": "3fccacd9-4b02-4b96-8d02-8e865865e9eb",
"type": "worker",
"command": "[PRIVATE DATA HIDDEN IN LISTS]",
"instances": 1,
"memory_in_mb": 256,
"disk_in_mb": 1024,
"health_check": {
"type": "process",
"data": {
"timeout": null,
"invocation_timeout": null
}
},
"relationships": {
"app": {
"data": {
"guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
}
},
},
"metadata": {
"labels": {},
"annotations": {}
},
"created_at": "2016-03-23T18:48:22Z",
"updated_at": "2016-03-23T18:48:42Z",
"links": {
"self": {
"href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb"
},
"scale": {
"href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/actions/scale",
"method": "POST"
},
"app": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
},
"space": {
"href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576"
},
"stats": {
"href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/stats"
}
}
}
]
}
```
*Note*: we're omitting the `revision` key entirely. The `metadata` key will always contain empty hashes, as in other stories.
---
### App with No Processes
**GIVEN** I have a CFApp and **no** CFProcesses are associated with it
**WHEN** I make the following API request:
```bash
curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \
-X GET \
-H "Authorization: bearer <placeholder-bearer-token>"
```
**THEN** I get back a response with an empty resources array
```json
HTTP/1.1 200 OK
Content-Type: application/json
{
"pagination": {
"total_results": 0,
"total_pages": 1,
"first": {
"href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50"
},
"last": {
"href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50"
},
"next": null,
"previous": null
},
"resources": [
]
}
```
---
### App doesn't exist
**GIVEN** I have do not have a CFApp with the guid below
**WHEN** I make the following API request:
```bash
curl "https://api-shim.example.org/v3/apps/<non-existant-app-guid>/processes" \
-X GET \
-H "Authorization: bearer <placeholder-bearer-token>"
```
**THEN** I get back a 404 response
```json
HTTP/1.1 404 Not Found
{
"errors": [
{
"detail": "App not found",
"title": "CF-ResourceNotFound",
"code": 10010
}
]
}
```
### Dev Notes
* V3 API Docs: https://v3-apidocs.cloudfoundry.org/version/3.107.0/index.html#list-processes
* Pagination: Always return all results for now (as we have been doing on other stories)
* Query parameters: Ignore filter parameters for now. We can add them in later
* Be sure to add the necessary RBAC annotations in the new Process repository. Otherwise the app will error when deployed to a real cluster (but will work locally)
|
1.0
|
[Feature]: API Client can List Processes for an App via `GET /v3/apps/:guid/processes` - ### Blockers/Dependencies
_No response_
### Background
**As a** client of the API Shim
**I want** to be able to list all Processes for my App
**So that** I can discover information about my Processes
The CF CLI hits this endpoint during `cf push` so that it can get a list of Process guids to make future API calls (such as fetching stats for a Process).
### Acceptance Criteria
## Scenarios
### Happy Path (App with Processes)
**GIVEN** I have a CFApp and CFProcesses are associated with it
**WHEN** I make the following API request:
```bash
curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \
-X GET \
-H "Authorization: bearer <placeholder-bearer-token>"
```
**THEN** I see a response that reflects the information on the CFProcesses
```json
HTTP/1.1 200 OK
Content-Type: application/json
{
"pagination": {
"total_results": 2,
"total_pages": 1,
"first": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=1&per_page=2"
},
"last": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2"
},
"next": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5/processes?page=2&per_page=2"
},
"previous": null
},
"resources": [
{
"guid": "6a901b7c-9417-4dc1-8189-d3234aa0ab82",
"type": "web",
"command": "[PRIVATE DATA HIDDEN IN LISTS]",
"instances": 5,
"memory_in_mb": 256,
"disk_in_mb": 1024,
"health_check": {
"type": "port",
"data": {
"timeout": null,
"invocation_timeout": null
}
},
"relationships": {
"app": {
"data": {
"guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
}
}
},
"metadata": {
"labels": {},
"annotations": {}
},
"created_at": "2016-03-23T18:48:22Z",
"updated_at": "2016-03-23T18:48:42Z",
"links": {
"self": {
"href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82"
},
"scale": {
"href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/actions/scale",
"method": "POST"
},
"app": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
},
"space": {
"href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576"
},
"stats": {
"href": "https://api-shim.example.org/v3/processes/6a901b7c-9417-4dc1-8189-d3234aa0ab82/stats"
}
}
},
{
"guid": "3fccacd9-4b02-4b96-8d02-8e865865e9eb",
"type": "worker",
"command": "[PRIVATE DATA HIDDEN IN LISTS]",
"instances": 1,
"memory_in_mb": 256,
"disk_in_mb": 1024,
"health_check": {
"type": "process",
"data": {
"timeout": null,
"invocation_timeout": null
}
},
"relationships": {
"app": {
"data": {
"guid": "ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
}
},
},
"metadata": {
"labels": {},
"annotations": {}
},
"created_at": "2016-03-23T18:48:22Z",
"updated_at": "2016-03-23T18:48:42Z",
"links": {
"self": {
"href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb"
},
"scale": {
"href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/actions/scale",
"method": "POST"
},
"app": {
"href": "https://api-shim.example.org/v3/apps/ccc25a0f-c8f4-4b39-9f1b-de9f328d0ee5"
},
"space": {
"href": "https://api-shim.example.org/v3/spaces/2f35885d-0c9d-4423-83ad-fd05066f8576"
},
"stats": {
"href": "https://api-shim.example.org/v3/processes/3fccacd9-4b02-4b96-8d02-8e865865e9eb/stats"
}
}
}
]
}
```
*Note*: we're omitting the `revision` key entirely. The `metadata` key will always contain empty hashes, as in other stories.
---
### App with No Processes
**GIVEN** I have a CFApp and **no** CFProcesses are associated with it
**WHEN** I make the following API request:
```bash
curl "https://api-shim.example.org/v3/apps/<app-guid>/processes" \
-X GET \
-H "Authorization: bearer <placeholder-bearer-token>"
```
**THEN** I get back a response with an empty resources array
```json
HTTP/1.1 200 OK
Content-Type: application/json
{
"pagination": {
"total_results": 0,
"total_pages": 1,
"first": {
"href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50"
},
"last": {
"href": "https://api.bramble-quester.capi.land/v3/apps/ea2501a0-a579-40a8-8cc9-2da76cb1d72d/processes?page=1&per_page=50"
},
"next": null,
"previous": null
},
"resources": [
]
}
```
---
### App doesn't exist
**GIVEN** I have do not have a CFApp with the guid below
**WHEN** I make the following API request:
```bash
curl "https://api-shim.example.org/v3/apps/<non-existant-app-guid>/processes" \
-X GET \
-H "Authorization: bearer <placeholder-bearer-token>"
```
**THEN** I get back a 404 response
```json
HTTP/1.1 404 Not Found
{
"errors": [
{
"detail": "App not found",
"title": "CF-ResourceNotFound",
"code": 10010
}
]
}
```
### Dev Notes
* V3 API Docs: https://v3-apidocs.cloudfoundry.org/version/3.107.0/index.html#list-processes
* Pagination: Always return all results for now (as we have been doing on other stories)
* Query parameters: Ignore filter parameters for now. We can add them in later
* Be sure to add the necessary RBAC annotations in the new Process repository. Otherwise the app will error when deployed to a real cluster (but will work locally)
|
process
|
api client can list processes for an app via get apps guid processes blockers dependencies no response background as a client of the api shim i want to be able to list all processes for my app so that i can discover information about my processes the cf cli hits this endpoint during cf push so that it can get a list of process guids to make future api calls such as fetching stats for a process acceptance criteria scenarios happy path app with processes given i have a cfapp and cfprocesses are associated with it when i make the following api request bash curl x get h authorization bearer then i see a response that reflects the information on the cfprocesses json http ok content type application json pagination total results total pages first href last href next href previous null resources guid type web command instances memory in mb disk in mb health check type port data timeout null invocation timeout null relationships app data guid metadata labels annotations created at updated at links self href scale href method post app href space href stats href guid type worker command instances memory in mb disk in mb health check type process data timeout null invocation timeout null relationships app data guid metadata labels annotations created at updated at links self href scale href method post app href space href stats href note we re omitting the revision key entirely the metadata key will always contain empty hashes as in other stories app with no processes given i have a cfapp and no cfprocesses are associated with it when i make the following api request bash curl x get h authorization bearer then i get back a response with an empty resources array json http ok content type application json pagination total results total pages first href last href next null previous null resources app doesn t exist given i have do not have a cfapp with the guid below when i make the following api request bash curl x get h authorization bearer then i get back a response json http not found errors detail app not found title cf resourcenotfound code dev notes api docs pagination always return all results for now as we have been doing on other stories query parameters ignore filter parameters for now we can add them in later be sure to add the necessary rbac annotations in the new process repository otherwise the app will error when deployed to a real cluster but will work locally
| 1
|
271,136
| 29,299,168,536
|
IssuesEvent
|
2023-05-25 01:06:58
|
hshivhare67/kernel_v4.19.72_CVE-2023-0461
|
https://api.github.com/repos/hshivhare67/kernel_v4.19.72_CVE-2023-0461
|
opened
|
CVE-2023-33203 (Medium) detected in linuxlinux-4.19.282
|
Mend: dependency security vulnerability
|
## CVE-2023-33203 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72_CVE-2023-0461/commit/20984407a51d9f25ee9889e4b1304489f480d36e">20984407a51d9f25ee9889e4b1304489f480d36e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before 6.2.9 has a race condition and resultant use-after-free in drivers/net/ethernet/qualcomm/emac/emac.c if a physically proximate attacker unplugs an emac based device.
<p>Publish Date: 2023-05-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-33203>CVE-2023-33203</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-33203">https://www.linuxkernelcves.com/cves/CVE-2023-33203</a></p>
<p>Release Date: 2023-05-18</p>
<p>Fix Resolution: v4.14.312,v4.19.280,v5.4.240,v5.10.177,v5.15.105,v6.1.22,v6.2.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-33203 (Medium) detected in linuxlinux-4.19.282 - ## CVE-2023-33203 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72_CVE-2023-0461/commit/20984407a51d9f25ee9889e4b1304489f480d36e">20984407a51d9f25ee9889e4b1304489f480d36e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before 6.2.9 has a race condition and resultant use-after-free in drivers/net/ethernet/qualcomm/emac/emac.c if a physically proximate attacker unplugs an emac based device.
<p>Publish Date: 2023-05-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-33203>CVE-2023-33203</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-33203">https://www.linuxkernelcves.com/cves/CVE-2023-33203</a></p>
<p>Release Date: 2023-05-18</p>
<p>Fix Resolution: v4.14.312,v4.19.280,v5.4.240,v5.10.177,v5.15.105,v6.1.22,v6.2.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net ethernet qualcomm emac emac c drivers net ethernet qualcomm emac emac c vulnerability details the linux kernel before has a race condition and resultant use after free in drivers net ethernet qualcomm emac emac c if a physically proximate attacker unplugs an emac based device publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
2,858
| 5,824,282,508
|
IssuesEvent
|
2017-05-07 11:31:07
|
QCoDeS/Qcodes
|
https://api.github.com/repos/QCoDeS/Qcodes
|
closed
|
Leaking sockets?
|
bug mulitprocessing p2
|
### Steps to reproduce
1. Run code for some time on windows including possible shutdown and restart of notebook and instrument communication.
### Expected behaviour
Things should keep working
### Actual behaviour
Notebook fails with socket related connection issues from Tornado and ZMQ.
```
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\multikernelmanager.py", line 33, in wrapped
r = method(*args, **kwargs)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\ioloop\manager.py", line 33, in wrapped
socket = f(self, *args, **kwargs)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\connect.py", line 492, in connect_shell
return self._create_connected_socket('shell', identity=identity)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\connect.py", line 476, in _create_connected_socket
sock = self.context.socket(socket_type)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\z
mq\sugar\context.py", line 146, in socket
s = self._socket_class(self, socket_type, **kwargs)
File "zmq\backend\cython\socket.pyx", line 285, in zmq.backend.cython.sock
et.Socket.__cinit__ (zmq\backend\cython\socket.c:3861)
zmq.error.ZMQError: No buffer space available
OSError: [WinError 10055] An operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full
```
Regular internet (browsing etc) is flaky too
We suspect that sockets are leaked and not cleaned up. This may be in the VISA driver of
network instruments such as ZNMB20 VNA
### System
**operating system**
Windows
**qcodes branch**
master
**qcodes commit**
?
|
1.0
|
Leaking sockets? -
### Steps to reproduce
1. Run code for some time on windows including possible shutdown and restart of notebook and instrument communication.
### Expected behaviour
Things should keep working
### Actual behaviour
Notebook fails with socket related connection issues from Tornado and ZMQ.
```
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\multikernelmanager.py", line 33, in wrapped
r = method(*args, **kwargs)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\ioloop\manager.py", line 33, in wrapped
socket = f(self, *args, **kwargs)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\connect.py", line 492, in connect_shell
return self._create_connected_socket('shell', identity=identity)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\j
upyter_client\connect.py", line 476, in _create_connected_socket
sock = self.context.socket(socket_type)
File "c:\users\triton2acq\anaconda3\envs\qcodes-master\lib\site-packages\z
mq\sugar\context.py", line 146, in socket
s = self._socket_class(self, socket_type, **kwargs)
File "zmq\backend\cython\socket.pyx", line 285, in zmq.backend.cython.sock
et.Socket.__cinit__ (zmq\backend\cython\socket.c:3861)
zmq.error.ZMQError: No buffer space available
OSError: [WinError 10055] An operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full
```
Regular internet (browsing etc) is flaky too
We suspect that sockets are leaked and not cleaned up. This may be in the VISA driver of
network instruments such as ZNMB20 VNA
### System
**operating system**
Windows
**qcodes branch**
master
**qcodes commit**
?
|
process
|
leaking sockets steps to reproduce run code for some time on windows including possible shutdown and restart of notebook and instrument communication expected behaviour things should keep working actual behaviour notebook fails with socket related connection issues from tornado and zmq file c users envs qcodes master lib site packages j upyter client multikernelmanager py line in wrapped r method args kwargs file c users envs qcodes master lib site packages j upyter client ioloop manager py line in wrapped socket f self args kwargs file c users envs qcodes master lib site packages j upyter client connect py line in connect shell return self create connected socket shell identity identity file c users envs qcodes master lib site packages j upyter client connect py line in create connected socket sock self context socket socket type file c users envs qcodes master lib site packages z mq sugar context py line in socket s self socket class self socket type kwargs file zmq backend cython socket pyx line in zmq backend cython sock et socket cinit zmq backend cython socket c zmq error zmqerror no buffer space available oserror an operation on a socket could not be performed because the system lacked sufficient buffer space or because a queue was full regular internet browsing etc is flaky too we suspect that sockets are leaked and not cleaned up this may be in the visa driver of network instruments such as vna system operating system windows qcodes branch master qcodes commit
| 1
|
117,932
| 25,216,777,579
|
IssuesEvent
|
2022-11-14 09:45:05
|
appsmithorg/appsmith
|
https://api.github.com/repos/appsmithorg/appsmith
|
closed
|
[Feature][Custom JS Lib Epic] Delete installed JS package on demand
|
JS Evaluation Task FE Coders Pod
|
## Summary
Remove installed package from application when user clicks on the delete JS package button.
|
1.0
|
[Feature][Custom JS Lib Epic] Delete installed JS package on demand - ## Summary
Remove installed package from application when user clicks on the delete JS package button.
|
non_process
|
delete installed js package on demand summary remove installed package from application when user clicks on the delete js package button
| 0
|
160,013
| 25,095,864,333
|
IssuesEvent
|
2022-11-08 10:07:13
|
metacpan/metacpan-web
|
https://api.github.com/repos/metacpan/metacpan-web
|
closed
|
multiple levels of indentation not handled
|
type:Bug design-2022-followup
|
Multiple levels of `=over 4` are not visible in the site redesign -- everything appears at the same level, so it's impossible to tell what content is meant to be nested.
example: https://metacpan.org/pod/Net::IDN::Encode -- there is a list of functions, and inside the function description is a list of options, but the options are shown at the same level as the functions themselves.
|
1.0
|
multiple levels of indentation not handled - Multiple levels of `=over 4` are not visible in the site redesign -- everything appears at the same level, so it's impossible to tell what content is meant to be nested.
example: https://metacpan.org/pod/Net::IDN::Encode -- there is a list of functions, and inside the function description is a list of options, but the options are shown at the same level as the functions themselves.
|
non_process
|
multiple levels of indentation not handled multiple levels of over are not visible in the site redesign everything appears at the same level so it s impossible to tell what content is meant to be nested example there is a list of functions and inside the function description is a list of options but the options are shown at the same level as the functions themselves
| 0
|
350,710
| 31,931,967,598
|
IssuesEvent
|
2023-09-19 08:03:21
|
Convergence-Project/step-backend
|
https://api.github.com/repos/Convergence-Project/step-backend
|
opened
|
[2주차] (문제집) 게시판 좋아요 기능 구현
|
🎯test ✨feature
|
✏️Description
-
작업사항을 입력해주세요
✅TODO
-
- [ ] 테스트 코드 작성
- [ ] 컨트롤러 프론트 연동
🐾ETC
-
|
1.0
|
[2주차] (문제집) 게시판 좋아요 기능 구현 - ✏️Description
-
작업사항을 입력해주세요
✅TODO
-
- [ ] 테스트 코드 작성
- [ ] 컨트롤러 프론트 연동
🐾ETC
-
|
non_process
|
문제집 게시판 좋아요 기능 구현 ✏️description 작업사항을 입력해주세요 ✅todo 테스트 코드 작성 컨트롤러 프론트 연동 🐾etc
| 0
|
18,485
| 24,550,797,339
|
IssuesEvent
|
2022-10-12 12:27:50
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] 'View consent' and 'View website' buttons are not getting displayed on the study overview screen
|
Bug P1 iOS Process: Fixed Process: Tested dev
|
Steps:
1. Sign up or sign in to the app
2. Click on any study to enroll
3. Click on the participate
4. I enrollment flow, click on the 'Cancel' button
5. Click on 'End task' or 'Click on 'Discard result' and observe
AR: 'View consent' and 'View website' buttons are not getting displayed on the study overview screen
ER: ''View consent' and 'View website' buttons should get displayed on the study overview screen
|
2.0
|
[iOS] 'View consent' and 'View website' buttons are not getting displayed on the study overview screen - Steps:
1. Sign up or sign in to the app
2. Click on any study to enroll
3. Click on the participate
4. I enrollment flow, click on the 'Cancel' button
5. Click on 'End task' or 'Click on 'Discard result' and observe
AR: 'View consent' and 'View website' buttons are not getting displayed on the study overview screen
ER: ''View consent' and 'View website' buttons should get displayed on the study overview screen
|
process
|
view consent and view website buttons are not getting displayed on the study overview screen steps sign up or sign in to the app click on any study to enroll click on the participate i enrollment flow click on the cancel button click on end task or click on discard result and observe ar view consent and view website buttons are not getting displayed on the study overview screen er view consent and view website buttons should get displayed on the study overview screen
| 1
|
14,604
| 17,703,628,988
|
IssuesEvent
|
2021-08-25 03:26:03
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - basisOfRecord
|
Term - change Class - Record-level non-normative Process - complete
|
## Term change
* Submitter: John Wieczorek
* Efficacy Justification (why is this change necessary?): completeness
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): Result of recent public review
* Stability Justification (what concerns are there that this might affect existing implementations?): An addition, no effect on stability except to promote standardization on a ratified term.
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Current Term definition: https://dwc.tdwg.org/list/#dwc_basisOfRecord
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): basisOfRecord
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): None
* Definition of the term (normative): The specific nature of the data record.
* Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use the standard label of one of the Darwin Core classes.
* Examples (not normative): `PreservedSpecimen`, `FossilSpecimen`, `LivingSpecimen`, `MaterialSample`, `Event`, `HumanObservation`, `MachineObservation`, `Taxon`, `Occurrence`, **`MaterialCitation`**
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/basisOfRecord-2017-10-06**
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/RecordBasis
This is to accommodate the addition of the MaterialCitation class.
|
1.0
|
Change term - basisOfRecord - ## Term change
* Submitter: John Wieczorek
* Efficacy Justification (why is this change necessary?): completeness
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): Result of recent public review
* Stability Justification (what concerns are there that this might affect existing implementations?): An addition, no effect on stability except to promote standardization on a ratified term.
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Current Term definition: https://dwc.tdwg.org/list/#dwc_basisOfRecord
Proposed attributes of the new term:
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): basisOfRecord
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): None
* Definition of the term (normative): The specific nature of the data record.
* Usage comments (recommendations regarding content, etc., not normative): Recommended best practice is to use the standard label of one of the Darwin Core classes.
* Examples (not normative): `PreservedSpecimen`, `FossilSpecimen`, `LivingSpecimen`, `MaterialSample`, `Event`, `HumanObservation`, `MachineObservation`, `Taxon`, `Occurrence`, **`MaterialCitation`**
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/basisOfRecord-2017-10-06**
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/RecordBasis
This is to accommodate the addition of the MaterialCitation class.
|
process
|
change term basisofrecord term change submitter john wieczorek efficacy justification why is this change necessary completeness demand justification if the change is semantic in nature name at least two organizations that independently need this term result of recent public review stability justification what concerns are there that this might affect existing implementations an addition no effect on stability except to promote standardization on a ratified term implications for dwciri namespace does this change affect a dwciri term version none current term definition proposed attributes of the new term term name in lowercamelcase for properties uppercamelcase for classes basisofrecord organized in class e g occurrence event location taxon none definition of the term normative the specific nature of the data record usage comments recommendations regarding content etc not normative recommended best practice is to use the standard label of one of the darwin core classes examples not normative preservedspecimen fossilspecimen livingspecimen materialsample event humanobservation machineobservation taxon occurrence materialcitation refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative abcd xpath of the equivalent term in abcd or efg not normative datasets dataset units unit recordbasis this is to accommodate the addition of the materialcitation class
| 1
|
438,189
| 12,623,665,407
|
IssuesEvent
|
2020-06-14 00:30:23
|
hack4impact-uiuc/kids-save-ocean
|
https://api.github.com/repos/hack4impact-uiuc/kids-save-ocean
|
closed
|
Remove fake links
|
high priority
|
Feed:
- Saved projects
- My projects
- Updates
- Followers
- Following
Navbar:
- Notifications
- Current project
Also remove all hardcoded data and comment out features projects from homepage and feed
|
1.0
|
Remove fake links - Feed:
- Saved projects
- My projects
- Updates
- Followers
- Following
Navbar:
- Notifications
- Current project
Also remove all hardcoded data and comment out features projects from homepage and feed
|
non_process
|
remove fake links feed saved projects my projects updates followers following navbar notifications current project also remove all hardcoded data and comment out features projects from homepage and feed
| 0
|
336,706
| 10,195,758,226
|
IssuesEvent
|
2019-08-12 18:57:07
|
jenkins-x/jx
|
https://api.github.com/repos/jenkins-x/jx
|
opened
|
Add kaniko image version to version stream
|
area/tekton area/versions kind/enhancement priority/important-soon
|
We hardcode the default Kaniko image and version in the code currently. That's kinda silly when we've got this whole version stream thing here. =) So let's add a Kaniko version to the version stream, and then update the logic in the CRD translation to use that.
|
1.0
|
Add kaniko image version to version stream - We hardcode the default Kaniko image and version in the code currently. That's kinda silly when we've got this whole version stream thing here. =) So let's add a Kaniko version to the version stream, and then update the logic in the CRD translation to use that.
|
non_process
|
add kaniko image version to version stream we hardcode the default kaniko image and version in the code currently that s kinda silly when we ve got this whole version stream thing here so let s add a kaniko version to the version stream and then update the logic in the crd translation to use that
| 0
|
7,647
| 10,738,585,256
|
IssuesEvent
|
2019-10-29 15:00:40
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
opened
|
Move USAJOBS data pull from Apply button to Next Steps "Continue"
|
Apply Process State Dept.
|
Who:
What:
Why:
Acceptance Criteria:
- Currently the USAJOBS one profile data is pulled for a student when they select "Apply"
- Change the data pull to when they click "Continue" on the Next Steps page
|
1.0
|
Move USAJOBS data pull from Apply button to Next Steps "Continue" - Who:
What:
Why:
Acceptance Criteria:
- Currently the USAJOBS one profile data is pulled for a student when they select "Apply"
- Change the data pull to when they click "Continue" on the Next Steps page
|
process
|
move usajobs data pull from apply button to next steps continue who what why acceptance criteria currently the usajobs one profile data is pulled for a student when they select apply change the data pull to when they click continue on the next steps page
| 1
|
129,603
| 12,414,793,498
|
IssuesEvent
|
2020-05-22 15:07:38
|
alpheios-project/alpheios-core
|
https://api.github.com/repos/alpheios-project/alpheios-core
|
closed
|
selective enabling of Alpheios on components
|
components documentation
|
for alpheios-project/components#129 we disabled alpheios on the panel and popup . sometimes we want to be able to enable it selectively. This requires some thought about the best way to do it but, for example, the ge'ez parser provides short definitions for its words in Latin. We should able to enable Alpheios on those definitions in the popup.
One way to do this might be to look for the language code on the text that is displayed in a component and compare that to available languages to determine if Alpheios can be activated. But we might need even finer grained control of that.
Probably an issue for after refactoring of component state and data.
|
1.0
|
selective enabling of Alpheios on components - for alpheios-project/components#129 we disabled alpheios on the panel and popup . sometimes we want to be able to enable it selectively. This requires some thought about the best way to do it but, for example, the ge'ez parser provides short definitions for its words in Latin. We should able to enable Alpheios on those definitions in the popup.
One way to do this might be to look for the language code on the text that is displayed in a component and compare that to available languages to determine if Alpheios can be activated. But we might need even finer grained control of that.
Probably an issue for after refactoring of component state and data.
|
non_process
|
selective enabling of alpheios on components for alpheios project components we disabled alpheios on the panel and popup sometimes we want to be able to enable it selectively this requires some thought about the best way to do it but for example the ge ez parser provides short definitions for its words in latin we should able to enable alpheios on those definitions in the popup one way to do this might be to look for the language code on the text that is displayed in a component and compare that to available languages to determine if alpheios can be activated but we might need even finer grained control of that probably an issue for after refactoring of component state and data
| 0
|
1,659
| 4,288,680,282
|
IssuesEvent
|
2016-07-17 16:31:09
|
log2timeline/plaso
|
https://api.github.com/repos/log2timeline/plaso
|
closed
|
Preprocessor not working for Windows
|
bug preprocessing
|
Preprocessor not working for Windows
```
2015-12-26 20:53:12,946 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: sysregistry to /WINDOWS/system32/config
2015-12-26 20:53:12,951 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: systemroot to /WINDOWS
2015-12-26 20:53:12,955 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: windir to /WINDOWS
2015-12-26 20:53:12,985 [INFO] (MainProcess) PID:4027 <extraction_frontend> Parser filter expression changed to: win7
```
Should be:
```
2015-12-26 20:53:27,192 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: sysregistry to \WINDOWS\system32\config
2015-12-26 20:53:27,196 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: systemroot to \WINDOWS
2015-12-26 20:53:27,200 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: windir to \WINDOWS
2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: code_page to cp1252
2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: hostname to TEST
2015-12-26 20:53:27,521 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: programfiles to \Program Files
```
* [x] fix issue
* ~~introduces a complication for preg and requires http://codereview.appspot.com/284880043/~~
* ~~https://codereview.appspot.com/276600043/~~
|
1.0
|
Preprocessor not working for Windows - Preprocessor not working for Windows
```
2015-12-26 20:53:12,946 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: sysregistry to /WINDOWS/system32/config
2015-12-26 20:53:12,951 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: systemroot to /WINDOWS
2015-12-26 20:53:12,955 [INFO] (MainProcess) PID:4027 <interface> [PreProcess] Set attribute: windir to /WINDOWS
2015-12-26 20:53:12,985 [INFO] (MainProcess) PID:4027 <extraction_frontend> Parser filter expression changed to: win7
```
Should be:
```
2015-12-26 20:53:27,192 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: sysregistry to \WINDOWS\system32\config
2015-12-26 20:53:27,196 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: systemroot to \WINDOWS
2015-12-26 20:53:27,200 [INFO] (MainProcess) PID:4128 <interface> [PreProcess] Set attribute: windir to \WINDOWS
2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: code_page to cp1252
2015-12-26 20:53:27,338 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: hostname to TEST
2015-12-26 20:53:27,521 [INFO] (MainProcess) PID:4128 <windows> [PreProcess] Set attribute: programfiles to \Program Files
```
* [x] fix issue
* ~~introduces a complication for preg and requires http://codereview.appspot.com/284880043/~~
* ~~https://codereview.appspot.com/276600043/~~
|
process
|
preprocessor not working for windows preprocessor not working for windows mainprocess pid set attribute sysregistry to windows config mainprocess pid set attribute systemroot to windows mainprocess pid set attribute windir to windows mainprocess pid parser filter expression changed to should be mainprocess pid set attribute sysregistry to windows config mainprocess pid set attribute systemroot to windows mainprocess pid set attribute windir to windows mainprocess pid set attribute code page to mainprocess pid set attribute hostname to test mainprocess pid set attribute programfiles to program files fix issue introduces a complication for preg and requires
| 1
|
681,397
| 23,309,660,230
|
IssuesEvent
|
2022-08-08 06:58:49
|
oasis-engine/engine
|
https://api.github.com/repos/oasis-engine/engine
|
closed
|
Improve the text system
|
feature 2D high priority
|
Design: @GuoLei1990 , @singlecoder
PR: @singlecoder @cptbtptpbcptdtptp
PR reviewers: @GuoLei1990 , @cptbtptpbcptdtptp , @gz65555
|
1.0
|
Improve the text system - Design: @GuoLei1990 , @singlecoder
PR: @singlecoder @cptbtptpbcptdtptp
PR reviewers: @GuoLei1990 , @cptbtptpbcptdtptp , @gz65555
|
non_process
|
improve the text system design singlecoder pr singlecoder cptbtptpbcptdtptp pr reviewers cptbtptpbcptdtptp
| 0
|
9,376
| 12,374,399,803
|
IssuesEvent
|
2020-05-19 01:27:35
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
change triage party meeeting to google meet
|
kind/process priority/important-soon
|
for better experience with most of the maintainers
|
1.0
|
change triage party meeeting to google meet - for better experience with most of the maintainers
|
process
|
change triage party meeeting to google meet for better experience with most of the maintainers
| 1
|
3,052
| 6,044,561,726
|
IssuesEvent
|
2017-06-12 06:22:52
|
javabird25/long-hour-and-a-half
|
https://api.github.com/repos/javabird25/long-hour-and-a-half
|
closed
|
"THIS IS A BUG"
|
bug will be processed soon
|
When character is peeing during a class without underwear or/and outerwear, "THIS IS A BUG" will appear.

Because there is no special wear handling in `ASK_TO_PEE` stage.
|
1.0
|
"THIS IS A BUG" - When character is peeing during a class without underwear or/and outerwear, "THIS IS A BUG" will appear.

Because there is no special wear handling in `ASK_TO_PEE` stage.
|
process
|
this is a bug when character is peeing during a class without underwear or and outerwear this is a bug will appear because there is no special wear handling in ask to pee stage
| 1
|
12,114
| 14,740,543,303
|
IssuesEvent
|
2021-01-07 09:15:25
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Bogus Email Address
|
anc-process anp-2 ant-enhancement has attachment
|
In GitLab by @kdjstudios on Nov 19, 2018, 16:01
Hello Team,
I just recently noticed one of the errors we are receiving: [SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg](/uploads/67c74a6f96dfbf330c7c90eba35a7cc1/SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg)
It would appear that we are having an issue with being able to send to "none@none.com"; which is completely valid. My thoughts on this would be the following to resolve this. I did a quick internal search and we have over 15 customers that use this address, and over 600 accounts that use the address. The email is not a required field correct? So we should be able to remove them.
1. Notify operations that on a certain date we will be removing all "bogus" email addresses from the system.
2. If any accounts are setup for email invoicing that have a bogus email address. Notify Operations to correct those accounts accordingly.
3. On that date remove all bogus email addresses.
4. Update the validation on all email address fields to check for bogus email addresses before allowing to save.
This will then in turn any accounts that did not get updated to a valid email and are setup for email invoicing will be displayed with an error to the site managers on the next billing cycle right?
|
1.0
|
Bogus Email Address - In GitLab by @kdjstudios on Nov 19, 2018, 16:01
Hello Team,
I just recently noticed one of the errors we are receiving: [SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg](/uploads/67c74a6f96dfbf330c7c90eba35a7cc1/SA_Billing_Error_Report_customers_update__NetSMTPFatalError__554_5.7.1_none_none.com_Recipient_address_rejected....msg)
It would appear that we are having an issue with being able to send to "none@none.com"; which is completely valid. My thoughts on this would be the following to resolve this. I did a quick internal search and we have over 15 customers that use this address, and over 600 accounts that use the address. The email is not a required field correct? So we should be able to remove them.
1. Notify operations that on a certain date we will be removing all "bogus" email addresses from the system.
2. If any accounts are setup for email invoicing that have a bogus email address. Notify Operations to correct those accounts accordingly.
3. On that date remove all bogus email addresses.
4. Update the validation on all email address fields to check for bogus email addresses before allowing to save.
This will then in turn any accounts that did not get updated to a valid email and are setup for email invoicing will be displayed with an error to the site managers on the next billing cycle right?
|
process
|
bogus email address in gitlab by kdjstudios on nov hello team i just recently noticed one of the errors we are receiving uploads sa billing error report customers update netsmtpfatalerror none none com recipient address rejected msg it would appear that we are having an issue with being able to send to none none com which is completely valid my thoughts on this would be the following to resolve this i did a quick internal search and we have over customers that use this address and over accounts that use the address the email is not a required field correct so we should be able to remove them notify operations that on a certain date we will be removing all bogus email addresses from the system if any accounts are setup for email invoicing that have a bogus email address notify operations to correct those accounts accordingly on that date remove all bogus email addresses update the validation on all email address fields to check for bogus email addresses before allowing to save this will then in turn any accounts that did not get updated to a valid email and are setup for email invoicing will be displayed with an error to the site managers on the next billing cycle right
| 1
|
193,146
| 22,216,072,041
|
IssuesEvent
|
2022-06-08 01:53:00
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
reopened
|
CVE-2017-18552 (High) detected in linux-stable-rtv4.1.33
|
security vulnerability
|
## CVE-2017-18552 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in net/rds/af_rds.c in the Linux kernel before 4.11. There is an out of bounds write and read in the function rds_recv_track_latency.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18552>CVE-2017-18552</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552</a></p>
<p>Release Date: 2019-08-19</p>
<p>Fix Resolution: v4.11-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-18552 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-18552 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/af_rds.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in net/rds/af_rds.c in the Linux kernel before 4.11. There is an out of bounds write and read in the function rds_recv_track_latency.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-18552>CVE-2017-18552</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-18552</a></p>
<p>Release Date: 2019-08-19</p>
<p>Fix Resolution: v4.11-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net rds af rds c net rds af rds c vulnerability details an issue was discovered in net rds af rds c in the linux kernel before there is an out of bounds write and read in the function rds recv track latency publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
587
| 3,062,247,959
|
IssuesEvent
|
2015-08-16 11:45:30
|
yeoman/yeoman-app
|
https://api.github.com/repos/yeoman/yeoman-app
|
opened
|
Align with the Inquirer.js API
|
help-wanted renderer process
|
[Inquirer.js](https://github.com/SBoudrias/Inquirer.js) can handle either an Array of `objects` or an Array of `strings` for the attribute `choices`. Our current implementation supports only an Array of `objects`. To be compatible with the Inquirer.js API we should support this format too.
|
1.0
|
Align with the Inquirer.js API - [Inquirer.js](https://github.com/SBoudrias/Inquirer.js) can handle either an Array of `objects` or an Array of `strings` for the attribute `choices`. Our current implementation supports only an Array of `objects`. To be compatible with the Inquirer.js API we should support this format too.
|
process
|
align with the inquirer js api can handle either an array of objects or an array of strings for the attribute choices our current implementation supports only an array of objects to be compatible with the inquirer js api we should support this format too
| 1
|
13,582
| 16,131,672,904
|
IssuesEvent
|
2021-04-29 06:18:00
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
Need to Support multiple Email Providers
|
Feature request Process: Fixed Process: Tested dev
|
Currently, we are supporting only Gmail but we need to support other email providers as well with MailGun API Support
**Priority**
High
**Labels**
Feature request
|
2.0
|
Need to Support multiple Email Providers - Currently, we are supporting only Gmail but we need to support other email providers as well with MailGun API Support
**Priority**
High
**Labels**
Feature request
|
process
|
need to support multiple email providers currently we are supporting only gmail but we need to support other email providers as well with mailgun api support priority high labels feature request
| 1
|
4,177
| 7,111,594,565
|
IssuesEvent
|
2018-01-17 14:41:29
|
mono/mono
|
https://api.github.com/repos/mono/mono
|
opened
|
Process.WaitForExit() gets slower the larger your open file limit (ulimit -n) is on Linux
|
area-Runtime: Process os-linux
|
## Steps to Reproduce
1. Compile this: `csc test.cs`:
```csharp
public static class Program {
public static int Main () {
for (int i = 0; i < 100; i++) {
var p = System.Diagnostics.Process.Start ("echo", "hello");
if (!p.WaitForExit (10000)) return 1;
}
return 0;
}
}
```
2. `$ (ulimit -n 1000; time mono test.exe)`
```
...
real 0m0.402s
user 0m0.028s
sys 0m0.184s
```
3. `$ (ulimit -n 100000; time mono test.exe)`
```
...
real 0m1.277s
user 0m0.380s
sys 0m0.592s
```
4. `$ (ulimit -n 1000000; time mono test.exe)`
```
...
real 0m8.758s
user 0m3.424s
sys 0m4.424s
```
## Current Behavior
It gets slower the higher your ulimit -n is.
Note: you may need to change your system settings (`/etc/security/limits.conf`) to allow higher limits.
<!--
What is the current behavior?
-->
## Expected Behavior
Not getting slower.
<!--
Please describe the behavior you are expecting
-->
### On which platforms did you notice this
[ ] macOS
[ X ] Linux, Ubuntu 14.04/16.04
[ ] Windows
**Version Used**: master, 5.8.088 and 4.2.1 (so either a very old bug or something outside our control)
This was the root cause behind an issue (https://github.com/mono/mono/pull/6537) we had on Jenkins because the Azure Linux VM builders had ulimit -n set to 1048576.
|
1.0
|
Process.WaitForExit() gets slower the larger your open file limit (ulimit -n) is on Linux - ## Steps to Reproduce
1. Compile this: `csc test.cs`:
```csharp
public static class Program {
public static int Main () {
for (int i = 0; i < 100; i++) {
var p = System.Diagnostics.Process.Start ("echo", "hello");
if (!p.WaitForExit (10000)) return 1;
}
return 0;
}
}
```
2. `$ (ulimit -n 1000; time mono test.exe)`
```
...
real 0m0.402s
user 0m0.028s
sys 0m0.184s
```
3. `$ (ulimit -n 100000; time mono test.exe)`
```
...
real 0m1.277s
user 0m0.380s
sys 0m0.592s
```
4. `$ (ulimit -n 1000000; time mono test.exe)`
```
...
real 0m8.758s
user 0m3.424s
sys 0m4.424s
```
## Current Behavior
It gets slower the higher your ulimit -n is.
Note: you may need to change your system settings (`/etc/security/limits.conf`) to allow higher limits.
<!--
What is the current behavior?
-->
## Expected Behavior
Not getting slower.
<!--
Please describe the behavior you are expecting
-->
### On which platforms did you notice this
[ ] macOS
[ X ] Linux, Ubuntu 14.04/16.04
[ ] Windows
**Version Used**: master, 5.8.088 and 4.2.1 (so either a very old bug or something outside our control)
This was the root cause behind an issue (https://github.com/mono/mono/pull/6537) we had on Jenkins because the Azure Linux VM builders had ulimit -n set to 1048576.
|
process
|
process waitforexit gets slower the larger your open file limit ulimit n is on linux steps to reproduce compile this csc test cs csharp public static class program public static int main for int i i i var p system diagnostics process start echo hello if p waitforexit return return ulimit n time mono test exe real user sys ulimit n time mono test exe real user sys ulimit n time mono test exe real user sys current behavior it gets slower the higher your ulimit n is note you may need to change your system settings etc security limits conf to allow higher limits what is the current behavior expected behavior not getting slower please describe the behavior you are expecting on which platforms did you notice this macos linux ubuntu windows version used master and so either a very old bug or something outside our control this was the root cause behind an issue we had on jenkins because the azure linux vm builders had ulimit n set to
| 1
|
14,611
| 17,754,602,078
|
IssuesEvent
|
2021-08-28 13:56:33
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
While resolving toolchains for target @org_tensorflow//tensorflow/tools/git:gen_git_source: No matching toolchains found for types @bazel_tools//tools/cpp:toolchain_type.
|
type: support / not a bug (process) team-Configurability untriaged
|
> ATTENTION! Please read and follow:
> - if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to bazel-discuss@googlegroups.com
> - if this is a _bug_ or _feature request_, fill the form below as best as you can.
### Description of the problem / feature request:
This is a problem of toolchain resolution, and my platform is m1 apple mbp.
I'm using org_tensorflow to build my project, and fail with the message above.
after some research, I found that it was the @local_execution_config_platform that cause the problem.
According to the BUILD file here:
`
platform(
constraint_values = [
"@bazel_tools//platforms:arm",
"@bazel_tools//platforms:osx",
],
`
platform should use @bazel_tools//platforms:arm to specify the cpu type, which in @bazel_tools//platform , is like:
`alias(
name = "arm",
actual = "@platforms//cpu:arm",
)
`
which in @platform , is like:
`
constraint_value(
name = "arm",
constraint_setting = ":cpu",
)
`
which leads to cpu type mismatch.
According to the @platform, this value in local_execution_config_platform should be @bazel_tools//platforms:aarch64 instead of @bazel_tools//platforms:arm
Now I have to manually change it to make it works.
### Feature requests: what underlying problem are you trying to solve with this feature?
According to the similar problem here: https://github.com/tensorflow/tensorflow/issues/38840
this could be the problem of $MACHTYPE, it's `x86_64` in /bin/zsh and `arm64-apple-darwin20` in /bin/bash
but when I change this to `arm64-apple-darwin20`, still no luck.
Is it a bug or what? Would it be fixed in the future?
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
on a M1 apple mbp device, load @org_tensorflow, should reproduce the problem.
### What operating system are you running Bazel on?
M1 apple mbp
### What's the output of `bazel info release`?
this try to download 3.4.0 bazel
but I already install 4.2.0rc3, but this problem can reproduce in 4.1.0
|
1.0
|
While resolving toolchains for target @org_tensorflow//tensorflow/tools/git:gen_git_source: No matching toolchains found for types @bazel_tools//tools/cpp:toolchain_type. - > ATTENTION! Please read and follow:
> - if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to bazel-discuss@googlegroups.com
> - if this is a _bug_ or _feature request_, fill the form below as best as you can.
### Description of the problem / feature request:
This is a problem of toolchain resolution, and my platform is m1 apple mbp.
I'm using org_tensorflow to build my project, and fail with the message above.
after some research, I found that it was the @local_execution_config_platform that cause the problem.
According to the BUILD file here:
`
platform(
constraint_values = [
"@bazel_tools//platforms:arm",
"@bazel_tools//platforms:osx",
],
`
platform should use @bazel_tools//platforms:arm to specify the cpu type, which in @bazel_tools//platform , is like:
`alias(
name = "arm",
actual = "@platforms//cpu:arm",
)
`
which in @platform , is like:
`
constraint_value(
name = "arm",
constraint_setting = ":cpu",
)
`
which leads to cpu type mismatch.
According to the @platform, this value in local_execution_config_platform should be @bazel_tools//platforms:aarch64 instead of @bazel_tools//platforms:arm
Now I have to manually change it to make it works.
### Feature requests: what underlying problem are you trying to solve with this feature?
According to the similar problem here: https://github.com/tensorflow/tensorflow/issues/38840
this could be the problem of $MACHTYPE, it's `x86_64` in /bin/zsh and `arm64-apple-darwin20` in /bin/bash
but when I change this to `arm64-apple-darwin20`, still no luck.
Is it a bug or what? Would it be fixed in the future?
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
on a M1 apple mbp device, load @org_tensorflow, should reproduce the problem.
### What operating system are you running Bazel on?
M1 apple mbp
### What's the output of `bazel info release`?
this try to download 3.4.0 bazel
but I already install 4.2.0rc3, but this problem can reproduce in 4.1.0
|
process
|
while resolving toolchains for target org tensorflow tensorflow tools git gen git source no matching toolchains found for types bazel tools tools cpp toolchain type attention please read and follow if this is a question about how to build test query deploy using bazel or a discussion starter send it to bazel discuss googlegroups com if this is a bug or feature request fill the form below as best as you can description of the problem feature request this is a problem of toolchain resolution and my platform is apple mbp i m using org tensorflow to build my project and fail with the message above after some research i found that it was the local execution config platform that cause the problem according to the build file here platform constraint values bazel tools platforms arm bazel tools platforms osx platform should use bazel tools platforms arm to specify the cpu type which in bazel tools platform is like alias name arm actual platforms cpu arm which in platform is like constraint value name arm constraint setting cpu which leads to cpu type mismatch according to the platform this value in local execution config platform should be bazel tools platforms instead of bazel tools platforms arm now i have to manually change it to make it works feature requests what underlying problem are you trying to solve with this feature according to the similar problem here this could be the problem of machtype it s in bin zsh and apple in bin bash but when i change this to apple still no luck is it a bug or what would it be fixed in the future bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible on a apple mbp device load org tensorflow should reproduce the problem what operating system are you running bazel on apple mbp what s the output of bazel info release this try to download bazel but i already install but this problem can reproduce in
| 1
|
319,170
| 23,759,642,214
|
IssuesEvent
|
2022-09-01 07:47:05
|
Unity-Technologies/com.unity.multiplayer.docs
|
https://api.github.com/repos/Unity-Technologies/com.unity.multiplayer.docs
|
closed
|
Feedback for /versioned_docs/version-1.0.0/migration/upgrade-guide.md
|
documentation IN JIRA
|
I've just upgraded from MLAPI to the new library and:
- there's no mention of the replacements/alternatives for NetworkManager.Singleton.StopHost() and NetworkManager.Singleton.StopClient() in the documentation
- messaging needs updating to say that CustomMessagingManager.RegisterNamedMessageHandler no longer works stand-alone but instead needs to be prefixed with "NetworkManager.Singleton."
- string arrays can no longer be used for RPC parameters. The page titled "C# Primitives" still lists "string" even though the migration page says we're now supposed to use Unity.Collections.FixedStringxxxxBytes structs instead
|
1.0
|
Feedback for /versioned_docs/version-1.0.0/migration/upgrade-guide.md - I've just upgraded from MLAPI to the new library and:
- there's no mention of the replacements/alternatives for NetworkManager.Singleton.StopHost() and NetworkManager.Singleton.StopClient() in the documentation
- messaging needs updating to say that CustomMessagingManager.RegisterNamedMessageHandler no longer works stand-alone but instead needs to be prefixed with "NetworkManager.Singleton."
- string arrays can no longer be used for RPC parameters. The page titled "C# Primitives" still lists "string" even though the migration page says we're now supposed to use Unity.Collections.FixedStringxxxxBytes structs instead
|
non_process
|
feedback for versioned docs version migration upgrade guide md i ve just upgraded from mlapi to the new library and there s no mention of the replacements alternatives for networkmanager singleton stophost and networkmanager singleton stopclient in the documentation messaging needs updating to say that custommessagingmanager registernamedmessagehandler no longer works stand alone but instead needs to be prefixed with networkmanager singleton string arrays can no longer be used for rpc parameters the page titled c primitives still lists string even though the migration page says we re now supposed to use unity collections fixedstringxxxxbytes structs instead
| 0
|
5,309
| 8,125,526,762
|
IssuesEvent
|
2018-08-16 21:14:54
|
MetaMask/metamask-extension
|
https://api.github.com/repos/MetaMask/metamask-extension
|
closed
|
Process: Reduce number of head branches
|
L09-process P2-sooner T02-refactor
|
we have a lot of "head" branches and we should reduce the number to reduce complexity:
- master
- uat
- mascara-prod
- newUI-flat
i suggest:
- master
- uat <- mascara-prod + newUI-flat
|
1.0
|
Process: Reduce number of head branches - we have a lot of "head" branches and we should reduce the number to reduce complexity:
- master
- uat
- mascara-prod
- newUI-flat
i suggest:
- master
- uat <- mascara-prod + newUI-flat
|
process
|
process reduce number of head branches we have a lot of head branches and we should reduce the number to reduce complexity master uat mascara prod newui flat i suggest master uat mascara prod newui flat
| 1
|
289,752
| 31,997,403,853
|
IssuesEvent
|
2023-09-21 10:02:54
|
Mend-developer-platform-load/8945136_33
|
https://api.github.com/repos/Mend-developer-platform-load/8945136_33
|
opened
|
CVE-2022-38750 (Medium) detected in snakeyaml-1.19.jar
|
Mend: dependency security vulnerability
|
## CVE-2022-38750 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.19.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p>
<p>
Dependency Hierarchy:
- rocketmq-broker-4.6.0.jar (Root Library)
- rocketmq-acl-4.6.0.jar
- :x: **snakeyaml-1.19.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Mend-developer-platform-load/8945136_33/commit/3b8eebc04951436067314b2f84da6260dad2fbda">3b8eebc04951436067314b2f84da6260dad2fbda</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38750>CVE-2022-38750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution (org.yaml:snakeyaml): 1.31</p>
<p>Direct dependency fix Resolution (org.apache.rocketmq:rocketmq-broker): 5.1.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation will be attempted for this issue.
|
True
|
CVE-2022-38750 (Medium) detected in snakeyaml-1.19.jar - ## CVE-2022-38750 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.19.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p>
<p>
Dependency Hierarchy:
- rocketmq-broker-4.6.0.jar (Root Library)
- rocketmq-acl-4.6.0.jar
- :x: **snakeyaml-1.19.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Mend-developer-platform-load/8945136_33/commit/3b8eebc04951436067314b2f84da6260dad2fbda">3b8eebc04951436067314b2f84da6260dad2fbda</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38750>CVE-2022-38750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47027</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution (org.yaml:snakeyaml): 1.31</p>
<p>Direct dependency fix Resolution (org.apache.rocketmq:rocketmq-broker): 5.1.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation will be attempted for this issue.
|
non_process
|
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy rocketmq broker jar root library rocketmq acl jar x snakeyaml jar vulnerable library found in head commit a href found in base branch main vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution org apache rocketmq rocketmq broker rescue worker helmet automatic remediation will be attempted for this issue
| 0
|
105,797
| 4,242,024,952
|
IssuesEvent
|
2016-07-06 18:09:18
|
easydigitaldownloads/easy-digital-downloads
|
https://api.github.com/repos/easydigitaldownloads/easy-digital-downloads
|
closed
|
PayPal: Negative fees getting included twice
|
Bug High Priority Payments PayPal
|
When an order has negative fees (such as through Discounts Pro), the discounts are taken into account twice once the customer reaches PayPal. See screenshots below.


This is almost certainly due to changes introduced in #4632 and #4658
The issue affects PayPal Standard for sure but I believe it also affects at minimum the PayPal Pro / Express and 2Checkout gateways. We may have to fix it in the gateways in order to prevent undoing our bug fixes with negative fees.
|
1.0
|
PayPal: Negative fees getting included twice - When an order has negative fees (such as through Discounts Pro), the discounts are taken into account twice once the customer reaches PayPal. See screenshots below.


This is almost certainly due to changes introduced in #4632 and #4658
The issue affects PayPal Standard for sure but I believe it also affects at minimum the PayPal Pro / Express and 2Checkout gateways. We may have to fix it in the gateways in order to prevent undoing our bug fixes with negative fees.
|
non_process
|
paypal negative fees getting included twice when an order has negative fees such as through discounts pro the discounts are taken into account twice once the customer reaches paypal see screenshots below this is almost certainly due to changes introduced in and the issue affects paypal standard for sure but i believe it also affects at minimum the paypal pro express and gateways we may have to fix it in the gateways in order to prevent undoing our bug fixes with negative fees
| 0
|
9,517
| 12,498,729,083
|
IssuesEvent
|
2020-06-01 18:49:52
|
KasiaCieslik/sentiment-analysis-of-tweets-using-emoticons
|
https://api.github.com/repos/KasiaCieslik/sentiment-analysis-of-tweets-using-emoticons
|
opened
|
Write script that uses spacy to preprocess the tweet
|
data pre-processing high priority medium task
|
To Do:
- use spacy to transform a tweet into a list of grammatical parts of a sentence (e.g., verb, adjective etc.) and find out other elementary aspects of a tweet (length etc.)
|
1.0
|
Write script that uses spacy to preprocess the tweet - To Do:
- use spacy to transform a tweet into a list of grammatical parts of a sentence (e.g., verb, adjective etc.) and find out other elementary aspects of a tweet (length etc.)
|
process
|
write script that uses spacy to preprocess the tweet to do use spacy to transform a tweet into a list of grammatical parts of a sentence e g verb adjective etc and find out other elementary aspects of a tweet length etc
| 1
|
274,207
| 8,558,509,785
|
IssuesEvent
|
2018-11-08 18:28:39
|
Iridescent-CM/technovation-app
|
https://api.github.com/repos/Iridescent-CM/technovation-app
|
opened
|
As an RA I cannot save teams or judges that I invite to my event
|
Bug Priority - 1 bug
|
**Describe the bug**
As an RA I can select a team to add to my event, but I can't save them
**To Reproduce**
Steps to reproduce the behavior:
1. Log in as Ra@ra.com
2. Click on Events
3. Find an event and click on 'manage teams'
4. Invite a team and try to click 'save teams'
5. Nothing happens
6. When the page is refreshed the teams have not been added
**Expected behavior**
I expect to be able to add the teams and judges.
**Screenshots**

**Desktop (please complete the following information):**
- OS: MacOS Mojave
- Browser: Chrome
**Label**
You MUST use the label menu to add 'bug' and '0 - To discuss' as a label
(do *not* choose a bug priority label, devs will do that according to the Defect Priority matrix)
|
1.0
|
As an RA I cannot save teams or judges that I invite to my event - **Describe the bug**
As an RA I can select a team to add to my event, but I can't save them
**To Reproduce**
Steps to reproduce the behavior:
1. Log in as Ra@ra.com
2. Click on Events
3. Find an event and click on 'manage teams'
4. Invite a team and try to click 'save teams'
5. Nothing happens
6. When the page is refreshed the teams have not been added
**Expected behavior**
I expect to be able to add the teams and judges.
**Screenshots**

**Desktop (please complete the following information):**
- OS: MacOS Mojave
- Browser: Chrome
**Label**
You MUST use the label menu to add 'bug' and '0 - To discuss' as a label
(do *not* choose a bug priority label, devs will do that according to the Defect Priority matrix)
|
non_process
|
as an ra i cannot save teams or judges that i invite to my event describe the bug as an ra i can select a team to add to my event but i can t save them to reproduce steps to reproduce the behavior log in as ra ra com click on events find an event and click on manage teams invite a team and try to click save teams nothing happens when the page is refreshed the teams have not been added expected behavior i expect to be able to add the teams and judges screenshots desktop please complete the following information os macos mojave browser chrome label you must use the label menu to add bug and to discuss as a label do not choose a bug priority label devs will do that according to the defect priority matrix
| 0
|
387,951
| 11,472,636,300
|
IssuesEvent
|
2020-02-09 18:25:42
|
islos-efe-eme/auto-news
|
https://api.github.com/repos/islos-efe-eme/auto-news
|
opened
|
Implement "ping" endpoint in bots
|
good first issue new-feature priority:low slack telegram webhook
|
It's a useful thing to have a simple and fast way to check if a service is online or not, this doesn't means it's working properly but at least replying to external requests through the network.
- [ ] Ping endpoint for Slack bot
- [ ] Ping endpoint for Telegram bot (**blocked**)
- [ ] Unit test for Slack bot
- [ ] Unit test for Telegram bot (**blocked**)
Expected request:
```js
GET /ping
```
Expected response:
```js
// status 200
{ message: "Pong" }
```
|
1.0
|
Implement "ping" endpoint in bots - It's a useful thing to have a simple and fast way to check if a service is online or not, this doesn't means it's working properly but at least replying to external requests through the network.
- [ ] Ping endpoint for Slack bot
- [ ] Ping endpoint for Telegram bot (**blocked**)
- [ ] Unit test for Slack bot
- [ ] Unit test for Telegram bot (**blocked**)
Expected request:
```js
GET /ping
```
Expected response:
```js
// status 200
{ message: "Pong" }
```
|
non_process
|
implement ping endpoint in bots it s a useful thing to have a simple and fast way to check if a service is online or not this doesn t means it s working properly but at least replying to external requests through the network ping endpoint for slack bot ping endpoint for telegram bot blocked unit test for slack bot unit test for telegram bot blocked expected request js get ping expected response js status message pong
| 0
|
323
| 2,772,000,275
|
IssuesEvent
|
2015-05-02 07:10:12
|
FG-Team/HCJ-Website-Builder
|
https://api.github.com/repos/FG-Team/HCJ-Website-Builder
|
closed
|
[DEPRECATED] Tree structure
|
Feature No Processing
|
-Klasse TreeView: Graphik für Baumstruktur, beinhaltet einen javax.swing.JTree
-Klasse TreeModel: Datenmodell
->Inhalt in Vector<TreeNode> gespeichert.
-Klasse TreeNode: Repräsentation eines Baumastes/-blattes, beinhaltet Namen und Ebene des Blattes.
-Listener für Tree/Model
-Klasse TreePresenter: presenter
alle im core Package
|
1.0
|
[DEPRECATED] Tree structure - -Klasse TreeView: Graphik für Baumstruktur, beinhaltet einen javax.swing.JTree
-Klasse TreeModel: Datenmodell
->Inhalt in Vector<TreeNode> gespeichert.
-Klasse TreeNode: Repräsentation eines Baumastes/-blattes, beinhaltet Namen und Ebene des Blattes.
-Listener für Tree/Model
-Klasse TreePresenter: presenter
alle im core Package
|
process
|
tree structure klasse treeview graphik für baumstruktur beinhaltet einen javax swing jtree klasse treemodel datenmodell inhalt in vector gespeichert klasse treenode repräsentation eines baumastes blattes beinhaltet namen und ebene des blattes listener für tree model klasse treepresenter presenter alle im core package
| 1
|
13,536
| 16,067,667,023
|
IssuesEvent
|
2021-04-23 22:17:12
|
Azure/azure-event-hubs-java
|
https://api.github.com/repos/Azure/azure-event-hubs-java
|
closed
|
Unable to pass RetryPolicy to EventProcessorHost constructor with AzureStorageCheckpoint
|
EventProcessorHost enhancement
|
## Actual Behavior
EventProcessorHost has only one constructor which accepts RetryPolicy:
```
EventProcessorHost(
final String hostName,
final String eventHubPath,
final String consumerGroupName,
final String eventHubConnectionString,
ICheckpointManager checkpointManager,
ILeaseManager leaseManager,
ScheduledExecutorService executorService,
RetryPolicy retryPolicy)
```
As you can see I need to pass `ICheckpointManager` and `ILeaseManager`, but in case I want to reuse built-in implementation `AzureStorageCheckpointLeaseManager` I can't do it, because it is non-public class with package-level constructor.
## Expected Behavior
`EventProcessorHost` should have additional constructor overload where it is possible to pass RetryPolicy, ScheduledExecutorService and use Azure Storage checkpointing/leasing
## Versions
- OS platform and version: MacOs
- Maven package version or commit ID: azure-eventhubs-eph: 2.0.1
|
1.0
|
Unable to pass RetryPolicy to EventProcessorHost constructor with AzureStorageCheckpoint - ## Actual Behavior
EventProcessorHost has only one constructor which accepts RetryPolicy:
```
EventProcessorHost(
final String hostName,
final String eventHubPath,
final String consumerGroupName,
final String eventHubConnectionString,
ICheckpointManager checkpointManager,
ILeaseManager leaseManager,
ScheduledExecutorService executorService,
RetryPolicy retryPolicy)
```
As you can see I need to pass `ICheckpointManager` and `ILeaseManager`, but in case I want to reuse built-in implementation `AzureStorageCheckpointLeaseManager` I can't do it, because it is non-public class with package-level constructor.
## Expected Behavior
`EventProcessorHost` should have additional constructor overload where it is possible to pass RetryPolicy, ScheduledExecutorService and use Azure Storage checkpointing/leasing
## Versions
- OS platform and version: MacOs
- Maven package version or commit ID: azure-eventhubs-eph: 2.0.1
|
process
|
unable to pass retrypolicy to eventprocessorhost constructor with azurestoragecheckpoint actual behavior eventprocessorhost has only one constructor which accepts retrypolicy eventprocessorhost final string hostname final string eventhubpath final string consumergroupname final string eventhubconnectionstring icheckpointmanager checkpointmanager ileasemanager leasemanager scheduledexecutorservice executorservice retrypolicy retrypolicy as you can see i need to pass icheckpointmanager and ileasemanager but in case i want to reuse built in implementation azurestoragecheckpointleasemanager i can t do it because it is non public class with package level constructor expected behavior eventprocessorhost should have additional constructor overload where it is possible to pass retrypolicy scheduledexecutorservice and use azure storage checkpointing leasing versions os platform and version macos maven package version or commit id azure eventhubs eph
| 1
|
151,470
| 12,037,258,214
|
IssuesEvent
|
2020-04-13 21:26:43
|
ansible/ansible
|
https://api.github.com/repos/ansible/ansible
|
closed
|
l3_interfaces of junos_facts gather_network_resources returns error
|
affects_2.9 bug module needs_triage networking python3 support:core support:network system test traceback
|
<!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
- l3_interfaces submodule of junos_facts: gather_network_resources: throws a AttributeError: 'NoneType' object has no attribute 'text' error.
These submodules of junos_facts work just fine in the same playbook:
junos_facts:
gather_network_resources:
- interfaces
- l2_interfaces
- lag_interfaces
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
junos_facts:
gather_network_resources:
- l3_interfaces
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
``'
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible --version
ansible 2.9.6
config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg
configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible-config dump --only-changed
ANSIBLE_PIPELINING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = True
DEFAULT_GATHERING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = explicit
DEFAULT_HOST_LIST(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = [u'/opt/adi/nsg_hosts.py']
DEFAULT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260
DEFAULT_TRANSPORT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = local
ENABLE_TASK_DEBUGGER(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False
HOST_KEY_CHECKING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False
PERSISTENT_COMMAND_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260
PERSISTENT_CONNECT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260
RETRY_FILES_ENABLED(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ cat /etc/centos-release
CentOS Linux release 7.6.1810 (Core)
Running ansible 2.9.6, python 3.
Also tested in ansible 2.9.1
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
Run the playbook, without -l3_interfaces called the module will run and print out gathered facts successfully.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- name: Gathers juniper 1a switch facts
gather_facts: false
hosts: juniperhost
serial: 10
vars_prompt:
- name: username
prompt: "Enter Username"
private: no
- name: password
private: yes
prompt: "Enter Password"
confirm: yes
vars:
cli:
host: "{{ inventory_hostname }}"
username: "{{ username }}"
password: "{{ password }}"
pre_tasks:
- name: gather facts
junos_facts:
gather_network_resources:
- interfaces
- l2_interfaces
- lag_interfaces
- l3_interfaces
provider: "{{ cli }}"
- name: dubug
debug:
msg: "{{ ansible_facts }}"
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Expected that data will be gathered, printed to the screen
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ rpb pb.gatherJunosFacts.yaml -vvvv
ansible-playbook 2.9.6
config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg
configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible-playbook
python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)]
Using /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg as config file
setting up inventory plugins
host_list declined parsing /opt/adi/nsg_hosts.py as it did not pass its verify_file() method
Parsed /opt/adi/nsg_hosts.py inventory source with script plugin
Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/site-packages/ansible/plugins/callback/default.pyc
PLAYBOOK: pb.gatherJunosFacts.yaml *******************************************************************************************************************************************************************************
Positional arguments: pb.gatherJunosFacts.yaml
become_method: sudo
inventory: (u'/opt/adi/nsg_hosts.py',)
forks: 5
tags: (u'all',)
verbosity: 4
connection: local
timeout: 260
1 plays in pb.gatherJunosFacts.yaml
Enter Username:
Enter Password:
confirm Enter Password:
PLAY [Gathers juniper 1a switch facts] ***************************************************************************************************************************************************************************
TASK [gather facts] **********************************************************************************************************************************************************************************************
task path: /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/pb.gatherJunosFacts.yaml:24
<js922-1a> using connection plugin netconf (was local)
<js922-1a> starting connection from persistent connection plugin
<js922-1a> local domain socket does not exist, starting it
<js922-1a> control socket path is /home/username/.ansible/pc/314f4ddcc8
<js922-1a> local domain socket listeners started successfully
<js922-1a> loaded netconf plugin junos from path /usr/lib/python2.7/site-packages/ansible/plugins/netconf/junos.py for network_os junos
<js922-1a>
<js922-1a> local domain socket path is /home/username/.ansible/pc/314f4ddcc8
<js922-1a> socket_path: /home/username/.ansible/pc/314f4ddcc8
<js922-1a> Attempting python interpreter discovery
<js922-1a> ESTABLISH LOCAL CONNECTION FOR USER: username
<js922-1a> EXEC /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'/usr/bin/python'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'python3.6'"'"'; command -v '"'"'python3.5'"'"'; command -v '"'"'python2.7'"'"'; command -v '"'"'python2.6'"'"'; command -v '"'"'/usr/libexec/platform-python'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'python'"'"'; echo ENDFOUND && sleep 0'
<js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0'
Using module file /usr/lib/python2.7/site-packages/ansible/modules/network/junos/junos_facts.py
Pipelining is enabled.
<js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0'
The full traceback is:
Traceback (most recent call last):
File "<stdin>", line 102, in <module>
File "<stdin>", line 94, in _ansiballz_main
File "<stdin>", line 40, in invoke_module
File "/usr/lib64/python2.7/runpy.py", line 176, in run_module
fname, loader, pkg_name)
File "/usr/lib64/python2.7/runpy.py", line 82, in _run_module_code
mod_name, mod_fname, mod_loader, pkg_name)
File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 135, in <module>
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 126, in main
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py", line 62, in get_facts
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py", line 105, in get_network_resources_facts
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 72, in populate_facts
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 93, in parse_l3_if_resources
AttributeError: 'NoneType' object has no attribute 'text'
fatal: [js922-1a]: FAILED! => {
"ansible_facts": {
"discovered_interpreter_python": "/usr/bin/python"
},
"changed": false,
"module_stderr": "Traceback (most recent call last):\n File \"<stdin>\", line 102, in <module>\n File \"<stdin>\", line 94, in _ansiballz_main\n File \"<stdin>\", line 40, in invoke_module\n File \"/usr/lib64/python2.7/runpy.py\", line 176, in run_module\n fname, loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 82, in _run_module_code\n mod_name, mod_fname, mod_loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 72, in _run_code\n exec code in run_globals\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 135, in <module>\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 126, in main\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py\", line 62, in get_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py\", line 105, in get_network_resources_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 72, in populate_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 93, in parse_l3_if_resources\nAttributeError: 'NoneType' object has no attribute 'text'\n",
"module_stdout": "",
"msg": "MODULE FAILURE\nSee stdout/stderr for the exact error",
"rc": 1
}
PLAY RECAP *******************************************************************************************************************************************************************************************************
js922-1a : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0
<!--- Paste verbatim command output between quotes -->
```paste below
ansible-playbook pb.gatherJunosFacts.yaml -vvvv
```
|
1.0
|
l3_interfaces of junos_facts gather_network_resources returns error - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
- l3_interfaces submodule of junos_facts: gather_network_resources: throws a AttributeError: 'NoneType' object has no attribute 'text' error.
These submodules of junos_facts work just fine in the same playbook:
junos_facts:
gather_network_resources:
- interfaces
- l2_interfaces
- lag_interfaces
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
junos_facts:
gather_network_resources:
- l3_interfaces
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
``'
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible --version
ansible 2.9.6
config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg
configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ ansible-config dump --only-changed
ANSIBLE_PIPELINING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = True
DEFAULT_GATHERING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = explicit
DEFAULT_HOST_LIST(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = [u'/opt/adi/nsg_hosts.py']
DEFAULT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260
DEFAULT_TRANSPORT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = local
ENABLE_TASK_DEBUGGER(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False
HOST_KEY_CHECKING(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False
PERSISTENT_COMMAND_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260
PERSISTENT_CONNECT_TIMEOUT(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = 260
RETRY_FILES_ENABLED(/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg) = False
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ cat /etc/centos-release
CentOS Linux release 7.6.1810 (Core)
Running ansible 2.9.6, python 3.
Also tested in ansible 2.9.1
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
Run the playbook, without -l3_interfaces called the module will run and print out gathered facts successfully.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- name: Gathers juniper 1a switch facts
gather_facts: false
hosts: juniperhost
serial: 10
vars_prompt:
- name: username
prompt: "Enter Username"
private: no
- name: password
private: yes
prompt: "Enter Password"
confirm: yes
vars:
cli:
host: "{{ inventory_hostname }}"
username: "{{ username }}"
password: "{{ password }}"
pre_tasks:
- name: gather facts
junos_facts:
gather_network_resources:
- interfaces
- l2_interfaces
- lag_interfaces
- l3_interfaces
provider: "{{ cli }}"
- name: dubug
debug:
msg: "{{ ansible_facts }}"
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Expected that data will be gathered, printed to the screen
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
username@y0319p11981:/opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier$ rpb pb.gatherJunosFacts.yaml -vvvv
ansible-playbook 2.9.6
config file = /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg
configured module search path = [u'/home/username/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible-playbook
python version = 2.7.5 (default, Oct 30 2018, 23:45:53) [GCC 4.8.5 20150623 (Red Hat 4.8.5-36)]
Using /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/ansible.cfg as config file
setting up inventory plugins
host_list declined parsing /opt/adi/nsg_hosts.py as it did not pass its verify_file() method
Parsed /opt/adi/nsg_hosts.py inventory source with script plugin
Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/site-packages/ansible/plugins/callback/default.pyc
PLAYBOOK: pb.gatherJunosFacts.yaml *******************************************************************************************************************************************************************************
Positional arguments: pb.gatherJunosFacts.yaml
become_method: sudo
inventory: (u'/opt/adi/nsg_hosts.py',)
forks: 5
tags: (u'all',)
verbosity: 4
connection: local
timeout: 260
1 plays in pb.gatherJunosFacts.yaml
Enter Username:
Enter Password:
confirm Enter Password:
PLAY [Gathers juniper 1a switch facts] ***************************************************************************************************************************************************************************
TASK [gather facts] **********************************************************************************************************************************************************************************************
task path: /opt/nsgtools/NSG_storage/LR2020/pb.2_9/sitebuilder/mansibleRouters/mansible_jquerier/pb.gatherJunosFacts.yaml:24
<js922-1a> using connection plugin netconf (was local)
<js922-1a> starting connection from persistent connection plugin
<js922-1a> local domain socket does not exist, starting it
<js922-1a> control socket path is /home/username/.ansible/pc/314f4ddcc8
<js922-1a> local domain socket listeners started successfully
<js922-1a> loaded netconf plugin junos from path /usr/lib/python2.7/site-packages/ansible/plugins/netconf/junos.py for network_os junos
<js922-1a>
<js922-1a> local domain socket path is /home/username/.ansible/pc/314f4ddcc8
<js922-1a> socket_path: /home/username/.ansible/pc/314f4ddcc8
<js922-1a> Attempting python interpreter discovery
<js922-1a> ESTABLISH LOCAL CONNECTION FOR USER: username
<js922-1a> EXEC /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'/usr/bin/python'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'python3.6'"'"'; command -v '"'"'python3.5'"'"'; command -v '"'"'python2.7'"'"'; command -v '"'"'python2.6'"'"'; command -v '"'"'/usr/libexec/platform-python'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'python'"'"'; echo ENDFOUND && sleep 0'
<js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0'
Using module file /usr/lib/python2.7/site-packages/ansible/modules/network/junos/junos_facts.py
Pipelining is enabled.
<js922-1a> EXEC /bin/sh -c '/usr/bin/python && sleep 0'
The full traceback is:
Traceback (most recent call last):
File "<stdin>", line 102, in <module>
File "<stdin>", line 94, in _ansiballz_main
File "<stdin>", line 40, in invoke_module
File "/usr/lib64/python2.7/runpy.py", line 176, in run_module
fname, loader, pkg_name)
File "/usr/lib64/python2.7/runpy.py", line 82, in _run_module_code
mod_name, mod_fname, mod_loader, pkg_name)
File "/usr/lib64/python2.7/runpy.py", line 72, in _run_code
exec code in run_globals
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 135, in <module>
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py", line 126, in main
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py", line 62, in get_facts
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py", line 105, in get_network_resources_facts
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 72, in populate_facts
File "/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py", line 93, in parse_l3_if_resources
AttributeError: 'NoneType' object has no attribute 'text'
fatal: [js922-1a]: FAILED! => {
"ansible_facts": {
"discovered_interpreter_python": "/usr/bin/python"
},
"changed": false,
"module_stderr": "Traceback (most recent call last):\n File \"<stdin>\", line 102, in <module>\n File \"<stdin>\", line 94, in _ansiballz_main\n File \"<stdin>\", line 40, in invoke_module\n File \"/usr/lib64/python2.7/runpy.py\", line 176, in run_module\n fname, loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 82, in _run_module_code\n mod_name, mod_fname, mod_loader, pkg_name)\n File \"/usr/lib64/python2.7/runpy.py\", line 72, in _run_code\n exec code in run_globals\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 135, in <module>\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/modules/network/junos/junos_facts.py\", line 126, in main\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/facts.py\", line 62, in get_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/common/facts/facts.py\", line 105, in get_network_resources_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 72, in populate_facts\n File \"/tmp/ansible_junos_facts_payload_BwEgMs/ansible_junos_facts_payload.zip/ansible/module_utils/network/junos/facts/l3_interfaces/l3_interfaces.py\", line 93, in parse_l3_if_resources\nAttributeError: 'NoneType' object has no attribute 'text'\n",
"module_stdout": "",
"msg": "MODULE FAILURE\nSee stdout/stderr for the exact error",
"rc": 1
}
PLAY RECAP *******************************************************************************************************************************************************************************************************
js922-1a : ok=0 changed=0 unreachable=0 failed=1 skipped=0 rescued=0 ignored=0
<!--- Paste verbatim command output between quotes -->
```paste below
ansible-playbook pb.gatherJunosFacts.yaml -vvvv
```
|
non_process
|
interfaces of junos facts gather network resources returns error summary interfaces submodule of junos facts gather network resources throws a attributeerror nonetype object has no attribute text error these submodules of junos facts work just fine in the same playbook junos facts gather network resources interfaces interfaces lag interfaces issue type bug report component name junos facts gather network resources interfaces ansible version username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible version ansible config file opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default oct configuration paste below username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible config dump only changed ansible pipelining opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg true default gathering opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg explicit default host list opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg default timeout opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg default transport opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg local enable task debugger opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg false host key checking opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg false persistent command timeout opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg persistent connect timeout opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg retry files enabled opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg false username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier os environment username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier cat etc centos release centos linux release core running ansible python also tested in ansible steps to reproduce run the playbook without interfaces called the module will run and print out gathered facts successfully yaml name gathers juniper switch facts gather facts false hosts juniperhost serial vars prompt name username prompt enter username private no name password private yes prompt enter password confirm yes vars cli host inventory hostname username username password password pre tasks name gather facts junos facts gather network resources interfaces interfaces lag interfaces interfaces provider cli name dubug debug msg ansible facts expected results expected that data will be gathered printed to the screen actual results username opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier rpb pb gatherjunosfacts yaml vvvv ansible playbook config file opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible playbook python version default oct using opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier ansible cfg as config file setting up inventory plugins host list declined parsing opt adi nsg hosts py as it did not pass its verify file method parsed opt adi nsg hosts py inventory source with script plugin loading callback plugin default of type stdout from usr lib site packages ansible plugins callback default pyc playbook pb gatherjunosfacts yaml positional arguments pb gatherjunosfacts yaml become method sudo inventory u opt adi nsg hosts py forks tags u all verbosity connection local timeout plays in pb gatherjunosfacts yaml enter username enter password confirm enter password play task task path opt nsgtools nsg storage pb sitebuilder mansiblerouters mansible jquerier pb gatherjunosfacts yaml using connection plugin netconf was local starting connection from persistent connection plugin local domain socket does not exist starting it control socket path is home username ansible pc local domain socket listeners started successfully loaded netconf plugin junos from path usr lib site packages ansible plugins netconf junos py for network os junos local domain socket path is home username ansible pc socket path home username ansible pc attempting python interpreter discovery establish local connection for user username exec bin sh c echo platform uname echo found command v usr bin python command v command v command v command v command v command v usr libexec platform python command v usr bin command v python echo endfound sleep exec bin sh c usr bin python sleep using module file usr lib site packages ansible modules network junos junos facts py pipelining is enabled exec bin sh c usr bin python sleep the full traceback is traceback most recent call last file line in file line in ansiballz main file line in invoke module file usr runpy py line in run module fname loader pkg name file usr runpy py line in run module code mod name mod fname mod loader pkg name file usr runpy py line in run code exec code in run globals file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in main file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts facts py line in get facts file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network common facts facts py line in get network resources facts file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in populate facts file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in parse if resources attributeerror nonetype object has no attribute text fatal failed ansible facts discovered interpreter python usr bin python changed false module stderr traceback most recent call last n file line in n file line in ansiballz main n file line in invoke module n file usr runpy py line in run module n fname loader pkg name n file usr runpy py line in run module code n mod name mod fname mod loader pkg name n file usr runpy py line in run code n exec code in run globals n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible modules network junos junos facts py line in main n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts facts py line in get facts n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network common facts facts py line in get network resources facts n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in populate facts n file tmp ansible junos facts payload bwegms ansible junos facts payload zip ansible module utils network junos facts interfaces interfaces py line in parse if resources nattributeerror nonetype object has no attribute text n module stdout msg module failure nsee stdout stderr for the exact error rc play recap ok changed unreachable failed skipped rescued ignored paste below ansible playbook pb gatherjunosfacts yaml vvvv
| 0
|
51,493
| 12,740,528,507
|
IssuesEvent
|
2020-06-26 02:53:32
|
carla-simulator/carla
|
https://api.github.com/repos/carla-simulator/carla
|
closed
|
Docker image for building Carla on Linux
|
backlog build system feature request help wanted
|
Make a Dockerfile that allows us building an image with a Linux environment set up for building Carla.
The main complication is having Unreal built without redistributing the code/binaries.
|
1.0
|
Docker image for building Carla on Linux - Make a Dockerfile that allows us building an image with a Linux environment set up for building Carla.
The main complication is having Unreal built without redistributing the code/binaries.
|
non_process
|
docker image for building carla on linux make a dockerfile that allows us building an image with a linux environment set up for building carla the main complication is having unreal built without redistributing the code binaries
| 0
|
34,806
| 16,685,373,121
|
IssuesEvent
|
2021-06-08 07:28:59
|
PrestaShop/PrestaShop
|
https://api.github.com/repos/PrestaShop/PrestaShop
|
closed
|
Update stock via webserivice with memcache exception and memory leak
|
1.7.7.2 Bug NMI Performance Stocks WS
|
#### Describe the bug
When i am updating stock via web service with memcached enable i am getting exception and also i have memory leak (ram 128GB)
When memcached is disabled everything works as expected.
#### Expected behavior
update stock
#### Steps to Reproduce
Steps to reproduce the behavior:
1. Enable memcached cache
2. Try to update stock via web service
3. HttpStatusCode: InternalServerError
4. Invalid argument supplied for foreach() (.../httpdocs/ps1772/classes/cache/Cache.php line 567
**Screenshots**
Request:
```
<?xml version="1.0" encoding="utf-8"?>
<prestashop>
<stock_available>
<id>52920</id>
<id_product>1024345</id_product>
<id_product_attribute>0</id_product_attribute>
<id_shop>1</id_shop>
<id_shop_group>0</id_shop_group>
<quantity>1</quantity>
<depends_on_stock>0</depends_on_stock>
<out_of_stock>2</out_of_stock>
</stock_available>
</prestashop>
```
Response:
```
<?xml version="1.0" encoding="UTF-8"?>
<prestashop xmlns:xlink="http://www.w3.org/1999/xlink">
<errors>
<error>
<code><![CDATA[3]]></code>
<message><![CDATA[[PHP Warning #2] Invalid argument supplied for foreach() (/var/www/vhosts/xxx/httpdocs/ps1772/classes/cache/Cache.php, line 567)]]></message>
</error>
</errors>
</prestashop>
HttpStatusCode: InternalServerError
```
#### Additional information
- Web server Apache + NGINX Reverse proxy + php fpm on apache.
- PrestaShop version: 1.7.7.2
- Server information Linux #1 SMP Debian 4.19.171-2 (2021-01-30) x86_64
- Server software version: Apache
- PHP version: 7.3.28
- Memory limit: 512M
- Max execution time: 300
- Upload Max File size: 24M
- MySQL version: 10.3.27-MariaDB-0+deb10u1
|
True
|
Update stock via webserivice with memcache exception and memory leak - #### Describe the bug
When i am updating stock via web service with memcached enable i am getting exception and also i have memory leak (ram 128GB)
When memcached is disabled everything works as expected.
#### Expected behavior
update stock
#### Steps to Reproduce
Steps to reproduce the behavior:
1. Enable memcached cache
2. Try to update stock via web service
3. HttpStatusCode: InternalServerError
4. Invalid argument supplied for foreach() (.../httpdocs/ps1772/classes/cache/Cache.php line 567
**Screenshots**
Request:
```
<?xml version="1.0" encoding="utf-8"?>
<prestashop>
<stock_available>
<id>52920</id>
<id_product>1024345</id_product>
<id_product_attribute>0</id_product_attribute>
<id_shop>1</id_shop>
<id_shop_group>0</id_shop_group>
<quantity>1</quantity>
<depends_on_stock>0</depends_on_stock>
<out_of_stock>2</out_of_stock>
</stock_available>
</prestashop>
```
Response:
```
<?xml version="1.0" encoding="UTF-8"?>
<prestashop xmlns:xlink="http://www.w3.org/1999/xlink">
<errors>
<error>
<code><![CDATA[3]]></code>
<message><![CDATA[[PHP Warning #2] Invalid argument supplied for foreach() (/var/www/vhosts/xxx/httpdocs/ps1772/classes/cache/Cache.php, line 567)]]></message>
</error>
</errors>
</prestashop>
HttpStatusCode: InternalServerError
```
#### Additional information
- Web server Apache + NGINX Reverse proxy + php fpm on apache.
- PrestaShop version: 1.7.7.2
- Server information Linux #1 SMP Debian 4.19.171-2 (2021-01-30) x86_64
- Server software version: Apache
- PHP version: 7.3.28
- Memory limit: 512M
- Max execution time: 300
- Upload Max File size: 24M
- MySQL version: 10.3.27-MariaDB-0+deb10u1
|
non_process
|
update stock via webserivice with memcache exception and memory leak describe the bug when i am updating stock via web service with memcached enable i am getting exception and also i have memory leak ram when memcached is disabled everything works as expected expected behavior update stock steps to reproduce steps to reproduce the behavior enable memcached cache try to update stock via web service httpstatuscode internalservererror invalid argument supplied for foreach httpdocs classes cache cache php line screenshots request response prestashop xmlns xlink httpstatuscode internalservererror additional information web server apache nginx reverse proxy php fpm on apache prestashop version server information linux smp debian server software version apache php version memory limit max execution time upload max file size mysql version mariadb
| 0
|
65,328
| 12,553,756,046
|
IssuesEvent
|
2020-06-06 23:17:45
|
Pokecube-Development/Pokecube-Issues-and-Wiki
|
https://api.github.com/repos/Pokecube-Development/Pokecube-Issues-and-Wiki
|
closed
|
Metagross dupped
|
1.14.x 1.15.2 Bug - Code Fixed
|
#### Issue Description:
Metagross ressurrect and dupped
#### What happens:
my metagross has a bug, when it died to some golbats when I was upgrading I ran to cure it so I wouldn't die, when I realized there was another metagross following me that had resurrected
#### What you expected to happen:
the metagross no duppe on die
#### Steps to reproduce:
1.the pokemon die
2.pokebolla returns to your inventory
3.the body that does not disappear resurrects and the pokemon duppe
...
____
#### Affected Versions (Do *not* use "latest"): Replace with a list of all mods you have in.
- Pokecube AIO:
- Minecraft: 1.15.2
- Forge: 31.1.18
|
1.0
|
Metagross dupped -
#### Issue Description:
Metagross ressurrect and dupped
#### What happens:
my metagross has a bug, when it died to some golbats when I was upgrading I ran to cure it so I wouldn't die, when I realized there was another metagross following me that had resurrected
#### What you expected to happen:
the metagross no duppe on die
#### Steps to reproduce:
1.the pokemon die
2.pokebolla returns to your inventory
3.the body that does not disappear resurrects and the pokemon duppe
...
____
#### Affected Versions (Do *not* use "latest"): Replace with a list of all mods you have in.
- Pokecube AIO:
- Minecraft: 1.15.2
- Forge: 31.1.18
|
non_process
|
metagross dupped issue description metagross ressurrect and dupped what happens my metagross has a bug when it died to some golbats when i was upgrading i ran to cure it so i wouldn t die when i realized there was another metagross following me that had resurrected what you expected to happen the metagross no duppe on die steps to reproduce the pokemon die pokebolla returns to your inventory the body that does not disappear resurrects and the pokemon duppe affected versions do not use latest replace with a list of all mods you have in pokecube aio minecraft forge
| 0
|
476
| 2,911,385,686
|
IssuesEvent
|
2015-06-22 09:13:47
|
haskell-distributed/distributed-process-simplelocalnet
|
https://api.github.com/repos/haskell-distributed/distributed-process-simplelocalnet
|
opened
|
SimpleLocalnet should remove peers that no longer exist
|
distributed-process-simplelocalnet Feature Request
|
_From @edsko on November 7, 2012 15:33_
This is not a *major* issue, because even if we do, there is still no guarantee that a peer still exists after it has been returned by `findPeers`.
_Copied from original issue: haskell-distributed/distributed-process#66_
|
1.0
|
SimpleLocalnet should remove peers that no longer exist - _From @edsko on November 7, 2012 15:33_
This is not a *major* issue, because even if we do, there is still no guarantee that a peer still exists after it has been returned by `findPeers`.
_Copied from original issue: haskell-distributed/distributed-process#66_
|
process
|
simplelocalnet should remove peers that no longer exist from edsko on november this is not a major issue because even if we do there is still no guarantee that a peer still exists after it has been returned by findpeers copied from original issue haskell distributed distributed process
| 1
|
10,304
| 4,038,834,705
|
IssuesEvent
|
2016-05-20 00:14:41
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
opened
|
Object+Barcode to Container
|
Function-ContainerOrBarcode
|
We are starting a large project to add barcodes to existing tissue vials. At least initially the upload of barcodes to specimen records with be done using the Object+Barcode to Container tool in Object Tracking. However, there a some issues that I would like to see revised for ease of use and error control.
1) "Parent Cont Type" should be changed to read "Move to Container Type" or "Barcode Container Type". My student was adding in cryovial barcodes and misunderstood this. She assumed the "parent" of a cryovial was a freezer box - this resulted in all the cryovials being converted to freezer boxes inside the freezer box positions.
2) The "Move to Container Type" field above should have a null default value so as not to change the type of an existing container unless the user specifically intends to do so (it certainly shouldn't be set on "Nalgene" as the automatic default requiring manual change to a different value.)
3) Please if at all possible add a live link to the specimen record from this page, so that the specimen record can be accessed easily for verification.
|
1.0
|
Object+Barcode to Container - We are starting a large project to add barcodes to existing tissue vials. At least initially the upload of barcodes to specimen records with be done using the Object+Barcode to Container tool in Object Tracking. However, there a some issues that I would like to see revised for ease of use and error control.
1) "Parent Cont Type" should be changed to read "Move to Container Type" or "Barcode Container Type". My student was adding in cryovial barcodes and misunderstood this. She assumed the "parent" of a cryovial was a freezer box - this resulted in all the cryovials being converted to freezer boxes inside the freezer box positions.
2) The "Move to Container Type" field above should have a null default value so as not to change the type of an existing container unless the user specifically intends to do so (it certainly shouldn't be set on "Nalgene" as the automatic default requiring manual change to a different value.)
3) Please if at all possible add a live link to the specimen record from this page, so that the specimen record can be accessed easily for verification.
|
non_process
|
object barcode to container we are starting a large project to add barcodes to existing tissue vials at least initially the upload of barcodes to specimen records with be done using the object barcode to container tool in object tracking however there a some issues that i would like to see revised for ease of use and error control parent cont type should be changed to read move to container type or barcode container type my student was adding in cryovial barcodes and misunderstood this she assumed the parent of a cryovial was a freezer box this resulted in all the cryovials being converted to freezer boxes inside the freezer box positions the move to container type field above should have a null default value so as not to change the type of an existing container unless the user specifically intends to do so it certainly shouldn t be set on nalgene as the automatic default requiring manual change to a different value please if at all possible add a live link to the specimen record from this page so that the specimen record can be accessed easily for verification
| 0
|
62,500
| 15,278,925,586
|
IssuesEvent
|
2021-02-23 02:41:13
|
Elaniobro/MMM-nyc-transit
|
https://api.github.com/repos/Elaniobro/MMM-nyc-transit
|
closed
|
[🐛 BUG] - Please put your title here.
|
:building_construction: dependencies :white_medium_square: minor
|
**Describe the bug**
Bump version of station pkg once npm published.
|
1.0
|
[🐛 BUG] - Please put your title here. - **Describe the bug**
Bump version of station pkg once npm published.
|
non_process
|
please put your title here describe the bug bump version of station pkg once npm published
| 0
|
8,231
| 11,415,575,212
|
IssuesEvent
|
2020-02-02 12:02:10
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
Changing `scss`/`css` file breaks style on livereload
|
:bug: Bug CSS Preprocessing HMR Stale
|
**Choose one:** is this a 🐛 bug report or 🙋 feature request?
🐛 bug report
### 🎛 Configuration (.babelrc, package.json, cli command)
The project I'm working is available [here](https://github.com/wescoder/tavern), you can checkout and reproduce the issue on it.
<!--- If describing a bug, tell us what your babel configuration looks like -->
.babelrc:
```js
{
"presets": ["env", "preact"],
"plugins": [
"transform-function-bind",
["transform-react-jsx", { "pragma": "h" }],
["module-resolver", {
"root": ["."],
"alias": {
"@app": "./src"
}
}],
"transform-object-rest-spread"
]
}
```
.postcssrc:
```
{
"modules": true,
"plugins": {
"autoprefixer": {
"grid": true
}
}
}
```
### 🤔 Expected Behavior
Changing a `scss` file with css modules enabled that is "named-imported" (`import s from './style.scss'`) in js and used in `className` by `preact`/`react`/etc should update the js that requires it providing the new `className` variable to it, without needing a manual full page reload.
### 😯 Current Behavior
Changing the `scss` updates the stylesheet causing the network fetch for the new asset but not for the js file, so all styles changed breaks on the page since the old classes doesn't exist anymore, to see the changes a manual full page reload is needed.
### 💁 Possible Solution
I really don't have a clue about this... Sorry...
### 🔦 Context
For example going into `src/home/home.scss` and changing `.Title` color will reproduce the bug (as of commit [451f3a2](https://github.com/wescoder/tavern/commit/451f3a2a5c23188f08419e33c8d2689e0c83e765))
### 💻 Code Sample
You can see the open sourced project here ([https://github.com/wescoder/tavern](https://github.com/wescoder/tavern))
### 🌍 Your Environment
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | ^1.4.1 |
| Node | v9.2.0 |
| npm/Yarn | 1.3.2
| Operating System | Linux Deepin 15.5 (Ubuntu based) |
|
1.0
|
Changing `scss`/`css` file breaks style on livereload - **Choose one:** is this a 🐛 bug report or 🙋 feature request?
🐛 bug report
### 🎛 Configuration (.babelrc, package.json, cli command)
The project I'm working is available [here](https://github.com/wescoder/tavern), you can checkout and reproduce the issue on it.
<!--- If describing a bug, tell us what your babel configuration looks like -->
.babelrc:
```js
{
"presets": ["env", "preact"],
"plugins": [
"transform-function-bind",
["transform-react-jsx", { "pragma": "h" }],
["module-resolver", {
"root": ["."],
"alias": {
"@app": "./src"
}
}],
"transform-object-rest-spread"
]
}
```
.postcssrc:
```
{
"modules": true,
"plugins": {
"autoprefixer": {
"grid": true
}
}
}
```
### 🤔 Expected Behavior
Changing a `scss` file with css modules enabled that is "named-imported" (`import s from './style.scss'`) in js and used in `className` by `preact`/`react`/etc should update the js that requires it providing the new `className` variable to it, without needing a manual full page reload.
### 😯 Current Behavior
Changing the `scss` updates the stylesheet causing the network fetch for the new asset but not for the js file, so all styles changed breaks on the page since the old classes doesn't exist anymore, to see the changes a manual full page reload is needed.
### 💁 Possible Solution
I really don't have a clue about this... Sorry...
### 🔦 Context
For example going into `src/home/home.scss` and changing `.Title` color will reproduce the bug (as of commit [451f3a2](https://github.com/wescoder/tavern/commit/451f3a2a5c23188f08419e33c8d2689e0c83e765))
### 💻 Code Sample
You can see the open sourced project here ([https://github.com/wescoder/tavern](https://github.com/wescoder/tavern))
### 🌍 Your Environment
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | ^1.4.1 |
| Node | v9.2.0 |
| npm/Yarn | 1.3.2
| Operating System | Linux Deepin 15.5 (Ubuntu based) |
|
process
|
changing scss css file breaks style on livereload choose one is this a 🐛 bug report or 🙋 feature request 🐛 bug report 🎛 configuration babelrc package json cli command the project i m working is available you can checkout and reproduce the issue on it babelrc js presets plugins transform function bind module resolver root alias app src transform object rest spread postcssrc modules true plugins autoprefixer grid true 🤔 expected behavior changing a scss file with css modules enabled that is named imported import s from style scss in js and used in classname by preact react etc should update the js that requires it providing the new classname variable to it without needing a manual full page reload 😯 current behavior changing the scss updates the stylesheet causing the network fetch for the new asset but not for the js file so all styles changed breaks on the page since the old classes doesn t exist anymore to see the changes a manual full page reload is needed 💁 possible solution i really don t have a clue about this sorry 🔦 context for example going into src home home scss and changing title color will reproduce the bug as of commit 💻 code sample you can see the open sourced project here 🌍 your environment software version s parcel node npm yarn operating system linux deepin ubuntu based
| 1
|
4,640
| 7,482,390,589
|
IssuesEvent
|
2018-04-05 01:04:27
|
UnbFeelings/unb-feelings-GQA
|
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
|
opened
|
Definir critérios de avalição
|
document help wanted organization process wiki
|
- [ ] Criar documento "Objetos de Auditoria"
- [ ] Listar todos os objetos a serem auditados
- [ ] Definir critérios de avaliação para cada objeto auditado
|
1.0
|
Definir critérios de avalição - - [ ] Criar documento "Objetos de Auditoria"
- [ ] Listar todos os objetos a serem auditados
- [ ] Definir critérios de avaliação para cada objeto auditado
|
process
|
definir critérios de avalição criar documento objetos de auditoria listar todos os objetos a serem auditados definir critérios de avaliação para cada objeto auditado
| 1
|
17,895
| 23,872,230,172
|
IssuesEvent
|
2022-09-07 15:42:10
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
closed
|
[SQL Connector] send avro message to topic and consume using SQL connector with avro format has EOF exception
|
compute/data-processing
|
```
Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records
at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:150)
at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
... 1 more
Caused by: java.io.IOException: java.io.IOException: Failed to deserialize Avro record.
at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:141)
at org.apache.flink.connector.pulsar.source.reader.split.PulsarOrderedPartitionSplitReader.fetch(PulsarOrderedPartitionSplitReader.java:50)
at org.apache.flink.connector.base.source.reader.fetcher.FetchTask.run(FetchTask.java:58)
at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:142)
... 6 more
Caused by: java.io.IOException: Failed to deserialize Avro record.
at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:106)
at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:46)
at org.apache.flink.api.common.serialization.DeserializationSchema.deserialize(DeserializationSchema.java:82)
at org.apache.flink.connector.pulsar.table.source.PulsarTableDeserializationSchema.deserialize(PulsarTableDeserializationSchema.java:88)
at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:123)
... 9 more
Caused by: java.io.EOFException
at org.apache.avro.io.BinaryDecoder$InputStreamByteSource.readRaw(BinaryDecoder.java:851)
at org.apache.avro.io.BinaryDecoder.doReadBytes(BinaryDecoder.java:373)
at org.apache.avro.io.BinaryDecoder.readString(BinaryDecoder.java:290)
at org.apache.avro.io.ResolvingDecoder.readString(ResolvingDecoder.java:208)
at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:469)
at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:459)
at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:191)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160)
at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:187)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160)
at org.apache.avro.generic.GenericDatumReader.readField(GenericDatumReader.java:259)
at org.apache.avro.generic.GenericDatumReader.readRecord(GenericDatumReader.java:247)
at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:179)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:153)
at org.apache.flink.formats.avro.AvroDeserializationSchema.deserialize(AvroDeserializationSchema.java:142)
at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:103)
... 13 more
```
It encounted this error in the sendMessageToTopicAndReadUsingAvroBasedSchema() method. We know that such use cases succeeded before. Need to figure out why it's failing now.
|
1.0
|
[SQL Connector] send avro message to topic and consume using SQL connector with avro format has EOF exception - ```
Caused by: java.lang.RuntimeException: SplitFetcher thread 0 received unexpected exception while polling the records
at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:150)
at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.run(SplitFetcher.java:105)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
... 1 more
Caused by: java.io.IOException: java.io.IOException: Failed to deserialize Avro record.
at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:141)
at org.apache.flink.connector.pulsar.source.reader.split.PulsarOrderedPartitionSplitReader.fetch(PulsarOrderedPartitionSplitReader.java:50)
at org.apache.flink.connector.base.source.reader.fetcher.FetchTask.run(FetchTask.java:58)
at org.apache.flink.connector.base.source.reader.fetcher.SplitFetcher.runOnce(SplitFetcher.java:142)
... 6 more
Caused by: java.io.IOException: Failed to deserialize Avro record.
at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:106)
at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:46)
at org.apache.flink.api.common.serialization.DeserializationSchema.deserialize(DeserializationSchema.java:82)
at org.apache.flink.connector.pulsar.table.source.PulsarTableDeserializationSchema.deserialize(PulsarTableDeserializationSchema.java:88)
at org.apache.flink.connector.pulsar.source.reader.split.PulsarPartitionSplitReaderBase.fetch(PulsarPartitionSplitReaderBase.java:123)
... 9 more
Caused by: java.io.EOFException
at org.apache.avro.io.BinaryDecoder$InputStreamByteSource.readRaw(BinaryDecoder.java:851)
at org.apache.avro.io.BinaryDecoder.doReadBytes(BinaryDecoder.java:373)
at org.apache.avro.io.BinaryDecoder.readString(BinaryDecoder.java:290)
at org.apache.avro.io.ResolvingDecoder.readString(ResolvingDecoder.java:208)
at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:469)
at org.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:459)
at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:191)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160)
at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:187)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160)
at org.apache.avro.generic.GenericDatumReader.readField(GenericDatumReader.java:259)
at org.apache.avro.generic.GenericDatumReader.readRecord(GenericDatumReader.java:247)
at org.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:179)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:160)
at org.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:153)
at org.apache.flink.formats.avro.AvroDeserializationSchema.deserialize(AvroDeserializationSchema.java:142)
at org.apache.flink.formats.avro.AvroRowDataDeserializationSchema.deserialize(AvroRowDataDeserializationSchema.java:103)
... 13 more
```
It encounted this error in the sendMessageToTopicAndReadUsingAvroBasedSchema() method. We know that such use cases succeeded before. Need to figure out why it's failing now.
|
process
|
send avro message to topic and consume using sql connector with avro format has eof exception caused by java lang runtimeexception splitfetcher thread received unexpected exception while polling the records at org apache flink connector base source reader fetcher splitfetcher runonce splitfetcher java at org apache flink connector base source reader fetcher splitfetcher run splitfetcher java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java more caused by java io ioexception java io ioexception failed to deserialize avro record at org apache flink connector pulsar source reader split pulsarpartitionsplitreaderbase fetch pulsarpartitionsplitreaderbase java at org apache flink connector pulsar source reader split pulsarorderedpartitionsplitreader fetch pulsarorderedpartitionsplitreader java at org apache flink connector base source reader fetcher fetchtask run fetchtask java at org apache flink connector base source reader fetcher splitfetcher runonce splitfetcher java more caused by java io ioexception failed to deserialize avro record at org apache flink formats avro avrorowdatadeserializationschema deserialize avrorowdatadeserializationschema java at org apache flink formats avro avrorowdatadeserializationschema deserialize avrorowdatadeserializationschema java at org apache flink api common serialization deserializationschema deserialize deserializationschema java at org apache flink connector pulsar table source pulsartabledeserializationschema deserialize pulsartabledeserializationschema java at org apache flink connector pulsar source reader split pulsarpartitionsplitreaderbase fetch pulsarpartitionsplitreaderbase java more caused by java io eofexception at org apache avro io binarydecoder inputstreambytesource readraw binarydecoder java at org apache avro io binarydecoder doreadbytes binarydecoder java at org apache avro io binarydecoder readstring binarydecoder java at org apache avro io resolvingdecoder readstring resolvingdecoder java at org apache avro generic genericdatumreader readstring genericdatumreader java at org apache avro generic genericdatumreader readstring genericdatumreader java at org apache avro generic genericdatumreader readwithoutconversion genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache avro generic genericdatumreader readwithoutconversion genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache avro generic genericdatumreader readfield genericdatumreader java at org apache avro generic genericdatumreader readrecord genericdatumreader java at org apache avro generic genericdatumreader readwithoutconversion genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache avro generic genericdatumreader read genericdatumreader java at org apache flink formats avro avrodeserializationschema deserialize avrodeserializationschema java at org apache flink formats avro avrorowdatadeserializationschema deserialize avrorowdatadeserializationschema java more it encounted this error in the sendmessagetotopicandreadusingavrobasedschema method we know that such use cases succeeded before need to figure out why it s failing now
| 1
|
17,770
| 23,698,732,518
|
IssuesEvent
|
2022-08-29 16:53:34
|
cloudfoundry/korifi
|
https://api.github.com/repos/cloudfoundry/korifi
|
opened
|
[Feature]: Developer can push apps using the top-level `disk_quota` field in the manifest
|
Top-level process config
|
### Blockers/Dependencies
_No response_
### Background
**As a** developer
**I want** top-level process configuration in manifests to be supported
**So that** I can use shortcut `cf push` flags like `-c`, `-i`, `-m` etc.
### Acceptance Criteria
**GIVEN** I have the sources of an application (e.g. `tests/smoke/assets/test-node-app`)
**WHEN I** push it with the following command:
```sh
cf push test -k 1GB
```
**THEN I** see the push succeeds with an output similar to this:
```
name: test
requested state: started
routes: test.vcap.me
last uploaded: Mon 29 Aug 16:28:36 UTC 2022
stack: cflinuxfs3
buildpacks:
name version detect output buildpack name
nodejs_buildpack 1.7.61 nodejs nodejs
type: web
sidecars:
instances: 1/1
memory usage: 256M
start command: npm start
state since cpu memory disk details
#0 running 2022-08-29T16:28:54Z 1.6% 42.3M of 1G 115.7M of 1G
```
### Dev Notes
The CLI will use the top-level `disk_quota` field on app objects for this, but [only if](https://github.com/cloudfoundry/cli/blob/68d897a19792bdffe20c39589148216b23c950d1/actor/v7pushaction/handle_disk_override.go#L14-L20) there's no `manifest.yml`, or if `manifest.yml` doesn't already have a `web` process definition.
|
1.0
|
[Feature]: Developer can push apps using the top-level `disk_quota` field in the manifest - ### Blockers/Dependencies
_No response_
### Background
**As a** developer
**I want** top-level process configuration in manifests to be supported
**So that** I can use shortcut `cf push` flags like `-c`, `-i`, `-m` etc.
### Acceptance Criteria
**GIVEN** I have the sources of an application (e.g. `tests/smoke/assets/test-node-app`)
**WHEN I** push it with the following command:
```sh
cf push test -k 1GB
```
**THEN I** see the push succeeds with an output similar to this:
```
name: test
requested state: started
routes: test.vcap.me
last uploaded: Mon 29 Aug 16:28:36 UTC 2022
stack: cflinuxfs3
buildpacks:
name version detect output buildpack name
nodejs_buildpack 1.7.61 nodejs nodejs
type: web
sidecars:
instances: 1/1
memory usage: 256M
start command: npm start
state since cpu memory disk details
#0 running 2022-08-29T16:28:54Z 1.6% 42.3M of 1G 115.7M of 1G
```
### Dev Notes
The CLI will use the top-level `disk_quota` field on app objects for this, but [only if](https://github.com/cloudfoundry/cli/blob/68d897a19792bdffe20c39589148216b23c950d1/actor/v7pushaction/handle_disk_override.go#L14-L20) there's no `manifest.yml`, or if `manifest.yml` doesn't already have a `web` process definition.
|
process
|
developer can push apps using the top level disk quota field in the manifest blockers dependencies no response background as a developer i want top level process configuration in manifests to be supported so that i can use shortcut cf push flags like c i m etc acceptance criteria given i have the sources of an application e g tests smoke assets test node app when i push it with the following command sh cf push test k then i see the push succeeds with an output similar to this name test requested state started routes test vcap me last uploaded mon aug utc stack buildpacks name version detect output buildpack name nodejs buildpack nodejs nodejs type web sidecars instances memory usage start command npm start state since cpu memory disk details running of of dev notes the cli will use the top level disk quota field on app objects for this but there s no manifest yml or if manifest yml doesn t already have a web process definition
| 1
|
5,196
| 7,974,012,858
|
IssuesEvent
|
2018-07-17 02:43:30
|
pelias/pelias
|
https://api.github.com/repos/pelias/pelias
|
closed
|
Harden dependency on libpostal
|
processed
|
Currently, if pelias-config specifies `parser: addressit` instead of the expected `parser: libpostal`, exceptions are thrown upon incoming requests. Since we have no intention of supporting Pelias without `libpostal` we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly.
Startup of API should also fail if libpostal is not properly installed or could not be loaded for some reason.
|
1.0
|
Harden dependency on libpostal - Currently, if pelias-config specifies `parser: addressit` instead of the expected `parser: libpostal`, exceptions are thrown upon incoming requests. Since we have no intention of supporting Pelias without `libpostal` we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly.
Startup of API should also fail if libpostal is not properly installed or could not be loaded for some reason.
|
process
|
harden dependency on libpostal currently if pelias config specifies parser addressit instead of the expected parser libpostal exceptions are thrown upon incoming requests since we have no intention of supporting pelias without libpostal we should make that clear to users by removing the config parameter and documenting the hard dependency more explicitly startup of api should also fail if libpostal is not properly installed or could not be loaded for some reason
| 1
|
249,104
| 7,953,807,638
|
IssuesEvent
|
2018-07-12 03:58:35
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
USER ISSUE: second layer items not visible
|
Medium Priority
|
**Version:** 0.7.2.3 beta
**Steps to Reproduce:**
drop 1st layer of items or wall down
drop 2nd layer down
2nd layer isnt visable
disconnect
reconnect
2nd layer visable
**Expected behavior:**
2nd layer of items be visable after placement
items taken from carry inventory
**Actual behavior:**
2nd layer of items invisable
items taken from carry inventory
|
1.0
|
USER ISSUE: second layer items not visible - **Version:** 0.7.2.3 beta
**Steps to Reproduce:**
drop 1st layer of items or wall down
drop 2nd layer down
2nd layer isnt visable
disconnect
reconnect
2nd layer visable
**Expected behavior:**
2nd layer of items be visable after placement
items taken from carry inventory
**Actual behavior:**
2nd layer of items invisable
items taken from carry inventory
|
non_process
|
user issue second layer items not visible version beta steps to reproduce drop layer of items or wall down drop layer down layer isnt visable disconnect reconnect layer visable expected behavior layer of items be visable after placement items taken from carry inventory actual behavior layer of items invisable items taken from carry inventory
| 0
|
177,192
| 13,686,148,637
|
IssuesEvent
|
2020-09-30 08:16:39
|
photoprism/photoprism
|
https://api.github.com/repos/photoprism/photoprism
|
closed
|
Fixed, Success Message in German
|
bug please-test
|
In the German language, when changing an album, after pressing the Save button, the message comes in English (Album saved).
<img width="774" alt="Bildschirmfoto 2020-09-01 um 21 36 48" src="https://user-images.githubusercontent.com/49251031/91898594-41ba4080-ec9c-11ea-923a-1370c6955855.png">
<img width="525" alt="Bildschirmfoto 2020-09-01 um 21 41 37" src="https://user-images.githubusercontent.com/49251031/91898610-454dc780-ec9c-11ea-9c27-696d18ebdbbf.png">
|
1.0
|
Fixed, Success Message in German - In the German language, when changing an album, after pressing the Save button, the message comes in English (Album saved).
<img width="774" alt="Bildschirmfoto 2020-09-01 um 21 36 48" src="https://user-images.githubusercontent.com/49251031/91898594-41ba4080-ec9c-11ea-923a-1370c6955855.png">
<img width="525" alt="Bildschirmfoto 2020-09-01 um 21 41 37" src="https://user-images.githubusercontent.com/49251031/91898610-454dc780-ec9c-11ea-9c27-696d18ebdbbf.png">
|
non_process
|
fixed success message in german in the german language when changing an album after pressing the save button the message comes in english album saved img width alt bildschirmfoto um src img width alt bildschirmfoto um src
| 0
|
318,225
| 23,708,577,631
|
IssuesEvent
|
2022-08-30 05:21:45
|
starlite-api/starlite
|
https://api.github.com/repos/starlite-api/starlite
|
closed
|
Bug: Documentation Site doesn't work properly in mobile.
|
bug documentation help wanted good first issue
|
The Doc site isn't working well in mobile. You need to press the back arrow to find the menu to see usage and reference.
|
1.0
|
Bug: Documentation Site doesn't work properly in mobile. - The Doc site isn't working well in mobile. You need to press the back arrow to find the menu to see usage and reference.
|
non_process
|
bug documentation site doesn t work properly in mobile the doc site isn t working well in mobile you need to press the back arrow to find the menu to see usage and reference
| 0
|
88,545
| 8,154,687,915
|
IssuesEvent
|
2018-08-23 04:48:45
|
trashgenerator/dcore
|
https://api.github.com/repos/trashgenerator/dcore
|
opened
|
DC-8-Config сделать Unit test
|
testing
|
Unit test на сохранение / загрузку /валидацию значений конфиг-файла
|
1.0
|
DC-8-Config сделать Unit test - Unit test на сохранение / загрузку /валидацию значений конфиг-файла
|
non_process
|
dc config сделать unit test unit test на сохранение загрузку валидацию значений конфиг файла
| 0
|
12,545
| 14,975,688,580
|
IssuesEvent
|
2021-01-28 06:40:54
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
opened
|
Update license and copyright to support rosetta
|
P3 enhancement process rosetta
|
**Problem**
Licenses and Copyright currently cover all the modules except rosetta.
**Solution**
Update license and copyright to support go rosetta files
**Alternatives**
A clear and concise description of any alternative solutions or features you've considered.
**Additional Context**
Add any other context or screenshots about the feature request here.
|
1.0
|
Update license and copyright to support rosetta - **Problem**
Licenses and Copyright currently cover all the modules except rosetta.
**Solution**
Update license and copyright to support go rosetta files
**Alternatives**
A clear and concise description of any alternative solutions or features you've considered.
**Additional Context**
Add any other context or screenshots about the feature request here.
|
process
|
update license and copyright to support rosetta problem licenses and copyright currently cover all the modules except rosetta solution update license and copyright to support go rosetta files alternatives a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
| 1
|
3,536
| 6,573,789,459
|
IssuesEvent
|
2017-09-11 10:07:26
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Should overwriting child_process.execFile result in changed behaviour of child_process.exec?
|
child_process
|
```sh
node -e "child_process.execFile = null; child_process.exec('cat package.json')"
child_process.js:138
return exports.execFile(opts.file,
TypeError: exports.execFile is not a function
at Object.exports.exec (child_process.js:138:18)
```
I couldn't find in documentation why child_process behave this way (but it behaves like that for at least [four years](https://github.com/nodejs/node/commit/e8df2676748e944388896dfd767e01906ae2e4eb)).
Relevant line https://github.com/nodejs/node/blob/58831b2/lib/child_process.js#L136
|
1.0
|
Should overwriting child_process.execFile result in changed behaviour of child_process.exec? - ```sh
node -e "child_process.execFile = null; child_process.exec('cat package.json')"
child_process.js:138
return exports.execFile(opts.file,
TypeError: exports.execFile is not a function
at Object.exports.exec (child_process.js:138:18)
```
I couldn't find in documentation why child_process behave this way (but it behaves like that for at least [four years](https://github.com/nodejs/node/commit/e8df2676748e944388896dfd767e01906ae2e4eb)).
Relevant line https://github.com/nodejs/node/blob/58831b2/lib/child_process.js#L136
|
process
|
should overwriting child process execfile result in changed behaviour of child process exec sh node e child process execfile null child process exec cat package json child process js return exports execfile opts file typeerror exports execfile is not a function at object exports exec child process js i couldn t find in documentation why child process behave this way but it behaves like that for at least relevant line
| 1
|
205,805
| 7,106,120,709
|
IssuesEvent
|
2018-01-16 15:43:19
|
flowup/ngx-swagger-client-generator
|
https://api.github.com/repos/flowup/ngx-swagger-client-generator
|
closed
|
Type Errors
|
bug priority:HIGH
|
Hi,
I'm having some issues with my generated code. On build I get a bunch of errors like:
```
.../api-client-service.ts(379,59): error TS2345: Argument of type 'HttpOptions' is not assignable to parameter of type '{ headers?: HttpHeaders | { [header: string]: string | string[]; }; observe?: "body"; params?: Ht...'.
Types of property 'responseType' are incompatible.
Type '"text" | "arraybuffer" | "blob" | "json"' is not assignable to type '"json"'.
Type '"text"' is not assignable to type '"json"'.
```
This seem to me like it's probably a versioning issue, e.g. the typescript or angular versions are wrong. I'm using Angular 4.x.x with typescript 2.4.0. I tried angular 5.x.x and TS 2.6.x and has the same issues. Maybe the code you generate targets an older angular version?
Additionally `HttpObserve` is not exported in `/common/http`. To get the generated code to build at all I've edited the HttpOptions interface as follows:
```
interface HttpOptions {
headers?: HttpHeaders,
params?: HttpParams,
reportProgress?: boolean,
withCredentials?: boolean,
}
```
Any ideas?
|
1.0
|
Type Errors - Hi,
I'm having some issues with my generated code. On build I get a bunch of errors like:
```
.../api-client-service.ts(379,59): error TS2345: Argument of type 'HttpOptions' is not assignable to parameter of type '{ headers?: HttpHeaders | { [header: string]: string | string[]; }; observe?: "body"; params?: Ht...'.
Types of property 'responseType' are incompatible.
Type '"text" | "arraybuffer" | "blob" | "json"' is not assignable to type '"json"'.
Type '"text"' is not assignable to type '"json"'.
```
This seem to me like it's probably a versioning issue, e.g. the typescript or angular versions are wrong. I'm using Angular 4.x.x with typescript 2.4.0. I tried angular 5.x.x and TS 2.6.x and has the same issues. Maybe the code you generate targets an older angular version?
Additionally `HttpObserve` is not exported in `/common/http`. To get the generated code to build at all I've edited the HttpOptions interface as follows:
```
interface HttpOptions {
headers?: HttpHeaders,
params?: HttpParams,
reportProgress?: boolean,
withCredentials?: boolean,
}
```
Any ideas?
|
non_process
|
type errors hi i m having some issues with my generated code on build i get a bunch of errors like api client service ts error argument of type httpoptions is not assignable to parameter of type headers httpheaders string string observe body params ht types of property responsetype are incompatible type text arraybuffer blob json is not assignable to type json type text is not assignable to type json this seem to me like it s probably a versioning issue e g the typescript or angular versions are wrong i m using angular x x with typescript i tried angular x x and ts x and has the same issues maybe the code you generate targets an older angular version additionally httpobserve is not exported in common http to get the generated code to build at all i ve edited the httpoptions interface as follows interface httpoptions headers httpheaders params httpparams reportprogress boolean withcredentials boolean any ideas
| 0
|
320,019
| 23,799,458,546
|
IssuesEvent
|
2022-09-03 03:42:22
|
ued123/modeling
|
https://api.github.com/repos/ued123/modeling
|
opened
|
음료수 구매하기 도메인 개선
|
documentation
|
## 요약
### project: cash-api
- 자판기 입금/출금 액을 관리할수 있는 도메인: cash
- 현금 거래 내역: transaction
### project: purchase-api
- cash, drink 로직을 참조해서 사용할수 있습니다
- 주문: purchase
|
1.0
|
음료수 구매하기 도메인 개선 - ## 요약
### project: cash-api
- 자판기 입금/출금 액을 관리할수 있는 도메인: cash
- 현금 거래 내역: transaction
### project: purchase-api
- cash, drink 로직을 참조해서 사용할수 있습니다
- 주문: purchase
|
non_process
|
음료수 구매하기 도메인 개선 요약 project cash api 자판기 입금 출금 액을 관리할수 있는 도메인 cash 현금 거래 내역 transaction project purchase api cash drink 로직을 참조해서 사용할수 있습니다 주문 purchase
| 0
|
21,724
| 30,232,574,680
|
IssuesEvent
|
2023-07-06 08:02:25
|
UnitTestBot/UTBotJava
|
https://api.github.com/repos/UnitTestBot/UTBotJava
|
closed
|
Perform every concrete execution in a separate transaction that is rollbacked right after execution
|
ctg-enhancement comp-instrumented-process comp-spring
|
**Description**
Right now database content is reset between concrete executions with hardcoded SQL statements tuned for `spring-boot-testing-main` project database scheme (e.g. `ALTER TABLE ${repositoryDefinition.tableName} ALTER COLUMN id RESTART WITH 1`).
**Expected behavior**
Every concrete execution run is performed in a separate transaction that is rollbacked right after execution.
|
1.0
|
Perform every concrete execution in a separate transaction that is rollbacked right after execution - **Description**
Right now database content is reset between concrete executions with hardcoded SQL statements tuned for `spring-boot-testing-main` project database scheme (e.g. `ALTER TABLE ${repositoryDefinition.tableName} ALTER COLUMN id RESTART WITH 1`).
**Expected behavior**
Every concrete execution run is performed in a separate transaction that is rollbacked right after execution.
|
process
|
perform every concrete execution in a separate transaction that is rollbacked right after execution description right now database content is reset between concrete executions with hardcoded sql statements tuned for spring boot testing main project database scheme e g alter table repositorydefinition tablename alter column id restart with expected behavior every concrete execution run is performed in a separate transaction that is rollbacked right after execution
| 1
|
238,693
| 7,782,180,625
|
IssuesEvent
|
2018-06-06 05:03:17
|
tyarkoni/pliers
|
https://api.github.com/repos/tyarkoni/pliers
|
closed
|
Offset of elements gets added again if ComplexTextStim iterated twice
|
bug priority: high
|
Problematic line: `elem.onset = offset if elem.onset is None else offset + elem.onset`
|
1.0
|
Offset of elements gets added again if ComplexTextStim iterated twice - Problematic line: `elem.onset = offset if elem.onset is None else offset + elem.onset`
|
non_process
|
offset of elements gets added again if complextextstim iterated twice problematic line elem onset offset if elem onset is none else offset elem onset
| 0
|
492,874
| 14,221,967,411
|
IssuesEvent
|
2020-11-17 16:19:22
|
StatCan/daaas
|
https://api.github.com/repos/StatCan/daaas
|
closed
|
I require write access to a private repo
|
area/security component/grafana priority/blocker
|
Hello,
I need to push a feature branch to the terraform-kubernetes-aks-platform-daaas-private repo.
It's specifically for integrating Grafana with Azure AD, using the prometheus operator.
Brendan mentioned that either one of you could grant me write access.
Thanks in advance!
Chuck
|
1.0
|
I require write access to a private repo - Hello,
I need to push a feature branch to the terraform-kubernetes-aks-platform-daaas-private repo.
It's specifically for integrating Grafana with Azure AD, using the prometheus operator.
Brendan mentioned that either one of you could grant me write access.
Thanks in advance!
Chuck
|
non_process
|
i require write access to a private repo hello i need to push a feature branch to the terraform kubernetes aks platform daaas private repo it s specifically for integrating grafana with azure ad using the prometheus operator brendan mentioned that either one of you could grant me write access thanks in advance chuck
| 0
|
22,734
| 10,766,017,909
|
IssuesEvent
|
2019-11-01 12:43:21
|
valdisiljuconoks/azfun-demo
|
https://api.github.com/repos/valdisiljuconoks/azfun-demo
|
opened
|
CVE-2019-13173 (High) detected in fstream-0.1.31.tgz
|
security vulnerability
|
## CVE-2019-13173 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-0.1.31.tgz</b></p></summary>
<p>Advanced file system stream things</p>
<p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz">https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz</a></p>
<p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/fstream/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- digdug-1.4.0.tgz
- decompress-0.2.3.tgz
- tar-0.1.20.tgz
- :x: **fstream-0.1.31.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/valdisiljuconoks/azfun-demo/commit/d6e21882f73fdf25bf4853c5914f2a50f94b2086">d6e21882f73fdf25bf4853c5914f2a50f94b2086</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable.
<p>Publish Date: 2019-07-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p>
<p>Release Date: 2019-07-02</p>
<p>Fix Resolution: 1.0.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-13173 (High) detected in fstream-0.1.31.tgz - ## CVE-2019-13173 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>fstream-0.1.31.tgz</b></p></summary>
<p>Advanced file system stream things</p>
<p>Library home page: <a href="https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz">https://registry.npmjs.org/fstream/-/fstream-0.1.31.tgz</a></p>
<p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_2bf874e3-c2e7-460c-b7d7-ae88880164d2/20191101123739_65547/ws-scm_depth_0/azfun-demo/Web1/modules/_protected/Shell/Shell/10.10.6.0/ClientResources/lib/xstyle/node_modules/fstream/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- digdug-1.4.0.tgz
- decompress-0.2.3.tgz
- tar-0.1.20.tgz
- :x: **fstream-0.1.31.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/valdisiljuconoks/azfun-demo/commit/d6e21882f73fdf25bf4853c5914f2a50f94b2086">d6e21882f73fdf25bf4853c5914f2a50f94b2086</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
fstream before 1.0.12 is vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink, will overwrite the system's file with the contents of the extracted file. The fstream.DirWriter() function is vulnerable.
<p>Publish Date: 2019-07-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173>CVE-2019-13173</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-13173</a></p>
<p>Release Date: 2019-07-02</p>
<p>Fix Resolution: 1.0.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in fstream tgz cve high severity vulnerability vulnerable library fstream tgz advanced file system stream things library home page a href path to dependency file tmp whitesource archiveextractor ws scm depth azfun demo modules protected shell shell clientresources lib xstyle package json path to vulnerable library tmp whitesource archiveextractor ws scm depth azfun demo modules protected shell shell clientresources lib xstyle node modules fstream package json dependency hierarchy intern geezer tgz root library digdug tgz decompress tgz tar tgz x fstream tgz vulnerable library found in head commit a href vulnerability details fstream before is vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file the fstream dirwriter function is vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
16,201
| 20,713,571,526
|
IssuesEvent
|
2022-03-12 09:11:23
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
PhpExecutableFinder may return directory instead of binary
|
Bug Process Status: Needs Review
|
### Symfony version(s) affected
3.4 | 4.4
### Description
We're using \Symfony\Component\Process\PhpExecutableFinder in Drupal's Upgrade Status module to execute PHPStan from a UI. In different environments leveraging Plesk or ISPConfig, the returned path from \Symfony\Component\Process\PhpExecutableFinder::find is a directory.
### How to reproduce
I haven't been able to reproduce since I do not have access to a Plesk server. But directories are technically executable.
### Possible Solution
The checks for `is_exectuable` may need a `!is_dir` check?
### Additional Context
_No response_
|
1.0
|
PhpExecutableFinder may return directory instead of binary - ### Symfony version(s) affected
3.4 | 4.4
### Description
We're using \Symfony\Component\Process\PhpExecutableFinder in Drupal's Upgrade Status module to execute PHPStan from a UI. In different environments leveraging Plesk or ISPConfig, the returned path from \Symfony\Component\Process\PhpExecutableFinder::find is a directory.
### How to reproduce
I haven't been able to reproduce since I do not have access to a Plesk server. But directories are technically executable.
### Possible Solution
The checks for `is_exectuable` may need a `!is_dir` check?
### Additional Context
_No response_
|
process
|
phpexecutablefinder may return directory instead of binary symfony version s affected description we re using symfony component process phpexecutablefinder in drupal s upgrade status module to execute phpstan from a ui in different environments leveraging plesk or ispconfig the returned path from symfony component process phpexecutablefinder find is a directory how to reproduce i haven t been able to reproduce since i do not have access to a plesk server but directories are technically executable possible solution the checks for is exectuable may need a is dir check additional context no response
| 1
|
20,627
| 27,299,815,214
|
IssuesEvent
|
2023-02-24 00:19:25
|
googleapis/google-cloud-node
|
https://api.github.com/repos/googleapis/google-cloud-node
|
closed
|
Turn off trusted-contributor bot for release-please
|
type: process
|
We expect release-please to keep an open PR with pending changes to be released. We do not need to run samples/system tests until we are ready to publish the packages. Otherwise, we will be running so many unnecessary integration tests when we are not ready to release.
When a maintainer is ready to release:
1. approve the PR
2. comment with `/gcbrun`
3. add the `automerge` label
|
1.0
|
Turn off trusted-contributor bot for release-please - We expect release-please to keep an open PR with pending changes to be released. We do not need to run samples/system tests until we are ready to publish the packages. Otherwise, we will be running so many unnecessary integration tests when we are not ready to release.
When a maintainer is ready to release:
1. approve the PR
2. comment with `/gcbrun`
3. add the `automerge` label
|
process
|
turn off trusted contributor bot for release please we expect release please to keep an open pr with pending changes to be released we do not need to run samples system tests until we are ready to publish the packages otherwise we will be running so many unnecessary integration tests when we are not ready to release when a maintainer is ready to release approve the pr comment with gcbrun add the automerge label
| 1
|
15,610
| 19,749,433,830
|
IssuesEvent
|
2022-01-15 00:08:31
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
MDBGA: Prevent users from trying to use referential actions on inline m-n's
|
process/candidate topic: schema validation team/migrations topic: mongodb topic: referential actions kind/subtask
|
_(This issue is a follow-up to: https://github.com/prisma/prisma/issues/10498)_
Currently, the Prisma Schema allows:
```prisma
model A {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
gql String?
b_ids String[] @db.Array(ObjectId)
bs B[] @relation(fields: [b_ids], onDelete: Cascade)
}
model B {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
gql String?
a_ids String[] @db.Array(ObjectId)
as A[] @relation(fields: [a_ids], onDelete: Cascade)
}
```
But we don't support referential actions on m-ns for MongoDB. We should:
- Invalidate this in prisma format and provide a helpful error. Perhaps link to the documentation to learn how to workaround it.
- Document this limitation in [our documentation](https://www.prisma.io/docs/concepts/components/prisma-schema/relations/many-to-many-relations#mongodb). We can tell people to use an explicit join table as a workaround.
Later on, we can consider adding referential action support to this type of schema.
|
1.0
|
MDBGA: Prevent users from trying to use referential actions on inline m-n's - _(This issue is a follow-up to: https://github.com/prisma/prisma/issues/10498)_
Currently, the Prisma Schema allows:
```prisma
model A {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
gql String?
b_ids String[] @db.Array(ObjectId)
bs B[] @relation(fields: [b_ids], onDelete: Cascade)
}
model B {
id String @id @default(dbgenerated()) @map("_id") @db.ObjectId
gql String?
a_ids String[] @db.Array(ObjectId)
as A[] @relation(fields: [a_ids], onDelete: Cascade)
}
```
But we don't support referential actions on m-ns for MongoDB. We should:
- Invalidate this in prisma format and provide a helpful error. Perhaps link to the documentation to learn how to workaround it.
- Document this limitation in [our documentation](https://www.prisma.io/docs/concepts/components/prisma-schema/relations/many-to-many-relations#mongodb). We can tell people to use an explicit join table as a workaround.
Later on, we can consider adding referential action support to this type of schema.
|
process
|
mdbga prevent users from trying to use referential actions on inline m n s this issue is a follow up to currently the prisma schema allows prisma model a id string id default dbgenerated map id db objectid gql string b ids string db array objectid bs b relation fields ondelete cascade model b id string id default dbgenerated map id db objectid gql string a ids string db array objectid as a relation fields ondelete cascade but we don t support referential actions on m ns for mongodb we should invalidate this in prisma format and provide a helpful error perhaps link to the documentation to learn how to workaround it document this limitation in we can tell people to use an explicit join table as a workaround later on we can consider adding referential action support to this type of schema
| 1
|
672
| 3,145,591,116
|
IssuesEvent
|
2015-09-14 18:44:53
|
K0zka/kerub
|
https://api.github.com/repos/K0zka/kerub
|
opened
|
Add optional description fields to entities
|
component:data processing component:ui priority: minimal
|
Markdown formatted description for the following entities:
* Network
* VM
* Virtual Disk
* Expectations
(followup of #81 )
|
1.0
|
Add optional description fields to entities - Markdown formatted description for the following entities:
* Network
* VM
* Virtual Disk
* Expectations
(followup of #81 )
|
process
|
add optional description fields to entities markdown formatted description for the following entities network vm virtual disk expectations followup of
| 1
|
3,502
| 6,557,593,319
|
IssuesEvent
|
2017-09-06 18:01:36
|
w3c/webauthn
|
https://api.github.com/repos/w3c/webauthn
|
closed
|
The W3C HTML spec is broken, and probably shouldn't be referenced
|
type:process type:technical
|
I was pretty confused by https://www.w3.org/TR/html51/browsers.html#relaxing-the-same-origin-restriction step 7.2, as referenced from https://w3c.github.io/webauthn/#dom-webauthentication-makecredential step 3, because it's completely nonsensical: it causes the algorithm to _always_ throw.
Then I realized that this is the W3C version of the HTML spec, which has all sorts of known issues like this. For comparison, https://html.spec.whatwg.org/multipage/browsers.html#relaxing-the-same-origin-restriction step 7.2 is correct...
Please reference the things that actually have the behavior you want. I guess if you have to reference W3C HTML, that makes it your problem to check that it has the behavior you want and push on its editors to fix it if it does not.
|
1.0
|
The W3C HTML spec is broken, and probably shouldn't be referenced - I was pretty confused by https://www.w3.org/TR/html51/browsers.html#relaxing-the-same-origin-restriction step 7.2, as referenced from https://w3c.github.io/webauthn/#dom-webauthentication-makecredential step 3, because it's completely nonsensical: it causes the algorithm to _always_ throw.
Then I realized that this is the W3C version of the HTML spec, which has all sorts of known issues like this. For comparison, https://html.spec.whatwg.org/multipage/browsers.html#relaxing-the-same-origin-restriction step 7.2 is correct...
Please reference the things that actually have the behavior you want. I guess if you have to reference W3C HTML, that makes it your problem to check that it has the behavior you want and push on its editors to fix it if it does not.
|
process
|
the html spec is broken and probably shouldn t be referenced i was pretty confused by step as referenced from step because it s completely nonsensical it causes the algorithm to always throw then i realized that this is the version of the html spec which has all sorts of known issues like this for comparison step is correct please reference the things that actually have the behavior you want i guess if you have to reference html that makes it your problem to check that it has the behavior you want and push on its editors to fix it if it does not
| 1
|
11,199
| 13,957,702,812
|
IssuesEvent
|
2020-10-24 08:13:43
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
CZ: Harvesting to thematic viewer
|
CZ - Czech Republic Geoportal Harvesting process
|
From: Jitka.Faugnerova@cenia.cz [Jitka.Faugnerova@cenia.cz]
Sent: 14 May 2018 14:03
To: TOMAS Robert (JRC-ISPRA)
Subject: Harvesting to thematic viewer
Dear Robert,
could you please check with Angelo or ask him to do the harvesting to your geoportal/viewer?
One of our data provider has made some changes in metadata regarding the priority data sets and it seems that last time the metadata was harvested is 09/05/2018. Deadline given by the letter from December 2018 is tomorrow, so we would like to see, whether we have succeeded in making data resources available to the Thematic Viewer. Of course we would like to score better than now :-)
Thank you very much in advance.
Best regards,
Jitka
Jitka Faugnerová
CENIA, ?eská informa?ní agentura ?ivotního prost?edí
|
1.0
|
CZ: Harvesting to thematic viewer - From: Jitka.Faugnerova@cenia.cz [Jitka.Faugnerova@cenia.cz]
Sent: 14 May 2018 14:03
To: TOMAS Robert (JRC-ISPRA)
Subject: Harvesting to thematic viewer
Dear Robert,
could you please check with Angelo or ask him to do the harvesting to your geoportal/viewer?
One of our data provider has made some changes in metadata regarding the priority data sets and it seems that last time the metadata was harvested is 09/05/2018. Deadline given by the letter from December 2018 is tomorrow, so we would like to see, whether we have succeeded in making data resources available to the Thematic Viewer. Of course we would like to score better than now :-)
Thank you very much in advance.
Best regards,
Jitka
Jitka Faugnerová
CENIA, ?eská informa?ní agentura ?ivotního prost?edí
|
process
|
cz harvesting to thematic viewer from jitka faugnerova cenia cz sent may to tomas robert jrc ispra subject harvesting to thematic viewer dear robert could you please check with angelo or ask him to do the harvesting to your geoportal viewer one of our data provider has made some changes in metadata regarding the priority data sets and it seems that last time the metadata was harvested is deadline given by the letter from december is tomorrow so we would like to see whether we have succeeded in making data resources available to the thematic viewer of course we would like to score better than now thank you very much in advance best regards jitka jitka faugnerov aacute cenia esk aacute informa n iacute agentura ivotn iacute ho prost ed iacute
| 1
|
582,078
| 17,352,338,381
|
IssuesEvent
|
2021-07-29 10:16:49
|
hedgedoc/hedgedoc
|
https://api.github.com/repos/hedgedoc/hedgedoc
|
closed
|
no link between slides
|
priority: high scope: frontend status: confirmed type: bug
|
**Description**
When presenting slides, the feature to create links between slides does not work.
**To Reproduce**
1. Go to demo slides (of the demo instance)
2. Go to [page 18](https://demo.hedgedoc.org/p/slide-example#/17)
3. Click on link "like this"
4. See that the browser is redirected to home page
**Expected behavior**
The link should lead to slide `1/3`.
**Logs**
n/a
**Config**
n/a
**Setup:**
- demo instance
**Additional context**
n/a
|
1.0
|
no link between slides - **Description**
When presenting slides, the feature to create links between slides does not work.
**To Reproduce**
1. Go to demo slides (of the demo instance)
2. Go to [page 18](https://demo.hedgedoc.org/p/slide-example#/17)
3. Click on link "like this"
4. See that the browser is redirected to home page
**Expected behavior**
The link should lead to slide `1/3`.
**Logs**
n/a
**Config**
n/a
**Setup:**
- demo instance
**Additional context**
n/a
|
non_process
|
no link between slides description when presenting slides the feature to create links between slides does not work to reproduce go to demo slides of the demo instance go to click on link like this see that the browser is redirected to home page expected behavior the link should lead to slide logs n a config n a setup demo instance additional context n a
| 0
|
10,139
| 13,044,162,440
|
IssuesEvent
|
2020-07-29 03:47:32
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `JsonArrayAppendSig` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `JsonArrayAppendSig` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `JsonArrayAppendSig` from TiDB -
## Description
Port the scalar function `JsonArrayAppendSig` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function jsonarrayappendsig from tidb description port the scalar function jsonarrayappendsig from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
14,645
| 25,331,436,165
|
IssuesEvent
|
2022-11-18 13:34:36
|
AAFC-BICoE/dina-planning
|
https://api.github.com/repos/AAFC-BICoE/dina-planning
|
closed
|
Allow to add material-sample to transaction
|
type: requirement
|
Goal: allow to search for material-sample and add them to a transaction.
Out-of-scope for this ticket: show that a material-sample is not available since it's part of a transaction.
|
1.0
|
Allow to add material-sample to transaction - Goal: allow to search for material-sample and add them to a transaction.
Out-of-scope for this ticket: show that a material-sample is not available since it's part of a transaction.
|
non_process
|
allow to add material sample to transaction goal allow to search for material sample and add them to a transaction out of scope for this ticket show that a material sample is not available since it s part of a transaction
| 0
|
31,668
| 15,036,210,141
|
IssuesEvent
|
2021-02-02 15:00:10
|
Automattic/wp-calypso
|
https://api.github.com/repos/Automattic/wp-calypso
|
closed
|
PlansGrid: API gets called twice
|
Performance [Goal] New Onboarding [Pri] High [Type] Bug
|
Probably after merging https://github.com/Automattic/wp-calypso/pull/48790, when loading PlansGrid in both Gutenboarding and launch flows (when sandboxed), the REST API calls for `/plans` and `/details` are each done 2 times.
<img width="731" alt="Screenshot 2021-01-21 at 20 52 03" src="https://user-images.githubusercontent.com/14192054/105399264-0d83f400-5c2c-11eb-8877-bdb0277b1307.png">
|
True
|
PlansGrid: API gets called twice - Probably after merging https://github.com/Automattic/wp-calypso/pull/48790, when loading PlansGrid in both Gutenboarding and launch flows (when sandboxed), the REST API calls for `/plans` and `/details` are each done 2 times.
<img width="731" alt="Screenshot 2021-01-21 at 20 52 03" src="https://user-images.githubusercontent.com/14192054/105399264-0d83f400-5c2c-11eb-8877-bdb0277b1307.png">
|
non_process
|
plansgrid api gets called twice probably after merging when loading plansgrid in both gutenboarding and launch flows when sandboxed the rest api calls for plans and details are each done times img width alt screenshot at src
| 0
|
94,120
| 15,962,337,510
|
IssuesEvent
|
2021-04-16 01:05:14
|
RG4421/nucleus
|
https://api.github.com/repos/RG4421/nucleus
|
opened
|
WS-2017-0266 (Low) detected in http-signature-0.10.1.tgz
|
security vulnerability
|
## WS-2017-0266 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-signature-0.10.1.tgz</b></p></summary>
<p>Reference implementation of Joyent's HTTP Signature scheme.</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz">https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz</a></p>
<p>Path to dependency file: nucleus/packages/@nucleus/package.json</p>
<p>Path to vulnerable library: nucleus/packages/@nucleus/node_modules/yui/node_modules/http-signature/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-addon-docs-0.6.16.tgz (Root Library)
- yuidocjs-0.10.2.tgz
- yui-3.18.1.tgz
- request-2.40.0.tgz
- :x: **http-signature-0.10.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks.
<p>Publish Date: 2015-01-22
<p>URL: <a href=https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56>WS-2017-0266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/joyent/node-http-signature/pull/36">https://github.com/joyent/node-http-signature/pull/36</a></p>
<p>Release Date: 2017-01-31</p>
<p>Fix Resolution: 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"http-signature","packageVersion":"0.10.1","packageFilePaths":["/packages/@nucleus/package.json"],"isTransitiveDependency":true,"dependencyTree":"ember-cli-addon-docs:0.6.16;yuidocjs:0.10.2;yui:3.18.1;request:2.40.0;http-signature:0.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0266","vulnerabilityDetails":"Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks.","vulnerabilityUrl":"https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56","cvss3Severity":"low","cvss3Score":"3.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Adjacent","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2017-0266 (Low) detected in http-signature-0.10.1.tgz - ## WS-2017-0266 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-signature-0.10.1.tgz</b></p></summary>
<p>Reference implementation of Joyent's HTTP Signature scheme.</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz">https://registry.npmjs.org/http-signature/-/http-signature-0.10.1.tgz</a></p>
<p>Path to dependency file: nucleus/packages/@nucleus/package.json</p>
<p>Path to vulnerable library: nucleus/packages/@nucleus/node_modules/yui/node_modules/http-signature/package.json</p>
<p>
Dependency Hierarchy:
- ember-cli-addon-docs-0.6.16.tgz (Root Library)
- yuidocjs-0.10.2.tgz
- yui-3.18.1.tgz
- request-2.40.0.tgz
- :x: **http-signature-0.10.1.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks.
<p>Publish Date: 2015-01-22
<p>URL: <a href=https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56>WS-2017-0266</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/joyent/node-http-signature/pull/36">https://github.com/joyent/node-http-signature/pull/36</a></p>
<p>Release Date: 2017-01-31</p>
<p>Fix Resolution: 1.0.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"http-signature","packageVersion":"0.10.1","packageFilePaths":["/packages/@nucleus/package.json"],"isTransitiveDependency":true,"dependencyTree":"ember-cli-addon-docs:0.6.16;yuidocjs:0.10.2;yui:3.18.1;request:2.40.0;http-signature:0.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0266","vulnerabilityDetails":"Affected versions (before 1.0.0) of the http-signature package are vulnerable to Timing Attacks.","vulnerabilityUrl":"https://github.com/joyent/node-http-signature/commit/78ab1da232f31f695f5c362d863593a143aa8b56","cvss3Severity":"low","cvss3Score":"3.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Adjacent","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
ws low detected in http signature tgz ws low severity vulnerability vulnerable library http signature tgz reference implementation of joyent s http signature scheme library home page a href path to dependency file nucleus packages nucleus package json path to vulnerable library nucleus packages nucleus node modules yui node modules http signature package json dependency hierarchy ember cli addon docs tgz root library yuidocjs tgz yui tgz request tgz x http signature tgz vulnerable library found in base branch master vulnerability details affected versions before of the http signature package are vulnerable to timing attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree ember cli addon docs yuidocjs yui request http signature isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier ws vulnerabilitydetails affected versions before of the http signature package are vulnerable to timing attacks vulnerabilityurl
| 0
|
193,982
| 15,394,038,524
|
IssuesEvent
|
2021-03-03 17:24:52
|
PawelDabrowski83/genAlgo
|
https://api.github.com/repos/PawelDabrowski83/genAlgo
|
closed
|
Main description
|
documentation
|
Main:
<problem - w ilu pokoleniach uda się uzyskać gen o wartości 1234>
GenePoolService(wszystkie składowe)
GenePool(wartość szukana 1234, liczba genów, genePoolService).solve() -> int z rozwiązaniem
|
1.0
|
Main description - Main:
<problem - w ilu pokoleniach uda się uzyskać gen o wartości 1234>
GenePoolService(wszystkie składowe)
GenePool(wartość szukana 1234, liczba genów, genePoolService).solve() -> int z rozwiązaniem
|
non_process
|
main description main genepoolservice wszystkie składowe genepool wartość szukana liczba genów genepoolservice solve int z rozwiązaniem
| 0
|
60,723
| 17,023,503,209
|
IssuesEvent
|
2021-07-03 02:21:37
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
strange routing decisions
|
Component: gosmore Priority: major Resolution: worksforme Type: defect
|
**[Submitted to the original trac issue database at 12.35pm, Saturday, 7th November 2009]**
Latest builds of Gosmore seem to take some strange routing decisions. The three cases I have seen so far all occur when a roundabout is involved.
Attached are screenshots of Gosmore routing decisions together with a text file showing the route as a www.yournavigation.org link (whilst I appreciate yournavigation doesn't use the current Gosmore version, providing these links seemed to me the best way of summarising the route I put into Gosmore).
|
1.0
|
strange routing decisions - **[Submitted to the original trac issue database at 12.35pm, Saturday, 7th November 2009]**
Latest builds of Gosmore seem to take some strange routing decisions. The three cases I have seen so far all occur when a roundabout is involved.
Attached are screenshots of Gosmore routing decisions together with a text file showing the route as a www.yournavigation.org link (whilst I appreciate yournavigation doesn't use the current Gosmore version, providing these links seemed to me the best way of summarising the route I put into Gosmore).
|
non_process
|
strange routing decisions latest builds of gosmore seem to take some strange routing decisions the three cases i have seen so far all occur when a roundabout is involved attached are screenshots of gosmore routing decisions together with a text file showing the route as a link whilst i appreciate yournavigation doesn t use the current gosmore version providing these links seemed to me the best way of summarising the route i put into gosmore
| 0
|
194,580
| 22,262,036,522
|
IssuesEvent
|
2022-06-10 02:01:08
|
Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034
|
https://api.github.com/repos/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034
|
reopened
|
CVE-2021-33033 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2021-33033 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before 5.11.14 has a use-after-free in cipso_v4_genopt in net/ipv4/cipso_ipv4.c because the CIPSO and CALIPSO refcounting for the DOI definitions is mishandled, aka CID-ad5d07f4a9cd. This leads to writing an arbitrary value.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33033>CVE-2021-33033</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-33033">https://www.linuxkernelcves.com/cves/CVE-2021-33033</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: v4.14.226, v4.19.181, v5.4.106, v5.10.24, v5.11.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-33033 (High) detected in multiple libraries - ## CVE-2021-33033 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b>, <b>linuxlinux-4.19.239</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before 5.11.14 has a use-after-free in cipso_v4_genopt in net/ipv4/cipso_ipv4.c because the CIPSO and CALIPSO refcounting for the DOI definitions is mishandled, aka CID-ad5d07f4a9cd. This leads to writing an arbitrary value.
<p>Publish Date: 2021-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33033>CVE-2021-33033</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-33033">https://www.linuxkernelcves.com/cves/CVE-2021-33033</a></p>
<p>Release Date: 2021-05-14</p>
<p>Fix Resolution: v4.14.226, v4.19.181, v5.4.106, v5.10.24, v5.11.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries linuxlinux linuxlinux linuxlinux vulnerability details the linux kernel before has a use after free in cipso genopt in net cipso c because the cipso and calipso refcounting for the doi definitions is mishandled aka cid this leads to writing an arbitrary value publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
240,662
| 18,364,263,303
|
IssuesEvent
|
2021-10-09 19:30:57
|
malodetz/Fl-project
|
https://api.github.com/repos/malodetz/Fl-project
|
closed
|
Поправить ConcreteSyntax.md
|
documentation good first issue
|
Кажется, что литералы и названия ключевых функций не являются ключевыми словами. Оставить в этой секции только управляющие конструкции и создать новую секцию: "Зарезервированные слова"
|
1.0
|
Поправить ConcreteSyntax.md - Кажется, что литералы и названия ключевых функций не являются ключевыми словами. Оставить в этой секции только управляющие конструкции и создать новую секцию: "Зарезервированные слова"
|
non_process
|
поправить concretesyntax md кажется что литералы и названия ключевых функций не являются ключевыми словами оставить в этой секции только управляющие конструкции и создать новую секцию зарезервированные слова
| 0
|
13,384
| 15,862,500,366
|
IssuesEvent
|
2021-04-08 11:41:09
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Update esbuild
|
kind/tech process/candidate team/client
|
Unfortunately, the latest esbuild is not compatible with webpack.
That's why we had to downgrade it.
We should investigate the problem and make sure we can upgrade to the latest esbuild.
https://github.com/prisma/prisma/pull/6260
|
1.0
|
Update esbuild - Unfortunately, the latest esbuild is not compatible with webpack.
That's why we had to downgrade it.
We should investigate the problem and make sure we can upgrade to the latest esbuild.
https://github.com/prisma/prisma/pull/6260
|
process
|
update esbuild unfortunately the latest esbuild is not compatible with webpack that s why we had to downgrade it we should investigate the problem and make sure we can upgrade to the latest esbuild
| 1
|
21,628
| 30,030,840,395
|
IssuesEvent
|
2023-06-27 09:28:51
|
unicode-org/icu4x
|
https://api.github.com/repos/unicode-org/icu4x
|
opened
|
ffi CI job is slow all of a sudden
|
T-bug C-process
|
It gets stuck here:
```
...
2023-06-27T06:21:24.1339208Z cp target/wasm32-unknown-unknown/release/icu_capi_cdylib.wasm icu_capi.wasm
2023-06-27T06:21:24.1355342Z cargo run -p icu_datagen -- \
2023-06-27T06:21:24.1355873Z --keys-for-bin icu_capi.wasm \
2023-06-27T06:21:24.1356257Z --locales full \
2023-06-27T06:21:24.1356610Z --format blob \
2023-06-27T06:21:24.1357007Z --out data.postcard
2023-06-27T06:21:24.7096518Z Finished dev [unoptimized + debuginfo] target(s) in 0.53s
2023-06-27T06:21:24.7154466Z Running `/home/runner/work/icu4x/icu4x/target/debug/icu4x-datagen --keys-for-bin icu_capi.wasm --locales full --format blob --out data.postcard`
2023-06-27T09:18:34.1380351Z INFO [icu_provider_blob::export::blob_exporter] Serializing blob to output stream...
2023-06-27T09:18:34.2128602Z rm -rf lib;
2023-06-27T09:18:34.2148391Z mkdir lib;
2023-06-27T09:18:34.2167144Z cp ../../include/* lib
...
```
The first commit exhibiting this behavior is https://github.com/unicode-org/icu4x/commit/c394c1c6c121e15b05aa7b0aa98826c0cf965694 but it's unclear why that would be at fault.
|
1.0
|
ffi CI job is slow all of a sudden - It gets stuck here:
```
...
2023-06-27T06:21:24.1339208Z cp target/wasm32-unknown-unknown/release/icu_capi_cdylib.wasm icu_capi.wasm
2023-06-27T06:21:24.1355342Z cargo run -p icu_datagen -- \
2023-06-27T06:21:24.1355873Z --keys-for-bin icu_capi.wasm \
2023-06-27T06:21:24.1356257Z --locales full \
2023-06-27T06:21:24.1356610Z --format blob \
2023-06-27T06:21:24.1357007Z --out data.postcard
2023-06-27T06:21:24.7096518Z Finished dev [unoptimized + debuginfo] target(s) in 0.53s
2023-06-27T06:21:24.7154466Z Running `/home/runner/work/icu4x/icu4x/target/debug/icu4x-datagen --keys-for-bin icu_capi.wasm --locales full --format blob --out data.postcard`
2023-06-27T09:18:34.1380351Z INFO [icu_provider_blob::export::blob_exporter] Serializing blob to output stream...
2023-06-27T09:18:34.2128602Z rm -rf lib;
2023-06-27T09:18:34.2148391Z mkdir lib;
2023-06-27T09:18:34.2167144Z cp ../../include/* lib
...
```
The first commit exhibiting this behavior is https://github.com/unicode-org/icu4x/commit/c394c1c6c121e15b05aa7b0aa98826c0cf965694 but it's unclear why that would be at fault.
|
process
|
ffi ci job is slow all of a sudden it gets stuck here cp target unknown unknown release icu capi cdylib wasm icu capi wasm cargo run p icu datagen keys for bin icu capi wasm locales full format blob out data postcard finished dev target s in running home runner work target debug datagen keys for bin icu capi wasm locales full format blob out data postcard info serializing blob to output stream rm rf lib mkdir lib cp include lib the first commit exhibiting this behavior is but it s unclear why that would be at fault
| 1
|
202,767
| 15,300,100,127
|
IssuesEvent
|
2021-02-24 11:51:03
|
tempesta-tech/tempesta
|
https://api.github.com/repos/tempesta-tech/tempesta
|
closed
|
HTTPtables blocking action ignores block_action directive
|
bug good to start test
|
# Scope
On configuration
```
listen 192.168.100.4:80;
block_action attack reply;
block_action error reply;
response_body 400 /var/www/tempesta-tech.com/400;
srv_group default {
server 127.0.0.1:9090;
}
vhost debian {
proxy_pass default;
}
cache 1;
cache_fulfill * *;
http_chain l7_rules {
hdr "Referer" == "http://badhost.com*" -> block;
}
http_chain {
mark == 1 -> l7_rules;
-> debian;
}
```
Note that there is no default host as well no default rule for `l7_rules` chain. Configure iptables as:
```
iptables -A INPUT -s 192.168.100.1 -j MARK --set-mark 1
```
Next try to send 2 requests from `192.168.100.1`:
1. with `Referer: http://goodhost.com`
2. with `Referer: http://badhost.com/foo`
In both the cases HTTPtables block the requests, but Tempesta does not send any HTTP response as expected. From browsing the code it seems we just ignore `block` argument returned from HTTPtables.
# Testing
For the scenario and configuration as the above:
- [ ] request with `Referer: http://goodhost.com`
- [ ] request with `Referer: http://badhost.com/foo`
- [ ] Check that configuration option `response_body` also does the work. In my demo scenario I used
```
block_action attack reply;
block_action error reply;
response_body 400 /var/www/tempesta-tech.com/400;
```
Please also add tests for the configuration described in https://github.com/tempesta-tech/tempesta/wiki/HTTP-tables#default-rules-and-chains to `./http_rules/test_http_tables.py`
- [ ] Drop a request if there is no matching rule in a chain
- [ ] implicit creation of the main chain
- [ ] Dropping all requests with an empy main chain
|
1.0
|
HTTPtables blocking action ignores block_action directive - # Scope
On configuration
```
listen 192.168.100.4:80;
block_action attack reply;
block_action error reply;
response_body 400 /var/www/tempesta-tech.com/400;
srv_group default {
server 127.0.0.1:9090;
}
vhost debian {
proxy_pass default;
}
cache 1;
cache_fulfill * *;
http_chain l7_rules {
hdr "Referer" == "http://badhost.com*" -> block;
}
http_chain {
mark == 1 -> l7_rules;
-> debian;
}
```
Note that there is no default host as well no default rule for `l7_rules` chain. Configure iptables as:
```
iptables -A INPUT -s 192.168.100.1 -j MARK --set-mark 1
```
Next try to send 2 requests from `192.168.100.1`:
1. with `Referer: http://goodhost.com`
2. with `Referer: http://badhost.com/foo`
In both the cases HTTPtables block the requests, but Tempesta does not send any HTTP response as expected. From browsing the code it seems we just ignore `block` argument returned from HTTPtables.
# Testing
For the scenario and configuration as the above:
- [ ] request with `Referer: http://goodhost.com`
- [ ] request with `Referer: http://badhost.com/foo`
- [ ] Check that configuration option `response_body` also does the work. In my demo scenario I used
```
block_action attack reply;
block_action error reply;
response_body 400 /var/www/tempesta-tech.com/400;
```
Please also add tests for the configuration described in https://github.com/tempesta-tech/tempesta/wiki/HTTP-tables#default-rules-and-chains to `./http_rules/test_http_tables.py`
- [ ] Drop a request if there is no matching rule in a chain
- [ ] implicit creation of the main chain
- [ ] Dropping all requests with an empy main chain
|
non_process
|
httptables blocking action ignores block action directive scope on configuration listen block action attack reply block action error reply response body var www tempesta tech com srv group default server vhost debian proxy pass default cache cache fulfill http chain rules hdr referer block http chain mark rules debian note that there is no default host as well no default rule for rules chain configure iptables as iptables a input s j mark set mark next try to send requests from with referer with referer in both the cases httptables block the requests but tempesta does not send any http response as expected from browsing the code it seems we just ignore block argument returned from httptables testing for the scenario and configuration as the above request with referer request with referer check that configuration option response body also does the work in my demo scenario i used block action attack reply block action error reply response body var www tempesta tech com please also add tests for the configuration described in to http rules test http tables py drop a request if there is no matching rule in a chain implicit creation of the main chain dropping all requests with an empy main chain
| 0
|
87,162
| 3,737,636,218
|
IssuesEvent
|
2016-03-08 19:58:02
|
aces/cbrain
|
https://api.github.com/repos/aces/cbrain
|
closed
|
Launch Task of the select tool overlay bug.
|
High Priority
|
It redirect to index page, don't know if it should be the case.
I think:
1) Should stay on the overlay.
2) Should redirect on the new task page.
3) Should definitively not put the launch at the bottom of the index table. Like this:

|
1.0
|
Launch Task of the select tool overlay bug. - It redirect to index page, don't know if it should be the case.
I think:
1) Should stay on the overlay.
2) Should redirect on the new task page.
3) Should definitively not put the launch at the bottom of the index table. Like this:

|
non_process
|
launch task of the select tool overlay bug it redirect to index page don t know if it should be the case i think should stay on the overlay should redirect on the new task page should definitively not put the launch at the bottom of the index table like this
| 0
|
469
| 2,731,262,838
|
IssuesEvent
|
2015-04-16 19:19:24
|
mozilla/id.webmaker.org
|
https://api.github.com/repos/mozilla/id.webmaker.org
|
closed
|
Implement CSRF
|
Security
|
- [ ] find out how this will work with the react router
- [ ] should do csrf on per-request basis (rather than per-page-load)?
- is this really more secure?
- hash per request or secret per request?
|
True
|
Implement CSRF - - [ ] find out how this will work with the react router
- [ ] should do csrf on per-request basis (rather than per-page-load)?
- is this really more secure?
- hash per request or secret per request?
|
non_process
|
implement csrf find out how this will work with the react router should do csrf on per request basis rather than per page load is this really more secure hash per request or secret per request
| 0
|
269,429
| 23,442,060,951
|
IssuesEvent
|
2022-08-15 15:47:30
|
RPTools/maptool
|
https://api.github.com/repos/RPTools/maptool
|
closed
|
[Bug]: occasional NullPointerException on local WebRTC client connection
|
bug tested
|
### Describe the Bug
After starting a WebRTC server it failed to start properly.
```
18:27:52.457 (ClientHandshake.java:145) [Thread-16] INFO - skullman :send: CLIENT_INIT_MSG
Exception in thread "Thread-16" java.lang.NullPointerException: Cannot invoke "dev.onvoid.webrtc.RTCPeerConnection.getConnectionState()" because "this.peerConnection" is null
at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:185)
at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:178)
at net.rptools.clientserver.hessian.client.MethodClientDecorator.sendMessage(MethodClientDecorator.java:42)
at net.rptools.maptool.server.ClientHandshake.sendMessage(ClientHandshake.java:146)
at net.rptools.maptool.server.ClientHandshake.startHandshake(ClientHandshake.java:139)
at net.rptools.maptool.client.MapToolConnection.start(MapToolConnection.java:76)
at net.rptools.maptool.client.MapTool.createConnection(MapTool.java:1227)
at net.rptools.maptool.client.AppActions$62.lambda$executeAction$1(AppActions.java:2326)
at net.rptools.maptool.client.AppActions$ClientAction.lambda$runBackground$0(AppActions.java:3496)
at java.base/java.lang.Thread.run(Thread.java:831)
18:27:52.546 (WebRTCServerConnection.java:70) [WebSocketConnectReadThread-81] INFO - S WebSocket connected
18:27:52.762 (WebRTCClientConnection.java:136) [WebSocketConnectReadThread-84] INFO - C WebSocket connected
18:27:52.922 (WebRTCClientConnection.java:408) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onRenegotiationNeeded
18:27:52.926 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: HAVE_LOCAL_OFFER
18:27:52.936 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange GATHERING
[sentry-pool-1-thread-2] WARN io.sentry.connection.AbstractConnection.lockdown - Initiated a temporary lockdown because of exception: {"detail":"bad sentry DSN public key"}
[sentry-pool-1-thread-2] ERROR io.sentry.connection.AsyncConnection - An exception occurred while sending the event to Sentry.
io.sentry.connection.ConnectionException: {"detail":"bad sentry DSN public key"}
at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:216)
at io.sentry.connection.AbstractConnection.send(AbstractConnection.java:88)
at io.sentry.connection.AsyncConnection$EventSubmitter.run(AsyncConnection.java:187)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630)
at java.base/java.lang.Thread.run(Thread.java:831)
Caused by: java.io.IOException: Server returned HTTP response code: 400 for URL: https://sentry.io/api/1404248/store/
at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1985)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1577)
at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getInputStream(HttpsURLConnectionImpl.java:224)
at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:172)
... 5 more
18:27:53.094 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: HAVE_REMOTE_OFFER
18:27:53.096 (WebRTCClientConnection.java:73) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S setRemoteDescription success.
18:27:53.098 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: STABLE
18:27:53.100 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange GATHERING
18:27:53.240 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: STABLE
18:27:53.240 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CHECKING
18:27:53.241 (WebRTCClientConnection.java:243) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C setRemoteDescription success.
18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTING
18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTING
18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange COMPLETE
18:27:53.257 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CHECKING
18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange COMPLETE
18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange COMPLETED
18:27:53.261 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C localDataChannel onStateChange OPEN
18:27:53.261 (WebRTCClientConnection.java:395) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onDataChannel
18:27:53.262 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S localDataChannel onStateChange OPEN
18:27:53.264 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_INIT_MSG
18:27:53.266 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: USE_AUTH_TYPE_MSG
18:27:53.267 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: USE_AUTH_TYPE_MSG
18:27:53.271 (ClientHandshake.java:145) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :send: CLIENT_AUTH_MESSAGE
18:27:53.272 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_AUTH_MESSAGE
18:27:53.272 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: CONNECTION_SUCCESSFUL_MSG
18:27:53.277 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: CONNECTION_SUCCESSFUL_MSG
```
### To Reproduce
Start a WebRTC server a bunch of times and get unlucky.
### Expected Behaviour
Local client is always able to send messages.
### Screenshots
_No response_
### MapTool Info
1.11.4 with WebRTC patch
### Desktop
Linux[Debian]
### Additional Context
_No response_
|
1.0
|
[Bug]: occasional NullPointerException on local WebRTC client connection - ### Describe the Bug
After starting a WebRTC server it failed to start properly.
```
18:27:52.457 (ClientHandshake.java:145) [Thread-16] INFO - skullman :send: CLIENT_INIT_MSG
Exception in thread "Thread-16" java.lang.NullPointerException: Cannot invoke "dev.onvoid.webrtc.RTCPeerConnection.getConnectionState()" because "this.peerConnection" is null
at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:185)
at net.rptools.clientserver.simple.client.WebRTCClientConnection.sendMessage(WebRTCClientConnection.java:178)
at net.rptools.clientserver.hessian.client.MethodClientDecorator.sendMessage(MethodClientDecorator.java:42)
at net.rptools.maptool.server.ClientHandshake.sendMessage(ClientHandshake.java:146)
at net.rptools.maptool.server.ClientHandshake.startHandshake(ClientHandshake.java:139)
at net.rptools.maptool.client.MapToolConnection.start(MapToolConnection.java:76)
at net.rptools.maptool.client.MapTool.createConnection(MapTool.java:1227)
at net.rptools.maptool.client.AppActions$62.lambda$executeAction$1(AppActions.java:2326)
at net.rptools.maptool.client.AppActions$ClientAction.lambda$runBackground$0(AppActions.java:3496)
at java.base/java.lang.Thread.run(Thread.java:831)
18:27:52.546 (WebRTCServerConnection.java:70) [WebSocketConnectReadThread-81] INFO - S WebSocket connected
18:27:52.762 (WebRTCClientConnection.java:136) [WebSocketConnectReadThread-84] INFO - C WebSocket connected
18:27:52.922 (WebRTCClientConnection.java:408) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onRenegotiationNeeded
18:27:52.926 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: HAVE_LOCAL_OFFER
18:27:52.936 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange GATHERING
[sentry-pool-1-thread-2] WARN io.sentry.connection.AbstractConnection.lockdown - Initiated a temporary lockdown because of exception: {"detail":"bad sentry DSN public key"}
[sentry-pool-1-thread-2] ERROR io.sentry.connection.AsyncConnection - An exception occurred while sending the event to Sentry.
io.sentry.connection.ConnectionException: {"detail":"bad sentry DSN public key"}
at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:216)
at io.sentry.connection.AbstractConnection.send(AbstractConnection.java:88)
at io.sentry.connection.AsyncConnection$EventSubmitter.run(AsyncConnection.java:187)
at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130)
at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:630)
at java.base/java.lang.Thread.run(Thread.java:831)
Caused by: java.io.IOException: Server returned HTTP response code: 400 for URL: https://sentry.io/api/1404248/store/
at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1985)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1577)
at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getInputStream(HttpsURLConnectionImpl.java:224)
at io.sentry.connection.HttpConnection.doSend(HttpConnection.java:172)
... 5 more
18:27:53.094 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: HAVE_REMOTE_OFFER
18:27:53.096 (WebRTCClientConnection.java:73) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S setRemoteDescription success.
18:27:53.098 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onSignalingChange: STABLE
18:27:53.100 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange GATHERING
18:27:53.240 (WebRTCClientConnection.java:305) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onSignalingChange: STABLE
18:27:53.240 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CHECKING
18:27:53.241 (WebRTCClientConnection.java:243) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C setRemoteDescription success.
18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTING
18:27:53.256 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTING
18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceGatheringChange COMPLETE
18:27:53.257 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CHECKING
18:27:53.257 (WebRTCClientConnection.java:344) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceGatheringChange COMPLETE
18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:310) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onIceConnectionChange CONNECTED
18:27:53.258 (WebRTCClientConnection.java:329) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C PeerConnection.onIceConnectionChange COMPLETED
18:27:53.261 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman] INFO - C localDataChannel onStateChange OPEN
18:27:53.261 (WebRTCClientConnection.java:395) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S PeerConnection.onDataChannel
18:27:53.262 (WebRTCClientConnection.java:440) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - S localDataChannel onStateChange OPEN
18:27:53.264 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_INIT_MSG
18:27:53.266 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: USE_AUTH_TYPE_MSG
18:27:53.267 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: USE_AUTH_TYPE_MSG
18:27:53.271 (ClientHandshake.java:145) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :send: CLIENT_AUTH_MESSAGE
18:27:53.272 (ServerHandshake.java:212) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :got: CLIENT_AUTH_MESSAGE
18:27:53.272 (ServerHandshake.java:202) [WebRTCClientConnection.WebRTCThread_skullman@Draconis] INFO - skullman@Draconis :send: CONNECTION_SUCCESSFUL_MSG
18:27:53.277 (ClientHandshake.java:155) [WebRTCClientConnection.WebRTCThread_skullman] INFO - skullman :got: CONNECTION_SUCCESSFUL_MSG
```
### To Reproduce
Start a WebRTC server a bunch of times and get unlucky.
### Expected Behaviour
Local client is always able to send messages.
### Screenshots
_No response_
### MapTool Info
1.11.4 with WebRTC patch
### Desktop
Linux[Debian]
### Additional Context
_No response_
|
non_process
|
occasional nullpointerexception on local webrtc client connection describe the bug after starting a webrtc server it failed to start properly clienthandshake java info skullman send client init msg exception in thread thread java lang nullpointerexception cannot invoke dev onvoid webrtc rtcpeerconnection getconnectionstate because this peerconnection is null at net rptools clientserver simple client webrtcclientconnection sendmessage webrtcclientconnection java at net rptools clientserver simple client webrtcclientconnection sendmessage webrtcclientconnection java at net rptools clientserver hessian client methodclientdecorator sendmessage methodclientdecorator java at net rptools maptool server clienthandshake sendmessage clienthandshake java at net rptools maptool server clienthandshake starthandshake clienthandshake java at net rptools maptool client maptoolconnection start maptoolconnection java at net rptools maptool client maptool createconnection maptool java at net rptools maptool client appactions lambda executeaction appactions java at net rptools maptool client appactions clientaction lambda runbackground appactions java at java base java lang thread run thread java webrtcserverconnection java info s websocket connected webrtcclientconnection java info c websocket connected webrtcclientconnection java info c peerconnection onrenegotiationneeded webrtcclientconnection java info c peerconnection onsignalingchange have local offer webrtcclientconnection java info c peerconnection onicegatheringchange gathering warn io sentry connection abstractconnection lockdown initiated a temporary lockdown because of exception detail bad sentry dsn public key error io sentry connection asyncconnection an exception occurred while sending the event to sentry io sentry connection connectionexception detail bad sentry dsn public key at io sentry connection httpconnection dosend httpconnection java at io sentry connection abstractconnection send abstractconnection java at io sentry connection asyncconnection eventsubmitter run asyncconnection java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java caused by java io ioexception server returned http response code for url at java base sun net at java base sun net at java base sun net at io sentry connection httpconnection dosend httpconnection java more webrtcclientconnection java info s peerconnection onsignalingchange have remote offer webrtcclientconnection java info s setremotedescription success webrtcclientconnection java info s peerconnection onsignalingchange stable webrtcclientconnection java info s peerconnection onicegatheringchange gathering webrtcclientconnection java info c peerconnection onsignalingchange stable webrtcclientconnection java info c peerconnection oniceconnectionchange checking webrtcclientconnection java info c setremotedescription success webrtcclientconnection java info c peerconnection onconnectionchange connecting webrtcclientconnection java info s peerconnection onconnectionchange connecting webrtcclientconnection java info c peerconnection onicegatheringchange complete webrtcclientconnection java info s peerconnection oniceconnectionchange checking webrtcclientconnection java info s peerconnection onicegatheringchange complete webrtcclientconnection java info c peerconnection onconnectionchange connected webrtcclientconnection java info s peerconnection onconnectionchange connected webrtcclientconnection java info c peerconnection oniceconnectionchange connected webrtcclientconnection java info s peerconnection oniceconnectionchange connected webrtcclientconnection java info c peerconnection oniceconnectionchange completed webrtcclientconnection java info c localdatachannel onstatechange open webrtcclientconnection java info s peerconnection ondatachannel webrtcclientconnection java info s localdatachannel onstatechange open serverhandshake java info skullman draconis got client init msg serverhandshake java info skullman draconis send use auth type msg clienthandshake java info skullman got use auth type msg clienthandshake java info skullman send client auth message serverhandshake java info skullman draconis got client auth message serverhandshake java info skullman draconis send connection successful msg clienthandshake java info skullman got connection successful msg to reproduce start a webrtc server a bunch of times and get unlucky expected behaviour local client is always able to send messages screenshots no response maptool info with webrtc patch desktop linux additional context no response
| 0
|
113,657
| 9,660,262,093
|
IssuesEvent
|
2019-05-20 15:08:01
|
OpenTechFund/opentech.fund
|
https://api.github.com/repos/OpenTechFund/opentech.fund
|
closed
|
Comment editing in submission detail view
|
RTBC needs tests todd_mini_sprint
|
Task from #1128
**Acceptance criteria**
- [x] Private comments in the detail view allow editing by the author of the comment
- [x] Internal comments in the detail view allow editing by the author of the comment
- [x] Edit button is visible on comment (styled as per edit button on submission detail page)
- [x] Inline editing of comment with 'submit' and 'cancel' button
- [x] Updated comment to display on edit submission
- [x] Last edited date to be displayed alongside post date
- [x] Comment sort order remains as it currently is (by original post date)
**QA Criteria**
Dev:
- [x] checked feature meets acceptance criteria/conforms exactly to the specification.
- [x] provided good unit test coverage (if this is non-trivial behaviour).
- [x] checked all tests for the project pass with this feature enabled.
- [x] checked code conforms to the project coding standards.
- [x] had code reviewed by another developer and resolved any issues raised.
- [x] tested this feature as an end user of the website/app (Can I get to it? Is it useable? Can I break it? Does it work in an end-to-end context?)
- [x] checked that this feature works on the server/s I am deploying it to.
QA:
- [ ] tested this feature as a front end user and it meets the acceptance criteria/conforms to the specification and design.
- [ ] checked that the feature works on the server/s deployed to
|
1.0
|
Comment editing in submission detail view - Task from #1128
**Acceptance criteria**
- [x] Private comments in the detail view allow editing by the author of the comment
- [x] Internal comments in the detail view allow editing by the author of the comment
- [x] Edit button is visible on comment (styled as per edit button on submission detail page)
- [x] Inline editing of comment with 'submit' and 'cancel' button
- [x] Updated comment to display on edit submission
- [x] Last edited date to be displayed alongside post date
- [x] Comment sort order remains as it currently is (by original post date)
**QA Criteria**
Dev:
- [x] checked feature meets acceptance criteria/conforms exactly to the specification.
- [x] provided good unit test coverage (if this is non-trivial behaviour).
- [x] checked all tests for the project pass with this feature enabled.
- [x] checked code conforms to the project coding standards.
- [x] had code reviewed by another developer and resolved any issues raised.
- [x] tested this feature as an end user of the website/app (Can I get to it? Is it useable? Can I break it? Does it work in an end-to-end context?)
- [x] checked that this feature works on the server/s I am deploying it to.
QA:
- [ ] tested this feature as a front end user and it meets the acceptance criteria/conforms to the specification and design.
- [ ] checked that the feature works on the server/s deployed to
|
non_process
|
comment editing in submission detail view task from acceptance criteria private comments in the detail view allow editing by the author of the comment internal comments in the detail view allow editing by the author of the comment edit button is visible on comment styled as per edit button on submission detail page inline editing of comment with submit and cancel button updated comment to display on edit submission last edited date to be displayed alongside post date comment sort order remains as it currently is by original post date qa criteria dev checked feature meets acceptance criteria conforms exactly to the specification provided good unit test coverage if this is non trivial behaviour checked all tests for the project pass with this feature enabled checked code conforms to the project coding standards had code reviewed by another developer and resolved any issues raised tested this feature as an end user of the website app can i get to it is it useable can i break it does it work in an end to end context checked that this feature works on the server s i am deploying it to qa tested this feature as a front end user and it meets the acceptance criteria conforms to the specification and design checked that the feature works on the server s deployed to
| 0
|
239,786
| 26,232,081,030
|
IssuesEvent
|
2023-01-05 01:45:10
|
pustovitDmytro/lalaps
|
https://api.github.com/repos/pustovitDmytro/lalaps
|
opened
|
CVE-2022-24999 (High) detected in qs-6.10.1.tgz
|
security vulnerability
|
## CVE-2022-24999 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/superagent/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- supertest-6.1.6.tgz (Root Library)
- superagent-6.1.0.tgz
- :x: **qs-6.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pustovitDmytro/lalaps/commit/0888eee49d490c03a2d773287f1270e8a617d8ca">0888eee49d490c03a2d773287f1270e8a617d8ca</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable).
<p>Publish Date: 2022-11-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p>
<p>Release Date: 2022-11-26</p>
<p>Fix Resolution (qs): 6.10.3</p>
<p>Direct dependency fix Resolution (supertest): 6.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-24999 (High) detected in qs-6.10.1.tgz - ## CVE-2022-24999 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.10.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.10.1.tgz">https://registry.npmjs.org/qs/-/qs-6.10.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/superagent/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- supertest-6.1.6.tgz (Root Library)
- superagent-6.1.0.tgz
- :x: **qs-6.10.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pustovitDmytro/lalaps/commit/0888eee49d490c03a2d773287f1270e8a617d8ca">0888eee49d490c03a2d773287f1270e8a617d8ca</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
qs before 6.10.3, as used in Express before 4.17.3 and other products, allows attackers to cause a Node process hang for an Express application because an __ proto__ key can be used. In many typical Express use cases, an unauthenticated remote attacker can place the attack payload in the query string of the URL that is used to visit the application, such as a[__proto__]=b&a[__proto__]&a[length]=100000000. The fix was backported to qs 6.9.7, 6.8.3, 6.7.3, 6.6.1, 6.5.3, 6.4.1, 6.3.3, and 6.2.4 (and therefore Express 4.17.3, which has "deps: qs@6.9.7" in its release description, is not vulnerable).
<p>Publish Date: 2022-11-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24999>CVE-2022-24999</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-24999">https://www.cve.org/CVERecord?id=CVE-2022-24999</a></p>
<p>Release Date: 2022-11-26</p>
<p>Fix Resolution (qs): 6.10.3</p>
<p>Direct dependency fix Resolution (supertest): 6.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules superagent node modules qs package json dependency hierarchy supertest tgz root library superagent tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details qs before as used in express before and other products allows attackers to cause a node process hang for an express application because an proto key can be used in many typical express use cases an unauthenticated remote attacker can place the attack payload in the query string of the url that is used to visit the application such as a b a a the fix was backported to qs and and therefore express which has deps qs in its release description is not vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs direct dependency fix resolution supertest step up your open source security game with mend
| 0
|
24,324
| 12,065,516,652
|
IssuesEvent
|
2020-04-16 10:06:04
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
closed
|
Document licenses for third party open source projects that we depend on
|
planned/3.15 team/core-services
|
We need something like this: https://neo4j.com/terms/thirdpartylicenses/
The page should be generated automatically. Tagging both @lguychard and @tsenart because the implementation is probably different for TypeScript and Go dependencies.
I would like to prioritize this for 3.15 release on April 20 but if you want to do it earlier, I wouldn't complain :)
slack thread: https://sourcegraph.slack.com/archives/C0B2RU51Q/p1582918787128100
|
1.0
|
Document licenses for third party open source projects that we depend on - We need something like this: https://neo4j.com/terms/thirdpartylicenses/
The page should be generated automatically. Tagging both @lguychard and @tsenart because the implementation is probably different for TypeScript and Go dependencies.
I would like to prioritize this for 3.15 release on April 20 but if you want to do it earlier, I wouldn't complain :)
slack thread: https://sourcegraph.slack.com/archives/C0B2RU51Q/p1582918787128100
|
non_process
|
document licenses for third party open source projects that we depend on we need something like this the page should be generated automatically tagging both lguychard and tsenart because the implementation is probably different for typescript and go dependencies i would like to prioritize this for release on april but if you want to do it earlier i wouldn t complain slack thread
| 0
|
6,438
| 9,540,966,250
|
IssuesEvent
|
2019-04-30 20:57:23
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
opened
|
Firestore: 'test_collection_add' flakes
|
api: firestore flaky testing type: process
|
From https://source.cloud.google.com/results/invocations/7deedac3-8b60-447d-bb5e-74a4a6653e57/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Ffirestore/log:
```python
_____________________________ test_collection_add ______________________________
client = <google.cloud.firestore_v1.client.Client object at 0x7fb9ca171ad0>
cleanup = <built-in method append of list object at 0x7fb9c7c0acb0>
def test_collection_add(client, cleanup):
collection1 = client.collection("collek")
collection2 = client.collection("collek", "shun", "child")
collection3 = client.collection("collek", "table", "child")
explicit_doc_id = "hula" + unique_resource_id("-")
> assert set(collection1.list_documents()) == set()
E assert set([<google....fb9c7c06290>]) == set([])
E Extra items in the left set:
E <google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290>
E Full diff:
E - set([<google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290>])
E + set([])
```
This test should be using a unique / name-mangled collection ID, in order to prevent collision with CI running for other PRs.
|
1.0
|
Firestore: 'test_collection_add' flakes - From https://source.cloud.google.com/results/invocations/7deedac3-8b60-447d-bb5e-74a4a6653e57/targets/cloud-devrel%2Fclient-libraries%2Fgoogle-cloud-python%2Fpresubmit%2Ffirestore/log:
```python
_____________________________ test_collection_add ______________________________
client = <google.cloud.firestore_v1.client.Client object at 0x7fb9ca171ad0>
cleanup = <built-in method append of list object at 0x7fb9c7c0acb0>
def test_collection_add(client, cleanup):
collection1 = client.collection("collek")
collection2 = client.collection("collek", "shun", "child")
collection3 = client.collection("collek", "table", "child")
explicit_doc_id = "hula" + unique_resource_id("-")
> assert set(collection1.list_documents()) == set()
E assert set([<google....fb9c7c06290>]) == set([])
E Extra items in the left set:
E <google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290>
E Full diff:
E - set([<google.cloud.firestore_v1.document.DocumentReference object at 0x7fb9c7c06290>])
E + set([])
```
This test should be using a unique / name-mangled collection ID, in order to prevent collision with CI running for other PRs.
|
process
|
firestore test collection add flakes from python test collection add client cleanup def test collection add client cleanup client collection collek client collection collek shun child client collection collek table child explicit doc id hula unique resource id assert set list documents set e assert set set e extra items in the left set e e full diff e set e set this test should be using a unique name mangled collection id in order to prevent collision with ci running for other prs
| 1
|
101,069
| 8,773,741,130
|
IssuesEvent
|
2018-12-18 17:42:32
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
no error message for bad behavior when unable to connect to sync servers
|
QA/Test-Plan-Specified QA/Yes feature/sync
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
Follow up to https://github.com/brave/brave-core/pull/1019
The final scenario in the test plan for the above PR did not pass QA. Scenario is `Bad behavior from Sync servers` in PR.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Create a sync chain and ensure you see two devices
2. Quit Brave
3. Launch Brave without internet
4. You should see a `Unable to connect to the Sync servers.` error dialog
## Actual result:
No error dialog
## Expected result:
Should be error dialog
## Reproduces how often:
easily
## Brave version (brave://version info)
Brave | 0.58.12 Chromium: 71.0.3578.80 (Official Build) (64-bit)
-- | --
Revision | 2ac50e7249fbd55e6f517a28131605c9fb9fe897-refs/branch-heads/3578@{#860}
OS | Mac OS X
### Reproducible on current release:
- Does it reproduce on brave-browser dev/beta builds?
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
### Additional Information
cc @brave/legacy_qa for verification on other platforms
|
1.0
|
no error message for bad behavior when unable to connect to sync servers - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
Follow up to https://github.com/brave/brave-core/pull/1019
The final scenario in the test plan for the above PR did not pass QA. Scenario is `Bad behavior from Sync servers` in PR.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Create a sync chain and ensure you see two devices
2. Quit Brave
3. Launch Brave without internet
4. You should see a `Unable to connect to the Sync servers.` error dialog
## Actual result:
No error dialog
## Expected result:
Should be error dialog
## Reproduces how often:
easily
## Brave version (brave://version info)
Brave | 0.58.12 Chromium: 71.0.3578.80 (Official Build) (64-bit)
-- | --
Revision | 2ac50e7249fbd55e6f517a28131605c9fb9fe897-refs/branch-heads/3578@{#860}
OS | Mac OS X
### Reproducible on current release:
- Does it reproduce on brave-browser dev/beta builds?
### Website problems only:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
### Additional Information
cc @brave/legacy_qa for verification on other platforms
|
non_process
|
no error message for bad behavior when unable to connect to sync servers have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description follow up to the final scenario in the test plan for the above pr did not pass qa scenario is bad behavior from sync servers in pr steps to reproduce create a sync chain and ensure you see two devices quit brave launch brave without internet you should see a unable to connect to the sync servers error dialog actual result no error dialog expected result should be error dialog reproduces how often easily brave version brave version info brave chromium official build bit revision refs branch heads os mac os x reproducible on current release does it reproduce on brave browser dev beta builds website problems only does the issue resolve itself when disabling brave shields n a is the issue reproducible on the latest version of chrome n a additional information cc brave legacy qa for verification on other platforms
| 0
|
60,171
| 6,673,557,552
|
IssuesEvent
|
2017-10-04 15:25:41
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
UWP: 2 System.IO.Pipes failing after enabling other tests on same project
|
area-System.IO test-run-uwp-ilc
|
```
System.IO.Pipes.Tests.NamedPipeTest_Specific.Windows_GetImpersonationUserName_Succeed [FAIL]
GetNamedPipeHandleState failed
Expected: True
Actual: False
Stack Trace:
at System.IO.Pipes.Tests.NamedPipeTest_Specific.<Windows_GetImpersonationUserName_Succeed>d__8.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.RunAsClient_Windows [FAIL]
Expected delegate to have been invoked
Expected: True
Actual: False
Stack Trace:
at System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.<RunAsClient_Windows>d__2.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
xUnit.net console test runner (64-bit .NET Core)
Copyright (C) 2014 Outercurve Foundation.
```
|
1.0
|
UWP: 2 System.IO.Pipes failing after enabling other tests on same project - ```
System.IO.Pipes.Tests.NamedPipeTest_Specific.Windows_GetImpersonationUserName_Succeed [FAIL]
GetNamedPipeHandleState failed
Expected: True
Actual: False
Stack Trace:
at System.IO.Pipes.Tests.NamedPipeTest_Specific.<Windows_GetImpersonationUserName_Succeed>d__8.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.RunAsClient_Windows [FAIL]
Expected delegate to have been invoked
Expected: True
Actual: False
Stack Trace:
at System.IO.Pipes.Tests.NamedPipeTest_RunAsClient.<RunAsClient_Windows>d__2.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
at Void Xunit.Sdk.TestInvoker`1+<>c__DisplayClass46_1+<<InvokeTestMethodAsync>b__1>d.MoveNext() + 0x2ea
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
--- End of stack trace from previous location where exception was thrown ---
at Void System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() + 0x21
at Void System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task) + 0x7e
at Void System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task) + 0x53
xUnit.net console test runner (64-bit .NET Core)
Copyright (C) 2014 Outercurve Foundation.
```
|
non_process
|
uwp system io pipes failing after enabling other tests on same project system io pipes tests namedpipetest specific windows getimpersonationusername succeed getnamedpipehandlestate failed expected true actual false stack trace at system io pipes tests namedpipetest specific d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task at void xunit sdk testinvoker c b d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task system io pipes tests namedpipetest runasclient runasclient windows expected delegate to have been invoked expected true actual false stack trace at system io pipes tests namedpipetest runasclient d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task at void xunit sdk testinvoker c b d movenext end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task end of stack trace from previous location where exception was thrown at void system runtime exceptionservices exceptiondispatchinfo throw at void system runtime compilerservices taskawaiter throwfornonsuccess task at void system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task xunit net console test runner bit net core copyright c outercurve foundation
| 0
|
51,191
| 7,680,269,405
|
IssuesEvent
|
2018-05-16 00:34:39
|
grpc/grpc
|
https://api.github.com/repos/grpc/grpc
|
closed
|
Add Objective-C API to disable fail fast mode
|
disposition/FOR DISCUSSION disposition/Needs Documentation kind/enhancement lang/ObjC priority/P3
|
As part of the 0.14 release, fail fast is now enabled by default.
In order to disable fail fast mode, a flag must be set on metadata requests.
https://github.com/grpc/grpc/blob/master/include/grpc/impl/codegen/grpc_types.h#L212
We need to ensure that there is some way a user can disable fail fast mode.
|
1.0
|
Add Objective-C API to disable fail fast mode - As part of the 0.14 release, fail fast is now enabled by default.
In order to disable fail fast mode, a flag must be set on metadata requests.
https://github.com/grpc/grpc/blob/master/include/grpc/impl/codegen/grpc_types.h#L212
We need to ensure that there is some way a user can disable fail fast mode.
|
non_process
|
add objective c api to disable fail fast mode as part of the release fail fast is now enabled by default in order to disable fail fast mode a flag must be set on metadata requests we need to ensure that there is some way a user can disable fail fast mode
| 0
|
14,607
| 17,703,636,347
|
IssuesEvent
|
2021-08-25 03:27:04
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - scientificName
|
Term - change Class - Taxon non-normative Process - complete
|
## Term change
* Submitter: Quentin Groom
* Efficacy Justification (why is this change necessary?): To improve clarity of the term usage, particularly to distinguish the different terms that can hold a scientific Latin name
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): This is largely for people and organizations publishing Darwin Core files to avoid repeated questions that keep cropping up. The issue #28 highlighted that the definitions of `scientificName`, `acceptedNameUsage `and `originalNameUsage` are all similar to one another, however, their intended usage is quite distinct, even though it is not clearly documented. The intension of this suggested change is to add to the comments of the term to help users understand the use of the terms more easily. The suggested explanations were given by @deepreef in #28, but they are only preliminary.
* Stability Justification (what concerns are there that this might affect existing implementations?): The intension is that the comments would reinforce the existing definition and thus improve stability.
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No implication
Current Term definition: https://dwc.tdwg.org/list/#dwc_scientificName
Proposed attributes of the new term:
* Usage comments (recommendations regarding content, etc., not normative): **This term should not contain identification qualifications, which should instead be supplied in the IdentificationQualifier term. When applied to an Organism or Occurrence, this term should be used to represent the scientific name that was applied to the associated Organism in accordance with the Taxon to which it was or is currently identified.**
* Examples: `Coleoptera` (order). `Vespertilionidae` (family). `Manis` (genus). `Ctenomys sociabilis` (genus + specificEpithet). `Ambystoma tigrinum diaboli` (genus + specificEpithet + infraspecificEpithet). `Roptrocerus typographi (Györfi, 1952)` (genus + specificEpithet + scientificNameAuthorship), `Quercus agrifolia var. oxyadenia (Torr.) J.T. Howell` (genus + specificEpithet + taxonRank + infraspecificEpithet + scientificNameAuthorship).
|
1.0
|
Change term - scientificName - ## Term change
* Submitter: Quentin Groom
* Efficacy Justification (why is this change necessary?): To improve clarity of the term usage, particularly to distinguish the different terms that can hold a scientific Latin name
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): This is largely for people and organizations publishing Darwin Core files to avoid repeated questions that keep cropping up. The issue #28 highlighted that the definitions of `scientificName`, `acceptedNameUsage `and `originalNameUsage` are all similar to one another, however, their intended usage is quite distinct, even though it is not clearly documented. The intension of this suggested change is to add to the comments of the term to help users understand the use of the terms more easily. The suggested explanations were given by @deepreef in #28, but they are only preliminary.
* Stability Justification (what concerns are there that this might affect existing implementations?): The intension is that the comments would reinforce the existing definition and thus improve stability.
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No implication
Current Term definition: https://dwc.tdwg.org/list/#dwc_scientificName
Proposed attributes of the new term:
* Usage comments (recommendations regarding content, etc., not normative): **This term should not contain identification qualifications, which should instead be supplied in the IdentificationQualifier term. When applied to an Organism or Occurrence, this term should be used to represent the scientific name that was applied to the associated Organism in accordance with the Taxon to which it was or is currently identified.**
* Examples: `Coleoptera` (order). `Vespertilionidae` (family). `Manis` (genus). `Ctenomys sociabilis` (genus + specificEpithet). `Ambystoma tigrinum diaboli` (genus + specificEpithet + infraspecificEpithet). `Roptrocerus typographi (Györfi, 1952)` (genus + specificEpithet + scientificNameAuthorship), `Quercus agrifolia var. oxyadenia (Torr.) J.T. Howell` (genus + specificEpithet + taxonRank + infraspecificEpithet + scientificNameAuthorship).
|
process
|
change term scientificname term change submitter quentin groom efficacy justification why is this change necessary to improve clarity of the term usage particularly to distinguish the different terms that can hold a scientific latin name demand justification if the change is semantic in nature name at least two organizations that independently need this term this is largely for people and organizations publishing darwin core files to avoid repeated questions that keep cropping up the issue highlighted that the definitions of scientificname acceptednameusage and originalnameusage are all similar to one another however their intended usage is quite distinct even though it is not clearly documented the intension of this suggested change is to add to the comments of the term to help users understand the use of the terms more easily the suggested explanations were given by deepreef in but they are only preliminary stability justification what concerns are there that this might affect existing implementations the intension is that the comments would reinforce the existing definition and thus improve stability implications for dwciri namespace does this change affect a dwciri term version no implication current term definition proposed attributes of the new term usage comments recommendations regarding content etc not normative this term should not contain identification qualifications which should instead be supplied in the identificationqualifier term when applied to an organism or occurrence this term should be used to represent the scientific name that was applied to the associated organism in accordance with the taxon to which it was or is currently identified examples coleoptera order vespertilionidae family manis genus ctenomys sociabilis genus specificepithet ambystoma tigrinum diaboli genus specificepithet infraspecificepithet roptrocerus typographi györfi genus specificepithet scientificnameauthorship quercus agrifolia var oxyadenia torr j t howell genus specificepithet taxonrank infraspecificepithet scientificnameauthorship
| 1
|
21,044
| 27,987,347,455
|
IssuesEvent
|
2023-03-26 20:49:17
|
bitfocus/companion-module-requests
|
https://api.github.com/repos/bitfocus/companion-module-requests
|
opened
|
Hive Player
|
NOT YET PROCESSED
|
- [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control: Hive Player
/
What you would like to be able to make it do from Companion: be able to control playback and layer opacity and skip cues etc.
Direct links or attachments to the ethernet control protocol or API: I can email a pdf with full UDP commands or contact Hive at www.hive.run
|
1.0
|
Hive Player - - [ ] **I have researched the list of existing Companion modules and requests and have determined this has not yet been requested**
The name of the device, hardware, or software you would like to control: Hive Player
/
What you would like to be able to make it do from Companion: be able to control playback and layer opacity and skip cues etc.
Direct links or attachments to the ethernet control protocol or API: I can email a pdf with full UDP commands or contact Hive at www.hive.run
|
process
|
hive player i have researched the list of existing companion modules and requests and have determined this has not yet been requested the name of the device hardware or software you would like to control hive player what you would like to be able to make it do from companion be able to control playback and layer opacity and skip cues etc direct links or attachments to the ethernet control protocol or api i can email a pdf with full udp commands or contact hive at
| 1
|
22,202
| 30,758,364,639
|
IssuesEvent
|
2023-07-29 11:08:21
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
skypilot-nightly 1.0.0.dev20230729 has 2 GuardDog issues
|
guarddog exec-base64 silent-process-execution
|
https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230729",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230729/sky/cloud_stores.py:113",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230729/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpwfkzzrpo/skypilot-nightly"
}
}```
|
1.0
|
skypilot-nightly 1.0.0.dev20230729 has 2 GuardDog issues - https://pypi.org/project/skypilot-nightly
https://inspector.pypi.io/project/skypilot-nightly
```{
"dependency": "skypilot-nightly",
"version": "1.0.0.dev20230729",
"result": {
"issues": 2,
"errors": {},
"results": {
"exec-base64": [
{
"location": "skypilot-nightly-1.0.0.dev20230729/sky/cloud_stores.py:113",
"code": " p = subprocess.run(command,\n stdout=subprocess.PIPE,\n shell=True,\n check=True,\n executable='/bin/bash')",
"message": "This package contains a call to the `eval` function with a `base64` encoded string as argument.\nThis is a common method used to hide a malicious payload in a module as static analysis will not decode the\nstring.\n"
}
],
"silent-process-execution": [
{
"location": "skypilot-nightly-1.0.0.dev20230729/sky/skylet/log_lib.py:219",
"code": " subprocess.Popen(\n daemon_cmd,\n start_new_session=True,\n # Suppress output\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n # Disa... )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpwfkzzrpo/skypilot-nightly"
}
}```
|
process
|
skypilot nightly has guarddog issues dependency skypilot nightly version result issues errors results exec location skypilot nightly sky cloud stores py code p subprocess run command n stdout subprocess pipe n shell true n check true n executable bin bash message this package contains a call to the eval function with a encoded string as argument nthis is a common method used to hide a malicious payload in a module as static analysis will not decode the nstring n silent process execution location skypilot nightly sky skylet log lib py code subprocess popen n daemon cmd n start new session true n suppress output n stdout subprocess devnull n stderr subprocess devnull n disa message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpwfkzzrpo skypilot nightly
| 1
|
637,702
| 20,675,805,666
|
IssuesEvent
|
2022-03-10 09:09:38
|
geosolutions-it/MapStore2-C027
|
https://api.github.com/repos/geosolutions-it/MapStore2-C027
|
closed
|
Requested changes to GeoNetwork
|
Task Priority: Medium C027-COMUNE_FI-2021-SUPPORT
|
The client is requesting few changes to the new GN 3 RNDT installed in the new SDI.
From the client:
>Per quanto riguarda l'interfaccia, invece avremmo la necessità di due migliorie che sono:
>- Nella pagina di ricerca dei dataset, tra i filtraggi aggiungere un filtro che consente di estrarre i risultati in base ad un range di "Data di pubblicazione" (filtrare tutti i dataset con Data di pubblicazione da gg/mm/aaaa a gg/mm/aaaa)

>- Nel pdf che è possibile esportare dopo aver selezionato i dataset eliminare i seguenti campi:
> - Update frequency
> - Record updated
> - Risorse
> - More onlineinformation
>Quindi gli unici campi visibili dovranno esssere:
> - Descrizione
> - Estensione
> - Parole chiave
> - Access and useconstraints
> - Data type
> - Format
> - Identificatoreunivoco

|
1.0
|
Requested changes to GeoNetwork - The client is requesting few changes to the new GN 3 RNDT installed in the new SDI.
From the client:
>Per quanto riguarda l'interfaccia, invece avremmo la necessità di due migliorie che sono:
>- Nella pagina di ricerca dei dataset, tra i filtraggi aggiungere un filtro che consente di estrarre i risultati in base ad un range di "Data di pubblicazione" (filtrare tutti i dataset con Data di pubblicazione da gg/mm/aaaa a gg/mm/aaaa)

>- Nel pdf che è possibile esportare dopo aver selezionato i dataset eliminare i seguenti campi:
> - Update frequency
> - Record updated
> - Risorse
> - More onlineinformation
>Quindi gli unici campi visibili dovranno esssere:
> - Descrizione
> - Estensione
> - Parole chiave
> - Access and useconstraints
> - Data type
> - Format
> - Identificatoreunivoco

|
non_process
|
requested changes to geonetwork the client is requesting few changes to the new gn rndt installed in the new sdi from the client per quanto riguarda l interfaccia invece avremmo la necessità di due migliorie che sono nella pagina di ricerca dei dataset tra i filtraggi aggiungere un filtro che consente di estrarre i risultati in base ad un range di data di pubblicazione filtrare tutti i dataset con data di pubblicazione da gg mm aaaa a gg mm aaaa nel pdf che è possibile esportare dopo aver selezionato i dataset eliminare i seguenti campi update frequency record updated risorse more onlineinformation quindi gli unici campi visibili dovranno esssere descrizione estensione parole chiave access and useconstraints data type format identificatoreunivoco
| 0
|
235,552
| 18,051,853,073
|
IssuesEvent
|
2021-09-19 21:55:03
|
EdwarMalpica/AsignacionOficinas
|
https://api.github.com/repos/EdwarMalpica/AsignacionOficinas
|
closed
|
Implementacion de la configuracion necesaria para la capa de persistencia
|
documentation enhancement
|
Codificar la configuracion necesaria para la creacion de la base de datos
|
1.0
|
Implementacion de la configuracion necesaria para la capa de persistencia - Codificar la configuracion necesaria para la creacion de la base de datos
|
non_process
|
implementacion de la configuracion necesaria para la capa de persistencia codificar la configuracion necesaria para la creacion de la base de datos
| 0
|
60,683
| 7,373,620,840
|
IssuesEvent
|
2018-03-13 17:46:58
|
ParabolInc/action
|
https://api.github.com/repos/ParabolInc/action
|
opened
|
Design: retrospective free→pro upgrade path
|
design enhancement groom-ready
|
## Issue - Enhancement
Let's make some money. Our bet is that adding a team handbook and retrospective meeting type will be enough differentiated value to drive adoption and conversion. Today, we're largely driving conversion through top-down sales. However, we should always have a experience we are testing to drive bottoms-up conversions.
### Some history
When we originally launched Parabol's payment features, Parabol was trial-ware. Teams could try us free for 30 days. They could get another 30 days free if they gave us their credit card number. When the trial expired, their users would lose access to all of their organization's teams until they paid.
*What we learned:* our value prop wasn't thick enough to drive self-adoption, so both free usage _and_ conversions were low.
*What we didn't learn:* trialing for premium features is bad. In fact, trialing for premium features might still be good.
### Goals & Acceptance Criteria (optional)
An designed, lo-fidelity experience where users can:
- Get to know all of the features of Parabol and get themselves and their teams hooked, meaning:
_(this is how we'll know this design has worked)_
- See clear:
- Messages for which features are pro and are subjected to being limited
- Calls to action for how to convert
- At least one team is running at least one meeting (action/retro) regularly (at least 1/mo)
- At the end of the evaluation period, the team converts at least 10% of the time
Validation criteria:
- The lo-fi design is checked with at least two of our partner/advisors
Once the design is completed, the following are created
- A list of implementation Issues/stories
- A list of supporting marketing workflows (for example, emails to send)
### Implementation considerations
#### Trial-ware
In addition to `personal` and `pro`, we create a new tier called `pro-trial` (or, extend the `pro` tier to have an expiration date). When a user creates their first org, their org is in the `pro-trial` tier.
We assume that it takes a quarter (3 months/3x4-week periods) to convert. After 7 weeks, we give them a banner that their trial is about to run out, but they can extend it for 4 more weeks free if they give us their cc now. After a maximal period of 12 weeks, if they do not convert, we offer them the a workflow to demote their account to the `personal` free tier.
We send marketing communications to support conversion.
When the account is in the `personal` tier, we disallow access to things such as Retrospective Meetings and access to the Team Handbook.
#### Limit-ware
All accounts start in the `personal` tier. Some features are limited, just as we limit access to a team's Task archive today.
We begin by adding a limit to the number of Retrospective meetings that can be run. For example, we offer a limit of 3 per org.
We limit the number of Handbook items that can be created. For example, 12 per org.
When one of these limits is exceeded, we decorate the interface with affordances to convert and remove these limits.
We send marketing communications to support conversion.
- **Estimated effort:** 15 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
|
1.0
|
Design: retrospective free→pro upgrade path - ## Issue - Enhancement
Let's make some money. Our bet is that adding a team handbook and retrospective meeting type will be enough differentiated value to drive adoption and conversion. Today, we're largely driving conversion through top-down sales. However, we should always have a experience we are testing to drive bottoms-up conversions.
### Some history
When we originally launched Parabol's payment features, Parabol was trial-ware. Teams could try us free for 30 days. They could get another 30 days free if they gave us their credit card number. When the trial expired, their users would lose access to all of their organization's teams until they paid.
*What we learned:* our value prop wasn't thick enough to drive self-adoption, so both free usage _and_ conversions were low.
*What we didn't learn:* trialing for premium features is bad. In fact, trialing for premium features might still be good.
### Goals & Acceptance Criteria (optional)
An designed, lo-fidelity experience where users can:
- Get to know all of the features of Parabol and get themselves and their teams hooked, meaning:
_(this is how we'll know this design has worked)_
- See clear:
- Messages for which features are pro and are subjected to being limited
- Calls to action for how to convert
- At least one team is running at least one meeting (action/retro) regularly (at least 1/mo)
- At the end of the evaluation period, the team converts at least 10% of the time
Validation criteria:
- The lo-fi design is checked with at least two of our partner/advisors
Once the design is completed, the following are created
- A list of implementation Issues/stories
- A list of supporting marketing workflows (for example, emails to send)
### Implementation considerations
#### Trial-ware
In addition to `personal` and `pro`, we create a new tier called `pro-trial` (or, extend the `pro` tier to have an expiration date). When a user creates their first org, their org is in the `pro-trial` tier.
We assume that it takes a quarter (3 months/3x4-week periods) to convert. After 7 weeks, we give them a banner that their trial is about to run out, but they can extend it for 4 more weeks free if they give us their cc now. After a maximal period of 12 weeks, if they do not convert, we offer them the a workflow to demote their account to the `personal` free tier.
We send marketing communications to support conversion.
When the account is in the `personal` tier, we disallow access to things such as Retrospective Meetings and access to the Team Handbook.
#### Limit-ware
All accounts start in the `personal` tier. Some features are limited, just as we limit access to a team's Task archive today.
We begin by adding a limit to the number of Retrospective meetings that can be run. For example, we offer a limit of 3 per org.
We limit the number of Handbook items that can be created. For example, 12 per org.
When one of these limits is exceeded, we decorate the interface with affordances to convert and remove these limits.
We send marketing communications to support conversion.
- **Estimated effort:** 15 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
|
non_process
|
design retrospective free→pro upgrade path issue enhancement let s make some money our bet is that adding a team handbook and retrospective meeting type will be enough differentiated value to drive adoption and conversion today we re largely driving conversion through top down sales however we should always have a experience we are testing to drive bottoms up conversions some history when we originally launched parabol s payment features parabol was trial ware teams could try us free for days they could get another days free if they gave us their credit card number when the trial expired their users would lose access to all of their organization s teams until they paid what we learned our value prop wasn t thick enough to drive self adoption so both free usage and conversions were low what we didn t learn trialing for premium features is bad in fact trialing for premium features might still be good goals acceptance criteria optional an designed lo fidelity experience where users can get to know all of the features of parabol and get themselves and their teams hooked meaning this is how we ll know this design has worked see clear messages for which features are pro and are subjected to being limited calls to action for how to convert at least one team is running at least one meeting action retro regularly at least mo at the end of the evaluation period the team converts at least of the time validation criteria the lo fi design is checked with at least two of our partner advisors once the design is completed the following are created a list of implementation issues stories a list of supporting marketing workflows for example emails to send implementation considerations trial ware in addition to personal and pro we create a new tier called pro trial or extend the pro tier to have an expiration date when a user creates their first org their org is in the pro trial tier we assume that it takes a quarter months week periods to convert after weeks we give them a banner that their trial is about to run out but they can extend it for more weeks free if they give us their cc now after a maximal period of weeks if they do not convert we offer them the a workflow to demote their account to the personal free tier we send marketing communications to support conversion when the account is in the personal tier we disallow access to things such as retrospective meetings and access to the team handbook limit ware all accounts start in the personal tier some features are limited just as we limit access to a team s task archive today we begin by adding a limit to the number of retrospective meetings that can be run for example we offer a limit of per org we limit the number of handbook items that can be created for example per org when one of these limits is exceeded we decorate the interface with affordances to convert and remove these limits we send marketing communications to support conversion estimated effort points
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.