Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
11,039
| 13,851,352,604
|
IssuesEvent
|
2020-10-15 03:45:05
|
CATcher-org/CATcher
|
https://api.github.com/repos/CATcher-org/CATcher
|
closed
|
Setup CI on Windows / MacOS + add automated builds
|
aspect-Process p.Medium
|
In our CI checks, we should not only run the tests, but also build the application.
This would give us assurance that the app can be built, after every merged PR.
Let's add this check to our existing CI checks on Ubuntu (via Travis CI).
At the same time, let's also setup CI checks on Windows and MacOS.
- [x] Build the application in our existing CI checks for Ubuntu
- [x] Setup CI for Windows (via AppVeyor)
- [x] Setup CI for MacOS (via Travis CI)
|
1.0
|
Setup CI on Windows / MacOS + add automated builds - In our CI checks, we should not only run the tests, but also build the application.
This would give us assurance that the app can be built, after every merged PR.
Let's add this check to our existing CI checks on Ubuntu (via Travis CI).
At the same time, let's also setup CI checks on Windows and MacOS.
- [x] Build the application in our existing CI checks for Ubuntu
- [x] Setup CI for Windows (via AppVeyor)
- [x] Setup CI for MacOS (via Travis CI)
|
process
|
setup ci on windows macos add automated builds in our ci checks we should not only run the tests but also build the application this would give us assurance that the app can be built after every merged pr let s add this check to our existing ci checks on ubuntu via travis ci at the same time let s also setup ci checks on windows and macos build the application in our existing ci checks for ubuntu setup ci for windows via appveyor setup ci for macos via travis ci
| 1
|
671,759
| 22,774,932,661
|
IssuesEvent
|
2022-07-08 13:38:28
|
opensquare-network/paid-qa
|
https://api.github.com/repos/opensquare-network/paid-qa
|
closed
|
refactor: don't use sub-component directly
|
UI priority:low
|
e.g.
- Don't use `MicromarkMd`, use `Preview`
- Don't use `MarkdownEditor`, use `RichEditor`
- Don't use `Dropdown(item)` and write a new Selector, use `DropdownSelector`
Sub-components is **not** for public.
If so, we'll have bulk of work to do/refactor.
|
1.0
|
refactor: don't use sub-component directly - e.g.
- Don't use `MicromarkMd`, use `Preview`
- Don't use `MarkdownEditor`, use `RichEditor`
- Don't use `Dropdown(item)` and write a new Selector, use `DropdownSelector`
Sub-components is **not** for public.
If so, we'll have bulk of work to do/refactor.
|
non_process
|
refactor don t use sub component directly e g don t use micromarkmd use preview don t use markdowneditor use richeditor don t use dropdown item and write a new selector use dropdownselector sub components is not for public if so we ll have bulk of work to do refactor
| 0
|
3,998
| 6,926,273,367
|
IssuesEvent
|
2017-11-30 18:35:44
|
syndesisio/syndesis
|
https://api.github.com/repos/syndesisio/syndesis
|
opened
|
Improve Decoupling of API & UI
|
cat/discussion cat/process cat/retro
|
## tl;dr
Following from our discussion in the retrospective, there is an inherent problem in the application development lifecycle that I think is directly impeding our ability to keep up with feature requests in a timely fashion. I know this is a LONG description, but please take a moment to read, as I think _everyone_ is affected by this. It's an opportunity to improve the development process. If we don't solve this soon, technical debt will continue to increase and the overall product roadmap will be jeopardized.
The two items to address are 1) decoupling the API & UI; 2) making a more parallel work flow rather than linear and blocking. Solutions proposed are at the bottom. We should collaborate and discuss any potential tradeoffs or alternatives.
## Introduction
At the beginning of the project we made a huge effort to decouple the UI from the backend by creating a Node.js API (now in Java). As features continue to be added to the iPaaS roadmap, we are finding it more and more challenging to develop these features without getting blocked by one or more areas.
## Theory
When you are spending as much time trying to get the local build of the REST API working properly (to be able to work with real data) as you do actually implementing a feature, it's a problem. It almost feels like we need yet another layer of abstraction to be able to work completely decoupled from the backend. From the UI perspective, which sits kind of right in the middle, it boils down to two factors:
1) separation of concerns and true decoupling
2) timing per sprint to develop these features (linear vs parallel work flow)
## Current Work Process
Here is a rough outline of our work process:
1. Discuss requirements for the sprint.
2. Assign feature owners. Feature owners are usually backend folk (not sure if this is intentional, but seems like a very important consideration we should be consistent with and have a philosophy it supports).
3. Feature owners write a design proposal, which includes the data structure the UI should adhere to.
4. UXD collaborates with feature owners to create designs.
5. UI receives designs, looks at design proposal for the REST API, and begins work.
6. Confusion about whether to mock out data or use existing REST API work ensues. Neither option is smooth, as using existing resources is risky for reasons stated below, and mocked data is great, but when it comes time to wire up the API, there are often huge discrepencies.
7. Scrambling to get things working properly together, whether using maintainable approaches or not, in time for the demo and sprint end.
## Testing the Theory
The problems described earlier manifest in a few different ways that I've noticed impede development:
### Separation of Concerns
- The REST API sometimes has continuously changing requirements anywhere in steps 3-5 that result in adjustments to the data it feeds to the UI (and its structure), possible changes in endpoints, sometimes the structure is not known at all, etc. This is often out of necessity due to potential issues that are later discovered in the design proposal (naturally, this is much easier to discover on implementation rather than in theory on a PR, which in and of itself is a problem). Other times they are dealing with CI issues.
- The UI tries to satisfy design requirements but they sometimes do not match the data structure provided by the API. Improvements have been made here by increasing communication between UXD and the backend, but sometimes, as stated earlier, structural issues and discrepencies arise mid-implementation, which is normal. Backend issues are discovered, such as a bug or server issues with using the REST API. Also, a lot of the data returned from the API is not relevant and the structure is inconsistent across features.
### Timing
- The REST API & UXD teams are essentially trying to work in parallel with the UI team, which puts an unnecessary amount of pressure on the former teams to get things completed in time for implementation in the UI, and early enough for QE and Docs to also make adjustments (a deadline the UI often can't meet). Parallel should be our goal, but execution is often linear and blocking.
- Since much of the sprint time in the UI is spent creating components and templates for the design implementation, this is often discovered at the very end of the sprint when wiring up the API.
The UI should be able to operate strictly with a known data structure and rough designs. Whether or not that has been implemented in the REST API is irrelevant. The data structure should be known before a single line of code is written on either side.
But who dictates what the structure should be? At the moment it is usually the feature owner, which is usually a backend engineer. There is nothing wrong with that, but if we choose to go that route, some adjustments need to be made.
## Proposed Solutions
Feel free to add to this:
- Decide where the data structure should come from. Considering that iPaaS is heavily user-focused and technically SaaS, to me, it makes sense that it either comes completely from the front end or is mediated between the API and UI (more on that below).
- Decide how that data structure is decided on:
- Consumer-driven contracts, as @zregvart suggested. ([Concept](https://martinfowler.com/articles/consumerDrivenContracts.html) & [example](https://cloud.spring.io/spring-cloud-contract/) for Spring). A pattern to adhere to an evolving service, and can come in many forms, as simple as a spreadsheet.
- UI-driven data models, to which the API then adapts. This doesn't mean UI engineers need to be feature owners, but just that the UI portion would dictate the structure. This would allow us to mock out data knowing that when we switch to the REST API it will be a seamless transition (in theory).
- Using an intermediary solution, like [GraphSQL](http://graphql.org/) and [Apollo](https://www.apollographql.com/client/) (here is a full stack [tutorial](https://www.howtographql.com/)). Not a huge fan, but it could work.
- Putting a freeze on design proposals, regardless of how bad the structure is, and reconvening the next sprint?
---
@syndesisio/all - Let's discuss benefits and tradeoffs, other possible solutions, and any considerations below!
|
1.0
|
Improve Decoupling of API & UI - ## tl;dr
Following from our discussion in the retrospective, there is an inherent problem in the application development lifecycle that I think is directly impeding our ability to keep up with feature requests in a timely fashion. I know this is a LONG description, but please take a moment to read, as I think _everyone_ is affected by this. It's an opportunity to improve the development process. If we don't solve this soon, technical debt will continue to increase and the overall product roadmap will be jeopardized.
The two items to address are 1) decoupling the API & UI; 2) making a more parallel work flow rather than linear and blocking. Solutions proposed are at the bottom. We should collaborate and discuss any potential tradeoffs or alternatives.
## Introduction
At the beginning of the project we made a huge effort to decouple the UI from the backend by creating a Node.js API (now in Java). As features continue to be added to the iPaaS roadmap, we are finding it more and more challenging to develop these features without getting blocked by one or more areas.
## Theory
When you are spending as much time trying to get the local build of the REST API working properly (to be able to work with real data) as you do actually implementing a feature, it's a problem. It almost feels like we need yet another layer of abstraction to be able to work completely decoupled from the backend. From the UI perspective, which sits kind of right in the middle, it boils down to two factors:
1) separation of concerns and true decoupling
2) timing per sprint to develop these features (linear vs parallel work flow)
## Current Work Process
Here is a rough outline of our work process:
1. Discuss requirements for the sprint.
2. Assign feature owners. Feature owners are usually backend folk (not sure if this is intentional, but seems like a very important consideration we should be consistent with and have a philosophy it supports).
3. Feature owners write a design proposal, which includes the data structure the UI should adhere to.
4. UXD collaborates with feature owners to create designs.
5. UI receives designs, looks at design proposal for the REST API, and begins work.
6. Confusion about whether to mock out data or use existing REST API work ensues. Neither option is smooth, as using existing resources is risky for reasons stated below, and mocked data is great, but when it comes time to wire up the API, there are often huge discrepencies.
7. Scrambling to get things working properly together, whether using maintainable approaches or not, in time for the demo and sprint end.
## Testing the Theory
The problems described earlier manifest in a few different ways that I've noticed impede development:
### Separation of Concerns
- The REST API sometimes has continuously changing requirements anywhere in steps 3-5 that result in adjustments to the data it feeds to the UI (and its structure), possible changes in endpoints, sometimes the structure is not known at all, etc. This is often out of necessity due to potential issues that are later discovered in the design proposal (naturally, this is much easier to discover on implementation rather than in theory on a PR, which in and of itself is a problem). Other times they are dealing with CI issues.
- The UI tries to satisfy design requirements but they sometimes do not match the data structure provided by the API. Improvements have been made here by increasing communication between UXD and the backend, but sometimes, as stated earlier, structural issues and discrepencies arise mid-implementation, which is normal. Backend issues are discovered, such as a bug or server issues with using the REST API. Also, a lot of the data returned from the API is not relevant and the structure is inconsistent across features.
### Timing
- The REST API & UXD teams are essentially trying to work in parallel with the UI team, which puts an unnecessary amount of pressure on the former teams to get things completed in time for implementation in the UI, and early enough for QE and Docs to also make adjustments (a deadline the UI often can't meet). Parallel should be our goal, but execution is often linear and blocking.
- Since much of the sprint time in the UI is spent creating components and templates for the design implementation, this is often discovered at the very end of the sprint when wiring up the API.
The UI should be able to operate strictly with a known data structure and rough designs. Whether or not that has been implemented in the REST API is irrelevant. The data structure should be known before a single line of code is written on either side.
But who dictates what the structure should be? At the moment it is usually the feature owner, which is usually a backend engineer. There is nothing wrong with that, but if we choose to go that route, some adjustments need to be made.
## Proposed Solutions
Feel free to add to this:
- Decide where the data structure should come from. Considering that iPaaS is heavily user-focused and technically SaaS, to me, it makes sense that it either comes completely from the front end or is mediated between the API and UI (more on that below).
- Decide how that data structure is decided on:
- Consumer-driven contracts, as @zregvart suggested. ([Concept](https://martinfowler.com/articles/consumerDrivenContracts.html) & [example](https://cloud.spring.io/spring-cloud-contract/) for Spring). A pattern to adhere to an evolving service, and can come in many forms, as simple as a spreadsheet.
- UI-driven data models, to which the API then adapts. This doesn't mean UI engineers need to be feature owners, but just that the UI portion would dictate the structure. This would allow us to mock out data knowing that when we switch to the REST API it will be a seamless transition (in theory).
- Using an intermediary solution, like [GraphSQL](http://graphql.org/) and [Apollo](https://www.apollographql.com/client/) (here is a full stack [tutorial](https://www.howtographql.com/)). Not a huge fan, but it could work.
- Putting a freeze on design proposals, regardless of how bad the structure is, and reconvening the next sprint?
---
@syndesisio/all - Let's discuss benefits and tradeoffs, other possible solutions, and any considerations below!
|
process
|
improve decoupling of api ui tl dr following from our discussion in the retrospective there is an inherent problem in the application development lifecycle that i think is directly impeding our ability to keep up with feature requests in a timely fashion i know this is a long description but please take a moment to read as i think everyone is affected by this it s an opportunity to improve the development process if we don t solve this soon technical debt will continue to increase and the overall product roadmap will be jeopardized the two items to address are decoupling the api ui making a more parallel work flow rather than linear and blocking solutions proposed are at the bottom we should collaborate and discuss any potential tradeoffs or alternatives introduction at the beginning of the project we made a huge effort to decouple the ui from the backend by creating a node js api now in java as features continue to be added to the ipaas roadmap we are finding it more and more challenging to develop these features without getting blocked by one or more areas theory when you are spending as much time trying to get the local build of the rest api working properly to be able to work with real data as you do actually implementing a feature it s a problem it almost feels like we need yet another layer of abstraction to be able to work completely decoupled from the backend from the ui perspective which sits kind of right in the middle it boils down to two factors separation of concerns and true decoupling timing per sprint to develop these features linear vs parallel work flow current work process here is a rough outline of our work process discuss requirements for the sprint assign feature owners feature owners are usually backend folk not sure if this is intentional but seems like a very important consideration we should be consistent with and have a philosophy it supports feature owners write a design proposal which includes the data structure the ui should adhere to uxd collaborates with feature owners to create designs ui receives designs looks at design proposal for the rest api and begins work confusion about whether to mock out data or use existing rest api work ensues neither option is smooth as using existing resources is risky for reasons stated below and mocked data is great but when it comes time to wire up the api there are often huge discrepencies scrambling to get things working properly together whether using maintainable approaches or not in time for the demo and sprint end testing the theory the problems described earlier manifest in a few different ways that i ve noticed impede development separation of concerns the rest api sometimes has continuously changing requirements anywhere in steps that result in adjustments to the data it feeds to the ui and its structure possible changes in endpoints sometimes the structure is not known at all etc this is often out of necessity due to potential issues that are later discovered in the design proposal naturally this is much easier to discover on implementation rather than in theory on a pr which in and of itself is a problem other times they are dealing with ci issues the ui tries to satisfy design requirements but they sometimes do not match the data structure provided by the api improvements have been made here by increasing communication between uxd and the backend but sometimes as stated earlier structural issues and discrepencies arise mid implementation which is normal backend issues are discovered such as a bug or server issues with using the rest api also a lot of the data returned from the api is not relevant and the structure is inconsistent across features timing the rest api uxd teams are essentially trying to work in parallel with the ui team which puts an unnecessary amount of pressure on the former teams to get things completed in time for implementation in the ui and early enough for qe and docs to also make adjustments a deadline the ui often can t meet parallel should be our goal but execution is often linear and blocking since much of the sprint time in the ui is spent creating components and templates for the design implementation this is often discovered at the very end of the sprint when wiring up the api the ui should be able to operate strictly with a known data structure and rough designs whether or not that has been implemented in the rest api is irrelevant the data structure should be known before a single line of code is written on either side but who dictates what the structure should be at the moment it is usually the feature owner which is usually a backend engineer there is nothing wrong with that but if we choose to go that route some adjustments need to be made proposed solutions feel free to add to this decide where the data structure should come from considering that ipaas is heavily user focused and technically saas to me it makes sense that it either comes completely from the front end or is mediated between the api and ui more on that below decide how that data structure is decided on consumer driven contracts as zregvart suggested for spring a pattern to adhere to an evolving service and can come in many forms as simple as a spreadsheet ui driven data models to which the api then adapts this doesn t mean ui engineers need to be feature owners but just that the ui portion would dictate the structure this would allow us to mock out data knowing that when we switch to the rest api it will be a seamless transition in theory using an intermediary solution like and here is a full stack not a huge fan but it could work putting a freeze on design proposals regardless of how bad the structure is and reconvening the next sprint syndesisio all let s discuss benefits and tradeoffs other possible solutions and any considerations below
| 1
|
752,266
| 26,278,352,655
|
IssuesEvent
|
2023-01-07 02:39:46
|
deckhouse/deckhouse
|
https://api.github.com/repos/deckhouse/deckhouse
|
closed
|
Do not use FQDN for service discovery
|
type/enhancement status/rotten priority/backlog source/deckhouse-team
|
### Preflight Checklist
- [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success.
### Use case. Why is this important?
FQDN service names are not required to access services. It is ok to use the combination of service name and namespace.
Using full names can cause problems during cluster domain migrations, e.g. if a user wants to migrate from the `cluster.local` domain to something specific.
### Proposed Solution
Audit the code of Deckhouse and find out where we use full names:
1. `external_auth.go` hook https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/modules/300-prometheus/hooks/external_auth.go#L24-L27
2. Ingress resource https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/modules/300-prometheus/templates/prometheus/ingress.yaml#L12-L15
3. kube-state-metrics address to scrape https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/modules/340-monitoring-kubernetes/templates/kube-state-metrics/additional-scrape-config.yaml#L12
4. Madison - Grafana connection
https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/ee/modules/600-flant-integration/templates/madison/service.yaml#L42-L45
There may be more.
### Additional Information
_No response_
|
1.0
|
Do not use FQDN for service discovery - ### Preflight Checklist
- [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to.
- [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success.
### Use case. Why is this important?
FQDN service names are not required to access services. It is ok to use the combination of service name and namespace.
Using full names can cause problems during cluster domain migrations, e.g. if a user wants to migrate from the `cluster.local` domain to something specific.
### Proposed Solution
Audit the code of Deckhouse and find out where we use full names:
1. `external_auth.go` hook https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/modules/300-prometheus/hooks/external_auth.go#L24-L27
2. Ingress resource https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/modules/300-prometheus/templates/prometheus/ingress.yaml#L12-L15
3. kube-state-metrics address to scrape https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/modules/340-monitoring-kubernetes/templates/kube-state-metrics/additional-scrape-config.yaml#L12
4. Madison - Grafana connection
https://github.com/deckhouse/deckhouse/blob/415d874918a4dce4d2e4302da57f9eb67b9207c9/ee/modules/600-flant-integration/templates/madison/service.yaml#L42-L45
There may be more.
### Additional Information
_No response_
|
non_process
|
do not use fqdn for service discovery preflight checklist i agree to follow the that this project adheres to i have searched the for an issue that matches the one i want to file without success use case why is this important fqdn service names are not required to access services it is ok to use the combination of service name and namespace using full names can cause problems during cluster domain migrations e g if a user wants to migrate from the cluster local domain to something specific proposed solution audit the code of deckhouse and find out where we use full names external auth go hook ingress resource kube state metrics address to scrape madison grafana connection there may be more additional information no response
| 0
|
292,325
| 25,206,555,215
|
IssuesEvent
|
2022-11-13 18:58:37
|
MinhazMurks/Bannerlord.Tweaks
|
https://api.github.com/repos/MinhazMurks/Bannerlord.Tweaks
|
opened
|
Test Castle Training Fields Experience Level 3
|
testing
|
Test to see if tweak: "Castle Training Fields Experience Level 3" works
|
1.0
|
Test Castle Training Fields Experience Level 3 - Test to see if tweak: "Castle Training Fields Experience Level 3" works
|
non_process
|
test castle training fields experience level test to see if tweak castle training fields experience level works
| 0
|
18,405
| 10,109,734,248
|
IssuesEvent
|
2019-07-30 08:42:43
|
influxdata/influxdb
|
https://api.github.com/repos/influxdata/influxdb
|
closed
|
Uncontrolled memory consumption
|
1.x performance wontfix
|
### Bug report
Host type : VM
OS : Ubuntu 16.04
4 vCPU
25 GB of memory
Influxdb data disk : 500GB of SSD
Influxdb version : 1.0.0-beta3
**Steps to reproduce:**
We have multiple source of data :
- telegraf (50GB of data)
- prometheus (99 GB of data)
- heapster (< 3GB of data)
**Expected behavior:**
We expect influxdb not to consume more memory than available in the system.
**Actual behavior:**
Very few minutes after startup, influxdb is eating more memory than the 25GB of RAM and get OOM killed.
This system has been working in production with great stability for a long time.
We do not use yet CQ so we are basically appending data.
We tried to purge some measurements but Influxdb got killed before dropping data.
**Gists:**
block : https://gist.github.com/rvrignaud/40ec0a3e2a36356e85a108d52d797d2b
goroutines : https://gist.github.com/rvrignaud/c8a96202628f0ec39518246d43e99961
iostat : https://gist.github.com/rvrignaud/920bc3d158d46b23d05d68ddf8bffdbe
vars : https://gist.github.com/rvrignaud/25c51d774f817c493ce7e6f9c2753e14
|
True
|
Uncontrolled memory consumption - ### Bug report
Host type : VM
OS : Ubuntu 16.04
4 vCPU
25 GB of memory
Influxdb data disk : 500GB of SSD
Influxdb version : 1.0.0-beta3
**Steps to reproduce:**
We have multiple source of data :
- telegraf (50GB of data)
- prometheus (99 GB of data)
- heapster (< 3GB of data)
**Expected behavior:**
We expect influxdb not to consume more memory than available in the system.
**Actual behavior:**
Very few minutes after startup, influxdb is eating more memory than the 25GB of RAM and get OOM killed.
This system has been working in production with great stability for a long time.
We do not use yet CQ so we are basically appending data.
We tried to purge some measurements but Influxdb got killed before dropping data.
**Gists:**
block : https://gist.github.com/rvrignaud/40ec0a3e2a36356e85a108d52d797d2b
goroutines : https://gist.github.com/rvrignaud/c8a96202628f0ec39518246d43e99961
iostat : https://gist.github.com/rvrignaud/920bc3d158d46b23d05d68ddf8bffdbe
vars : https://gist.github.com/rvrignaud/25c51d774f817c493ce7e6f9c2753e14
|
non_process
|
uncontrolled memory consumption bug report host type vm os ubuntu vcpu gb of memory influxdb data disk of ssd influxdb version steps to reproduce we have multiple source of data telegraf of data prometheus gb of data heapster of data expected behavior we expect influxdb not to consume more memory than available in the system actual behavior very few minutes after startup influxdb is eating more memory than the of ram and get oom killed this system has been working in production with great stability for a long time we do not use yet cq so we are basically appending data we tried to purge some measurements but influxdb got killed before dropping data gists block goroutines iostat vars
| 0
|
6,893
| 10,036,432,852
|
IssuesEvent
|
2019-07-18 10:38:48
|
nbh-digital/goldchain
|
https://api.github.com/repos/nbh-digital/goldchain
|
closed
|
Web UI for authorizing GFT addresses
|
process_wontfix type_feature
|
Web UI, can only be operated by NBH employees through wallet addresses created by them.
Multisign 1 on x (x being all NBH addresses).
|
1.0
|
Web UI for authorizing GFT addresses - Web UI, can only be operated by NBH employees through wallet addresses created by them.
Multisign 1 on x (x being all NBH addresses).
|
process
|
web ui for authorizing gft addresses web ui can only be operated by nbh employees through wallet addresses created by them multisign on x x being all nbh addresses
| 1
|
185,913
| 21,876,277,384
|
IssuesEvent
|
2022-05-19 10:25:46
|
turkdevops/graphql-tools
|
https://api.github.com/repos/turkdevops/graphql-tools
|
closed
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- eslint-plugin-import-2.23.4.tgz (Root Library)
- read-pkg-up-3.0.0.tgz
- read-pkg-3.0.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/graphql-tools/commit/522129ce265decf86028571eea566ef21c50fd7f">522129ce265decf86028571eea566ef21c50fd7f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution (hosted-git-info): 2.8.9</p>
<p>Direct dependency fix Resolution (eslint-plugin-import): 2.24.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - autoclosed - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- eslint-plugin-import-2.23.4.tgz (Root Library)
- read-pkg-up-3.0.0.tgz
- read-pkg-3.0.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/graphql-tools/commit/522129ce265decf86028571eea566ef21c50fd7f">522129ce265decf86028571eea566ef21c50fd7f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution (hosted-git-info): 2.8.9</p>
<p>Direct dependency fix Resolution (eslint-plugin-import): 2.24.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in hosted git info tgz autoclosed cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file package json path to vulnerable library node modules hosted git info package json dependency hierarchy eslint plugin import tgz root library read pkg up tgz read pkg tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info direct dependency fix resolution eslint plugin import step up your open source security game with whitesource
| 0
|
21,544
| 29,865,128,452
|
IssuesEvent
|
2023-06-20 02:39:01
|
cncf/tag-security
|
https://api.github.com/repos/cncf/tag-security
|
closed
|
guidelines for prioritizing projects for security assessements
|
assessment-process inactive
|
We need guidelines on how to prioritize projects for security assessments when there are multiple projects interested in engaging with us on an assessment.
see [draft guidelines](https://docs.google.com/document/d/1RUzTBHFqpEL27RvSRMJVcbQtqDj1v78eOu-5xEvVw6I/edit#) -- these need to be moved into a PR, linking from here for visibility
Related:
* Assessment Listing https://github.com/cncf/sig-security/issues/206
* Annual review process: https://github.com/cncf/sig-security/issues/152
|
1.0
|
guidelines for prioritizing projects for security assessements - We need guidelines on how to prioritize projects for security assessments when there are multiple projects interested in engaging with us on an assessment.
see [draft guidelines](https://docs.google.com/document/d/1RUzTBHFqpEL27RvSRMJVcbQtqDj1v78eOu-5xEvVw6I/edit#) -- these need to be moved into a PR, linking from here for visibility
Related:
* Assessment Listing https://github.com/cncf/sig-security/issues/206
* Annual review process: https://github.com/cncf/sig-security/issues/152
|
process
|
guidelines for prioritizing projects for security assessements we need guidelines on how to prioritize projects for security assessments when there are multiple projects interested in engaging with us on an assessment see these need to be moved into a pr linking from here for visibility related assessment listing annual review process
| 1
|
14,838
| 18,234,226,142
|
IssuesEvent
|
2021-10-01 03:38:02
|
edmobe/android-video-magnification
|
https://api.github.com/repos/edmobe/android-video-magnification
|
closed
|
OB-1004 La magnificación de vídeo continúa en segundo plano
|
video-reception video-processing video-output obstacle
|
Es recomendable que la magnificación se detenga, ya que si el usuario cambia de actividad para modificar algún parámetro de magnificación, esta continúa ejecutándose, consumiendo recursos y obstruyendo posteriores ejecuciones.
|
1.0
|
OB-1004 La magnificación de vídeo continúa en segundo plano - Es recomendable que la magnificación se detenga, ya que si el usuario cambia de actividad para modificar algún parámetro de magnificación, esta continúa ejecutándose, consumiendo recursos y obstruyendo posteriores ejecuciones.
|
process
|
ob la magnificación de vídeo continúa en segundo plano es recomendable que la magnificación se detenga ya que si el usuario cambia de actividad para modificar algún parámetro de magnificación esta continúa ejecutándose consumiendo recursos y obstruyendo posteriores ejecuciones
| 1
|
621,481
| 19,588,057,921
|
IssuesEvent
|
2022-01-05 09:37:52
|
apache/dolphinscheduler
|
https://api.github.com/repos/apache/dolphinscheduler
|
closed
|
[Bug] [MasterServer] NPE in netty processor
|
bug priority-high
|
### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
### What happened
```
[ERROR] 2021-12-10 16:36:58.596 org.apache.dolphinscheduler.remote.handler.NettyClientHandler:[156] - process command Command [type=TASK_EXECUTE_RESPONSE, opaque=13, bodyLen=130] exception
java.lang.NullPointerException: null
at org.apache.dolphinscheduler.server.master.processor.TaskResponseProcessor.process(TaskResponseProcessor.java:72)
at org.apache.dolphinscheduler.remote.handler.NettyClientHandler.lambda$processByCommandType$0(NettyClientHandler.java:154)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
```
### What you expected to happen
no error
### How to reproduce
just run a process instance.
### Anything else
_No response_
### Version
dev
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
|
1.0
|
[Bug] [MasterServer] NPE in netty processor - ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
### What happened
```
[ERROR] 2021-12-10 16:36:58.596 org.apache.dolphinscheduler.remote.handler.NettyClientHandler:[156] - process command Command [type=TASK_EXECUTE_RESPONSE, opaque=13, bodyLen=130] exception
java.lang.NullPointerException: null
at org.apache.dolphinscheduler.server.master.processor.TaskResponseProcessor.process(TaskResponseProcessor.java:72)
at org.apache.dolphinscheduler.remote.handler.NettyClientHandler.lambda$processByCommandType$0(NettyClientHandler.java:154)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
```
### What you expected to happen
no error
### How to reproduce
just run a process instance.
### Anything else
_No response_
### Version
dev
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
|
non_process
|
npe in netty processor search before asking i had searched in the and found no similar issues what happened org apache dolphinscheduler remote handler nettyclienthandler process command command exception java lang nullpointerexception null at org apache dolphinscheduler server master processor taskresponseprocessor process taskresponseprocessor java at org apache dolphinscheduler remote handler nettyclienthandler lambda processbycommandtype nettyclienthandler java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java what you expected to happen no error how to reproduce just run a process instance anything else no response version dev are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s
| 0
|
18,607
| 3,391,753,250
|
IssuesEvent
|
2015-11-30 16:42:35
|
18F/atf-eregs
|
https://api.github.com/repos/18F/atf-eregs
|
closed
|
Present HD nodes appropriately in UI
|
design
|
I think we're currently handling these reasonably well, but we should examine them and make sure we're happy with the current approach.
|
1.0
|
Present HD nodes appropriately in UI - I think we're currently handling these reasonably well, but we should examine them and make sure we're happy with the current approach.
|
non_process
|
present hd nodes appropriately in ui i think we re currently handling these reasonably well but we should examine them and make sure we re happy with the current approach
| 0
|
12,223
| 14,743,175,151
|
IssuesEvent
|
2021-01-07 13:31:00
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Need to Track the Cron Job folder in Git
|
anc-process anp-1 ant-feature
|
In GitLab by @pchaudhary on Jul 30, 2019, 07:31
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-08-14-19954/conversation
From #1507
[Tim] (@tim.traylor ) wrote - Sumeet, as per our discussion, please identify the mechanism by which these OS-level errors are trapped and reported. For an issue like this, where a cron job can't run (and it turn can't tell us there was an error), we need the OS to send us a notification at our criticalerrors@sahosted.com email address. Thx
|
1.0
|
Need to Track the Cron Job folder in Git - In GitLab by @pchaudhary on Jul 30, 2019, 07:31
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-08-14-19954/conversation
From #1507
[Tim] (@tim.traylor ) wrote - Sumeet, as per our discussion, please identify the mechanism by which these OS-level errors are trapped and reported. For an issue like this, where a cron job can't run (and it turn can't tell us there was an error), we need the OS to send us a notification at our criticalerrors@sahosted.com email address. Thx
|
process
|
need to track the cron job folder in git in gitlab by pchaudhary on jul helpdesk from tim traylor wrote sumeet as per our discussion please identify the mechanism by which these os level errors are trapped and reported for an issue like this where a cron job can t run and it turn can t tell us there was an error we need the os to send us a notification at our criticalerrors sahosted com email address thx
| 1
|
60,206
| 25,032,162,179
|
IssuesEvent
|
2022-11-04 13:18:35
|
astropy/astroquery
|
https://api.github.com/repos/astropy/astroquery
|
opened
|
Isochrone & Stellar modeling services
|
New Service
|
We could include interfaces to web services that provide isochrone model sets, such as MIST, Padova, Yale, etc.
We've mentioned this [before](https://github.com/astropy/astroquery/issues/954#issuecomment-991974614), but we didn't have an open issue for it.
The services I know of are:
* [ ] [MIST](https://waps.cfa.harvard.edu/MIST/)
* [ ] [Padova](http://stev.oapd.inaf.it/cgi-bin/cmd)
* [ ] [Yale](http://www.astro.yale.edu/yapsi/download_grids.html) (no web interface - they just let you download the grids)
If you know of more, please add them.
The fact that you can download and interpolate (some of) these yourself somewhat limits the need for an astroquery service, but at least some of these modeling groups regularly update their grids with new models or new filter sets, so it is probably still worthwhile to support web access.
|
1.0
|
Isochrone & Stellar modeling services - We could include interfaces to web services that provide isochrone model sets, such as MIST, Padova, Yale, etc.
We've mentioned this [before](https://github.com/astropy/astroquery/issues/954#issuecomment-991974614), but we didn't have an open issue for it.
The services I know of are:
* [ ] [MIST](https://waps.cfa.harvard.edu/MIST/)
* [ ] [Padova](http://stev.oapd.inaf.it/cgi-bin/cmd)
* [ ] [Yale](http://www.astro.yale.edu/yapsi/download_grids.html) (no web interface - they just let you download the grids)
If you know of more, please add them.
The fact that you can download and interpolate (some of) these yourself somewhat limits the need for an astroquery service, but at least some of these modeling groups regularly update their grids with new models or new filter sets, so it is probably still worthwhile to support web access.
|
non_process
|
isochrone stellar modeling services we could include interfaces to web services that provide isochrone model sets such as mist padova yale etc we ve mentioned this but we didn t have an open issue for it the services i know of are no web interface they just let you download the grids if you know of more please add them the fact that you can download and interpolate some of these yourself somewhat limits the need for an astroquery service but at least some of these modeling groups regularly update their grids with new models or new filter sets so it is probably still worthwhile to support web access
| 0
|
14,921
| 18,359,528,387
|
IssuesEvent
|
2021-10-09 01:45:36
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Script processing error with with `foreach` keyword in Knockout.js
|
TYPE: bug AREA: client FREQUENCY: level 1 SYSTEM: client side processing STATE: Stale
|
Error details:
```
JavaScript error details:
Error: Unable to process binding "foreach: function(){return tabs }"
Message: Unable to process binding "template: function(){return { name:template,data:data,afterRender:function(element,model){
$parent.onEditorTabRenderedCallback(name,element,model,$data);}} }"
```
Here is an example to reproduce the issue.
[example.zip](https://github.com/DevExpress/testcafe-hammerhead/files/4330351/example.zip)
Set the `showLogicTab` property on the `index.html` page to `true` to get the error under playground.
If you set the `showLogicTab` property the `false` value, the error will not happen
|
1.0
|
Script processing error with with `foreach` keyword in Knockout.js - Error details:
```
JavaScript error details:
Error: Unable to process binding "foreach: function(){return tabs }"
Message: Unable to process binding "template: function(){return { name:template,data:data,afterRender:function(element,model){
$parent.onEditorTabRenderedCallback(name,element,model,$data);}} }"
```
Here is an example to reproduce the issue.
[example.zip](https://github.com/DevExpress/testcafe-hammerhead/files/4330351/example.zip)
Set the `showLogicTab` property on the `index.html` page to `true` to get the error under playground.
If you set the `showLogicTab` property the `false` value, the error will not happen
|
process
|
script processing error with with foreach keyword in knockout js error details javascript error details error unable to process binding foreach function return tabs message unable to process binding template function return name template data data afterrender function element model parent oneditortabrenderedcallback name element model data here is an example to reproduce the issue set the showlogictab property on the index html page to true to get the error under playground if you set the showlogictab property the false value the error will not happen
| 1
|
26,527
| 20,194,313,902
|
IssuesEvent
|
2022-02-11 09:15:05
|
SonarSource/sonar-scanner-msbuild
|
https://api.github.com/repos/SonarSource/sonar-scanner-msbuild
|
opened
|
Compute conditional (branch) code coverage for S4NET
|
Infrastructure
|
Currently the VS Coverage does not support conditional coverage (well, it does in a very limited fashion, saying that a line is partially covered, but that's not helpful enough for SQ/SC, which needs the number of conditions and the number of covered conditions).
We should not use OpenCover as its been deprecated.
We should probably use coverlet to compute code coverage.
|
1.0
|
Compute conditional (branch) code coverage for S4NET - Currently the VS Coverage does not support conditional coverage (well, it does in a very limited fashion, saying that a line is partially covered, but that's not helpful enough for SQ/SC, which needs the number of conditions and the number of covered conditions).
We should not use OpenCover as its been deprecated.
We should probably use coverlet to compute code coverage.
|
non_process
|
compute conditional branch code coverage for currently the vs coverage does not support conditional coverage well it does in a very limited fashion saying that a line is partially covered but that s not helpful enough for sq sc which needs the number of conditions and the number of covered conditions we should not use opencover as its been deprecated we should probably use coverlet to compute code coverage
| 0
|
10,696
| 13,492,184,972
|
IssuesEvent
|
2020-09-11 17:38:54
|
GetTerminus/terminus-oss
|
https://api.github.com/repos/GetTerminus/terminus-oss
|
closed
|
Add remaining stories
|
Focus: consumer Goal: Process Improvement Type: chore
|
- [x] autofocus
- [x] cohort date range
- [x] date range
- [x] file upload
- [x] spacing
- [x] toggle
- [x] validators
|
1.0
|
Add remaining stories - - [x] autofocus
- [x] cohort date range
- [x] date range
- [x] file upload
- [x] spacing
- [x] toggle
- [x] validators
|
process
|
add remaining stories autofocus cohort date range date range file upload spacing toggle validators
| 1
|
212,380
| 23,884,112,277
|
IssuesEvent
|
2022-09-08 05:55:09
|
nidhi7598/packages_providers_MediaProvider_AOSP_10_r33
|
https://api.github.com/repos/nidhi7598/packages_providers_MediaProvider_AOSP_10_r33
|
opened
|
CVE-2021-0340 (High) detected in MediaProviderandroid-10.0.0_r31
|
security vulnerability
|
## CVE-2021-0340 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>MediaProviderandroid-10.0.0_r31</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/providers/MediaProvider>https://android.googlesource.com/platform/packages/providers/MediaProvider</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/packages_providers_MediaProvider_AOSP_10_r33/commit/6cff788f705b2cd5f95ca8d0f047922fedd0f53b">6cff788f705b2cd5f95ca8d0f047922fedd0f53b</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/providers/media/util/IsoInterface.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In parseNextBox of IsoInterface.java, there is a possible leak of unredacted location information due to improper input validation. This could lead to remote information disclosure with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-134155286
<p>Publish Date: 2021-02-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0340>CVE-2021-0340</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2021-02-01">https://source.android.com/security/bulletin/2021-02-01</a></p>
<p>Release Date: 2021-02-10</p>
<p>Fix Resolution: android-11.0.0_r1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-0340 (High) detected in MediaProviderandroid-10.0.0_r31 - ## CVE-2021-0340 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>MediaProviderandroid-10.0.0_r31</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/providers/MediaProvider>https://android.googlesource.com/platform/packages/providers/MediaProvider</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/packages_providers_MediaProvider_AOSP_10_r33/commit/6cff788f705b2cd5f95ca8d0f047922fedd0f53b">6cff788f705b2cd5f95ca8d0f047922fedd0f53b</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/providers/media/util/IsoInterface.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In parseNextBox of IsoInterface.java, there is a possible leak of unredacted location information due to improper input validation. This could lead to remote information disclosure with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-134155286
<p>Publish Date: 2021-02-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0340>CVE-2021-0340</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2021-02-01">https://source.android.com/security/bulletin/2021-02-01</a></p>
<p>Release Date: 2021-02-10</p>
<p>Fix Resolution: android-11.0.0_r1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in mediaproviderandroid cve high severity vulnerability vulnerable library mediaproviderandroid library home page a href found in head commit a href found in base branch master vulnerable source files src com android providers media util isointerface java vulnerability details in parsenextbox of isointerface java there is a possible leak of unredacted location information due to improper input validation this could lead to remote information disclosure with no additional execution privileges needed user interaction is needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
| 0
|
2,142
| 4,990,027,873
|
IssuesEvent
|
2016-12-08 13:55:52
|
jbarnoud/ALLATOM
|
https://api.github.com/repos/jbarnoud/ALLATOM
|
closed
|
Identify protocols that already ran
|
Component-Main program processus running
|
This will avoid running the same protocol twice. It is mandatory to resume a test suite.
|
1.0
|
Identify protocols that already ran - This will avoid running the same protocol twice. It is mandatory to resume a test suite.
|
process
|
identify protocols that already ran this will avoid running the same protocol twice it is mandatory to resume a test suite
| 1
|
118,046
| 4,731,417,162
|
IssuesEvent
|
2016-10-19 01:58:02
|
RTICWDT/college-scorecard
|
https://api.github.com/repos/RTICWDT/college-scorecard
|
closed
|
School Basics / Size: Small, Medium, Large
|
Area - Consumer Tool Bang 1 - Low Bang For Buck 1 - Low Buck 1 - Low Priority 2 Scrub (Sabrina) Stage 2 - Research Theme 3 - Students To Value Metrics
|
2000 is kind of a low cut-off for small schools. I would suggest raising it to 3000 or 4000.
|
1.0
|
School Basics / Size: Small, Medium, Large - 2000 is kind of a low cut-off for small schools. I would suggest raising it to 3000 or 4000.
|
non_process
|
school basics size small medium large is kind of a low cut off for small schools i would suggest raising it to or
| 0
|
8,964
| 12,069,743,642
|
IssuesEvent
|
2020-04-16 16:30:36
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
How to make parameter optional
|
Pri1 cba devops-cicd-process/tech devops/prod support-request
|
How to make parameter optional? I want to implement stepList optional parameter. I tried to add "default: []", but I can't to check this later when add template based on this:
parameters:
- name: stagename
- name: sqlsteps
type: stepList
default: []
stages:
- stage: ${{ parameters.stagename }}
jobs:
...
// I want to exclude this template if stepList wan't passed
- template: tmplt-job-deploysql.yml
parameters:
jobname: deploysql
sqlsteps: ${{ parameters.sqlsteps }}
...
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#parameters)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
How to make parameter optional - How to make parameter optional? I want to implement stepList optional parameter. I tried to add "default: []", but I can't to check this later when add template based on this:
parameters:
- name: stagename
- name: sqlsteps
type: stepList
default: []
stages:
- stage: ${{ parameters.stagename }}
jobs:
...
// I want to exclude this template if stepList wan't passed
- template: tmplt-job-deploysql.yml
parameters:
jobname: deploysql
sqlsteps: ${{ parameters.sqlsteps }}
...
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#parameters)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
how to make parameter optional how to make parameter optional i want to implement steplist optional parameter i tried to add default but i can t to check this later when add template based on this parameters name stagename name sqlsteps type steplist default stages stage parameters stagename jobs i want to exclude this template if steplist wan t passed template tmplt job deploysql yml parameters jobname deploysql sqlsteps parameters sqlsteps document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
264,436
| 23,119,505,924
|
IssuesEvent
|
2022-07-27 19:53:18
|
CryoVR/CryoVR_Leica-EM-GP2
|
https://api.github.com/repos/CryoVR/CryoVR_Leica-EM-GP2
|
closed
|
Overall height.
|
Ready to test
|
The height of the table is too high. Maybe lower the table 10-20cm and it will be better.
|
1.0
|
Overall height. - The height of the table is too high. Maybe lower the table 10-20cm and it will be better.
|
non_process
|
overall height the height of the table is too high maybe lower the table and it will be better
| 0
|
647,742
| 21,154,874,420
|
IssuesEvent
|
2022-04-07 01:20:13
|
papertek/CyDJ
|
https://api.github.com/repos/papertek/CyDJ
|
closed
|
Update layout to include more contrast
|
enhancement Priority 3 low effort required
|
currently thinking to use more pink and yellow and light green since it looks really nice when combined
### reference designs
**non hovering**

**elements hovered**

|
1.0
|
Update layout to include more contrast - currently thinking to use more pink and yellow and light green since it looks really nice when combined
### reference designs
**non hovering**

**elements hovered**

|
non_process
|
update layout to include more contrast currently thinking to use more pink and yellow and light green since it looks really nice when combined reference designs non hovering elements hovered
| 0
|
284,207
| 21,408,482,414
|
IssuesEvent
|
2022-04-22 01:17:40
|
JosephLoRusso/Memento-Mori
|
https://api.github.com/repos/JosephLoRusso/Memento-Mori
|
closed
|
Create presentation
|
documentation
|
Record the website functioning; we should not try to do a live demo. Prepare presentation.
|
1.0
|
Create presentation - Record the website functioning; we should not try to do a live demo. Prepare presentation.
|
non_process
|
create presentation record the website functioning we should not try to do a live demo prepare presentation
| 0
|
5,592
| 8,444,728,120
|
IssuesEvent
|
2018-10-18 19:18:05
|
aspnet/IISIntegration
|
https://api.github.com/repos/aspnet/IISIntegration
|
closed
|
In some cases we compress/decompress/compress again, heating up CPUs
|
bug out-of-process
|
We're doing unnecessary work with compression.
|
1.0
|
In some cases we compress/decompress/compress again, heating up CPUs - We're doing unnecessary work with compression.
|
process
|
in some cases we compress decompress compress again heating up cpus we re doing unnecessary work with compression
| 1
|
92,900
| 11,724,503,615
|
IssuesEvent
|
2020-03-10 11:05:13
|
celo-org/celo-monorepo
|
https://api.github.com/repos/celo-org/celo-monorepo
|
closed
|
Celostats Beta Design Optimizations
|
celostats need-design protocol-tools
|
- [x] The contrast of the upper panels is weak and the gradients make it a bit noisy. Maybe opt for a more simple color scheme in line with the Celo colors.
- [x] Update title and favicon
Thoughts on design:
- The initial pan in of components feels a little gimicky
- The fonts (sans vs serif) and colours dont line up with either blockscout or our web site (all three are different!)
- The monospace font is very clear and i prefer words like “Elected” rather than images (probably especially if im not looking at this page every day!)
- It’s unclear what the colors of the top panels represent. Dimi mentioned this above as well—I think the gradient’s make this a little ambiguous (ie. a panel is sometimes both green and yellow, some are blue). If a panel’s data is “neutral” (doesn’t have any good/bad -> green/red meaning associated with it), perhaps just remove the background.
- Agree with Tim’s comments above regarding design:
- Load animation feels unnecessary
- Ideally, align the fonts to https://celo.org/experience/brand/typography#overview (the monospace seems helpful and fine)
- Since we’re displaying a lot of data, could be helpful to highlight a row on hover to aid eye travel from one end to the other
- Font readable
- Overall layout feels overwhelming, crowded.
- Make colors pop out more?
- Agree w/ TIm feels a bit gimmicky and look & feel not aligned with other Celo sites
- Celostats dashboard should contain options to toggle pages to see different views.
- Beginner or Intro View for any interested users, Validator View, and Developer View.
- When I load up the page, I should be able to login to see my proxies and validators at the top.
- I should be able to get an alert when my validator was elected.
- Be able to see my validator performance compared to other validators.
|
1.0
|
Celostats Beta Design Optimizations - - [x] The contrast of the upper panels is weak and the gradients make it a bit noisy. Maybe opt for a more simple color scheme in line with the Celo colors.
- [x] Update title and favicon
Thoughts on design:
- The initial pan in of components feels a little gimicky
- The fonts (sans vs serif) and colours dont line up with either blockscout or our web site (all three are different!)
- The monospace font is very clear and i prefer words like “Elected” rather than images (probably especially if im not looking at this page every day!)
- It’s unclear what the colors of the top panels represent. Dimi mentioned this above as well—I think the gradient’s make this a little ambiguous (ie. a panel is sometimes both green and yellow, some are blue). If a panel’s data is “neutral” (doesn’t have any good/bad -> green/red meaning associated with it), perhaps just remove the background.
- Agree with Tim’s comments above regarding design:
- Load animation feels unnecessary
- Ideally, align the fonts to https://celo.org/experience/brand/typography#overview (the monospace seems helpful and fine)
- Since we’re displaying a lot of data, could be helpful to highlight a row on hover to aid eye travel from one end to the other
- Font readable
- Overall layout feels overwhelming, crowded.
- Make colors pop out more?
- Agree w/ TIm feels a bit gimmicky and look & feel not aligned with other Celo sites
- Celostats dashboard should contain options to toggle pages to see different views.
- Beginner or Intro View for any interested users, Validator View, and Developer View.
- When I load up the page, I should be able to login to see my proxies and validators at the top.
- I should be able to get an alert when my validator was elected.
- Be able to see my validator performance compared to other validators.
|
non_process
|
celostats beta design optimizations the contrast of the upper panels is weak and the gradients make it a bit noisy maybe opt for a more simple color scheme in line with the celo colors update title and favicon thoughts on design the initial pan in of components feels a little gimicky the fonts sans vs serif and colours dont line up with either blockscout or our web site all three are different the monospace font is very clear and i prefer words like “elected” rather than images probably especially if im not looking at this page every day it’s unclear what the colors of the top panels represent dimi mentioned this above as well—i think the gradient’s make this a little ambiguous ie a panel is sometimes both green and yellow some are blue if a panel’s data is “neutral” doesn’t have any good bad green red meaning associated with it perhaps just remove the background agree with tim’s comments above regarding design load animation feels unnecessary ideally align the fonts to the monospace seems helpful and fine since we’re displaying a lot of data could be helpful to highlight a row on hover to aid eye travel from one end to the other font readable overall layout feels overwhelming crowded make colors pop out more agree w tim feels a bit gimmicky and look feel not aligned with other celo sites celostats dashboard should contain options to toggle pages to see different views beginner or intro view for any interested users validator view and developer view when i load up the page i should be able to login to see my proxies and validators at the top i should be able to get an alert when my validator was elected be able to see my validator performance compared to other validators
| 0
|
16,537
| 21,564,111,881
|
IssuesEvent
|
2022-05-01 16:01:37
|
uncrustify/uncrustify
|
https://api.github.com/repos/uncrustify/uncrustify
|
closed
|
Crash with mod_enum_last_comma = add when input file contains __VA_ARGS__
|
C and C++11 Preprocessor
|
Since uncrustify 0.74, the following code and config make it crash:
``` test.h
#define MY_DEF(Type, ...) \
enum Type { \
__VA_ARGS__, \
};
```
```uncrustify.cfg
mod_enum_last_comma = add
```
|
1.0
|
Crash with mod_enum_last_comma = add when input file contains __VA_ARGS__ - Since uncrustify 0.74, the following code and config make it crash:
``` test.h
#define MY_DEF(Type, ...) \
enum Type { \
__VA_ARGS__, \
};
```
```uncrustify.cfg
mod_enum_last_comma = add
```
|
process
|
crash with mod enum last comma add when input file contains va args since uncrustify the following code and config make it crash test h define my def type enum type va args uncrustify cfg mod enum last comma add
| 1
|
404,313
| 11,855,041,260
|
IssuesEvent
|
2020-03-25 02:53:38
|
Ktt-Development/simplehttpserver
|
https://api.github.com/repos/Ktt-Development/simplehttpserver
|
closed
|
Jitpack does not compile correctly
|
priority
|
### Prerequisites
*If **all** checks are not passed then the issue will be closed*
- [x] I have checked that no other similar issue already exists
**Operating System:** *Operating system name and version*
Windows 10
**Release Version:** *Release version or branch where the issue occurred*
v02.01.00
### Issue
*Explain your issue. Add any screenshots here*
Jitpack compile has only meta inf and it is the wrong one.
### Expected Behavior
*Explain what was supposed to happen*
Files should compile.
### Steps To Reproduce
*Explain how and/or when the error occurred*
Import build v02.01.00
|
1.0
|
Jitpack does not compile correctly - ### Prerequisites
*If **all** checks are not passed then the issue will be closed*
- [x] I have checked that no other similar issue already exists
**Operating System:** *Operating system name and version*
Windows 10
**Release Version:** *Release version or branch where the issue occurred*
v02.01.00
### Issue
*Explain your issue. Add any screenshots here*
Jitpack compile has only meta inf and it is the wrong one.
### Expected Behavior
*Explain what was supposed to happen*
Files should compile.
### Steps To Reproduce
*Explain how and/or when the error occurred*
Import build v02.01.00
|
non_process
|
jitpack does not compile correctly prerequisites if all checks are not passed then the issue will be closed i have checked that no other similar issue already exists operating system operating system name and version windows release version release version or branch where the issue occurred issue explain your issue add any screenshots here jitpack compile has only meta inf and it is the wrong one expected behavior explain what was supposed to happen files should compile steps to reproduce explain how and or when the error occurred import build
| 0
|
12,891
| 15,282,794,829
|
IssuesEvent
|
2021-02-23 10:01:45
|
Today-I-Learn/backend-study
|
https://api.github.com/repos/Today-I-Learn/backend-study
|
opened
|
프로세스의 상태에 대해 설명할 수 있나요?
|
OS process
|
### 프로세스의 상태에 대해 설명할 수 있나요?
- [ ] 프로세스 각 상태에 대한 설명
### context switching(문맥교환)이 무엇인가요
- [ ] context switching의 정의
- [ ] context switching 단점
|
1.0
|
프로세스의 상태에 대해 설명할 수 있나요? - ### 프로세스의 상태에 대해 설명할 수 있나요?
- [ ] 프로세스 각 상태에 대한 설명
### context switching(문맥교환)이 무엇인가요
- [ ] context switching의 정의
- [ ] context switching 단점
|
process
|
프로세스의 상태에 대해 설명할 수 있나요 프로세스의 상태에 대해 설명할 수 있나요 프로세스 각 상태에 대한 설명 context switching 문맥교환 이 무엇인가요 context switching의 정의 context switching 단점
| 1
|
253,165
| 21,658,661,207
|
IssuesEvent
|
2022-05-06 16:37:45
|
zowe/vscode-extension-for-zowe
|
https://api.github.com/repos/zowe/vscode-extension-for-zowe
|
closed
|
Remove requirement of a testProfileData for build purposes
|
Tests
|
In order to build the source code, someone needs to create a dummy `resources/testProfileData.ts` because the tests won't compile without it.
Recommendations:
1. Split the Source and Test compilation into separate npm run scripts for ease of use and error handling.
2. Create a Test Environment class that helps manage test configuration as well as any utilities needed to be loaded prior to any test hooks (e.g. `beforeEach`, ...)
|
1.0
|
Remove requirement of a testProfileData for build purposes - In order to build the source code, someone needs to create a dummy `resources/testProfileData.ts` because the tests won't compile without it.
Recommendations:
1. Split the Source and Test compilation into separate npm run scripts for ease of use and error handling.
2. Create a Test Environment class that helps manage test configuration as well as any utilities needed to be loaded prior to any test hooks (e.g. `beforeEach`, ...)
|
non_process
|
remove requirement of a testprofiledata for build purposes in order to build the source code someone needs to create a dummy resources testprofiledata ts because the tests won t compile without it recommendations split the source and test compilation into separate npm run scripts for ease of use and error handling create a test environment class that helps manage test configuration as well as any utilities needed to be loaded prior to any test hooks e g beforeeach
| 0
|
1,145
| 28,967,327,248
|
IssuesEvent
|
2023-05-10 08:48:53
|
Swiss-Polar-Institute/project-application
|
https://api.github.com/repos/Swiss-Polar-Institute/project-application
|
closed
|
Check that applicant, overarching project leader and proposal partners have no duplicates
|
enhancement before next call opens proposal people
|
When creating a proposal right now the applicant could enter the same physical person (orcid) as applicant and overarching project leader.
It could also enter two proposal partners to be the same person.
On the proposal validation: make sure that the the applicant and overarching project leader are not the same person. Show an error message if this is the case and avoid saving/submitting the proposal
The same in case that the same proposal partner is entered twice.
|
1.0
|
Check that applicant, overarching project leader and proposal partners have no duplicates - When creating a proposal right now the applicant could enter the same physical person (orcid) as applicant and overarching project leader.
It could also enter two proposal partners to be the same person.
On the proposal validation: make sure that the the applicant and overarching project leader are not the same person. Show an error message if this is the case and avoid saving/submitting the proposal
The same in case that the same proposal partner is entered twice.
|
non_process
|
check that applicant overarching project leader and proposal partners have no duplicates when creating a proposal right now the applicant could enter the same physical person orcid as applicant and overarching project leader it could also enter two proposal partners to be the same person on the proposal validation make sure that the the applicant and overarching project leader are not the same person show an error message if this is the case and avoid saving submitting the proposal the same in case that the same proposal partner is entered twice
| 0
|
2,537
| 5,299,939,580
|
IssuesEvent
|
2017-02-10 02:12:19
|
mitchellh/packer
|
https://api.github.com/repos/mitchellh/packer
|
closed
|
Unable to extract .zip produced by 'Compress' post-processor
|
bug post-processor/compress
|
Hello,
I am unable to extract a **.zip** file produced by the **Compress** post-processor. The error received states that it is not a valid archive file. I am using Packer along with Packer-Windows on Windows 2012 R2.
## Snippet from Win_7.json
``` json
{
"type": "compress",
"only": ["vmware-iso"],
"keep_input_artifact": false,
"output": "win_7.zip"
}
```
## Command Executed
To run Packer, I am executing `packer build -only vmware-iso win_7.json` via an elevated Poweshell prompt.
Please let me know if I can provide any further details that may assist in investigating the issue.
|
1.0
|
Unable to extract .zip produced by 'Compress' post-processor - Hello,
I am unable to extract a **.zip** file produced by the **Compress** post-processor. The error received states that it is not a valid archive file. I am using Packer along with Packer-Windows on Windows 2012 R2.
## Snippet from Win_7.json
``` json
{
"type": "compress",
"only": ["vmware-iso"],
"keep_input_artifact": false,
"output": "win_7.zip"
}
```
## Command Executed
To run Packer, I am executing `packer build -only vmware-iso win_7.json` via an elevated Poweshell prompt.
Please let me know if I can provide any further details that may assist in investigating the issue.
|
process
|
unable to extract zip produced by compress post processor hello i am unable to extract a zip file produced by the compress post processor the error received states that it is not a valid archive file i am using packer along with packer windows on windows snippet from win json json type compress only keep input artifact false output win zip command executed to run packer i am executing packer build only vmware iso win json via an elevated poweshell prompt please let me know if i can provide any further details that may assist in investigating the issue
| 1
|
100,996
| 16,490,735,574
|
IssuesEvent
|
2021-05-25 03:10:03
|
valdisiljuconoks/AlloyTech
|
https://api.github.com/repos/valdisiljuconoks/AlloyTech
|
opened
|
WS-2019-0103 (Medium) detected in handlebars-1.3.0.tgz
|
security vulnerability
|
## WS-2019-0103 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p>
<p>Path to dependency file: AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- istanbul-0.2.16.tgz
- :x: **handlebars-1.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars.js before 4.1.0 has Remote Code Execution (RCE)
<p>Publish Date: 2019-01-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/issues/1267#issue-187151586>WS-2019-0103</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57">https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57</a></p>
<p>Release Date: 2019-05-30</p>
<p>Fix Resolution: 4.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0103 (Medium) detected in handlebars-1.3.0.tgz - ## WS-2019-0103 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p>
<p>Path to dependency file: AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- istanbul-0.2.16.tgz
- :x: **handlebars-1.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Handlebars.js before 4.1.0 has Remote Code Execution (RCE)
<p>Publish Date: 2019-01-30
<p>URL: <a href=https://github.com/wycats/handlebars.js/issues/1267#issue-187151586>WS-2019-0103</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57">https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57</a></p>
<p>Release Date: 2019-05-30</p>
<p>Fix Resolution: 4.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in handlebars tgz ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file alloytech modules protected shell shell clientresources lib xstyle package json path to vulnerable library alloytech modules protected shell shell clientresources lib xstyle node modules handlebars package json dependency hierarchy intern geezer tgz root library istanbul tgz x handlebars tgz vulnerable library vulnerability details handlebars js before has remote code execution rce publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
13,555
| 16,099,458,317
|
IssuesEvent
|
2021-04-27 07:25:35
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
'Using a webhook to start a Python runbook is not supported.' note seems obsolete
|
Pri2 automation/svc cxp doc-enhancement process-automation/subsvc product-question triaged
|
'Using a webhook to start a Python runbook is not supported.' note on this page is probably obsolete. There is no such limitation mentioned on[ runbook types page](https://docs.microsoft.com/en-us/azure/automation/automation-runbook-types#python-runbooks).
And it is possible to add webhooks to Python 2 runhook at Azure.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 17130548-aaae-b56d-7300-134c7d740f17
* Version Independent ID: fc0e9934-33db-5d34-307f-6170a345583d
* Content: [Create a Python runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual-python2#feedback)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-textual-python2.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual-python2.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
'Using a webhook to start a Python runbook is not supported.' note seems obsolete -
'Using a webhook to start a Python runbook is not supported.' note on this page is probably obsolete. There is no such limitation mentioned on[ runbook types page](https://docs.microsoft.com/en-us/azure/automation/automation-runbook-types#python-runbooks).
And it is possible to add webhooks to Python 2 runhook at Azure.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 17130548-aaae-b56d-7300-134c7d740f17
* Version Independent ID: fc0e9934-33db-5d34-307f-6170a345583d
* Content: [Create a Python runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual-python2#feedback)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-textual-python2.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual-python2.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
using a webhook to start a python runbook is not supported note seems obsolete using a webhook to start a python runbook is not supported note on this page is probably obsolete there is no such limitation mentioned on and it is possible to add webhooks to python runhook at azure document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id aaae version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
344,104
| 24,798,157,274
|
IssuesEvent
|
2022-10-24 19:10:02
|
ocf/ocfstatic
|
https://api.github.com/repos/ocf/ocfstatic
|
closed
|
Extract docs to their own repo
|
documentation help wanted
|
The documentation is currently duplicated across ocfweb and ocfstatic. This means that the content of them has diverged between the two websites, with some people making PRs to add content to docs on ocfweb and some PRs to ocfstatic. I've already started a repo at https://github.com/Kalissaac/ocfdocs with the docs content and I think I've been able to merge everything together so far. On the branch `kian/extract-docs` I've done the work of using a git submodule to point the docs folder to that repo which will allow us to update docs in one place and have it take effect everywhere on the site.
|
1.0
|
Extract docs to their own repo - The documentation is currently duplicated across ocfweb and ocfstatic. This means that the content of them has diverged between the two websites, with some people making PRs to add content to docs on ocfweb and some PRs to ocfstatic. I've already started a repo at https://github.com/Kalissaac/ocfdocs with the docs content and I think I've been able to merge everything together so far. On the branch `kian/extract-docs` I've done the work of using a git submodule to point the docs folder to that repo which will allow us to update docs in one place and have it take effect everywhere on the site.
|
non_process
|
extract docs to their own repo the documentation is currently duplicated across ocfweb and ocfstatic this means that the content of them has diverged between the two websites with some people making prs to add content to docs on ocfweb and some prs to ocfstatic i ve already started a repo at with the docs content and i think i ve been able to merge everything together so far on the branch kian extract docs i ve done the work of using a git submodule to point the docs folder to that repo which will allow us to update docs in one place and have it take effect everywhere on the site
| 0
|
33,214
| 27,310,538,986
|
IssuesEvent
|
2023-02-24 11:48:18
|
ministryofjustice/staff-infrastructure-monitoring-app-reachability
|
https://api.github.com/repos/ministryofjustice/staff-infrastructure-monitoring-app-reachability
|
closed
|
Collaborator review date expires soon for user emileswarts
|
infrastructure-monitoring
|
Hi there
The user @emileswarts has its access for this repository maintained in code here: https://github.com/ministryofjustice/github-collaborators
The review_after date is due to expire within one month, please update this via a PR if they still require access.
If you have any questions, please post in #ask-operations-engineering on Slack.
Failure to update the review_date will result in the collaborator being removed from the repository via our automation.
|
1.0
|
Collaborator review date expires soon for user emileswarts - Hi there
The user @emileswarts has its access for this repository maintained in code here: https://github.com/ministryofjustice/github-collaborators
The review_after date is due to expire within one month, please update this via a PR if they still require access.
If you have any questions, please post in #ask-operations-engineering on Slack.
Failure to update the review_date will result in the collaborator being removed from the repository via our automation.
|
non_process
|
collaborator review date expires soon for user emileswarts hi there the user emileswarts has its access for this repository maintained in code here the review after date is due to expire within one month please update this via a pr if they still require access if you have any questions please post in ask operations engineering on slack failure to update the review date will result in the collaborator being removed from the repository via our automation
| 0
|
4,568
| 3,393,217,263
|
IssuesEvent
|
2015-11-30 23:03:56
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
closed
|
new-build / new-app should report error if input and output image stream tags are identical
|
component/build priority/P2
|
if we generate a build config and the source tag and the destination tag are identical we should report an error (if the user hasn't specified --to https://github.com/openshift/origin/issues/4980)
@smarterclayton FYI
|
1.0
|
new-build / new-app should report error if input and output image stream tags are identical - if we generate a build config and the source tag and the destination tag are identical we should report an error (if the user hasn't specified --to https://github.com/openshift/origin/issues/4980)
@smarterclayton FYI
|
non_process
|
new build new app should report error if input and output image stream tags are identical if we generate a build config and the source tag and the destination tag are identical we should report an error if the user hasn t specified to smarterclayton fyi
| 0
|
1,416
| 3,980,328,605
|
IssuesEvent
|
2016-05-06 06:51:03
|
e-government-ua/iBP
|
https://api.github.com/repos/e-government-ua/iBP
|
closed
|
Черкаська область - м.Золотоноша - встановлення за погодженням з власником режиму роботи підприємств торгівлі, ресторанного господарства та сфери послуг незалежно від форм власності
|
In process of testing in work test
|
Координатор:
Колодич Олена - Координатор IGov в Черкаській області. тел.+380674704730,
**дуже велике прохання координатора** -
коли будь-які листи будуть надсилатися на контактних осіб (тестування, питання тощо)
ставити ії в копію elena.kolodich@privatbank.ua
та називати листи **IGov - район/місто -(назва послуги)**
Контактна особа:
Остроглазова Вікторія (0678924701),
Мірошник Іна(0969455819);
economic@zolo.ck.ua
|
1.0
|
Черкаська область - м.Золотоноша - встановлення за погодженням з власником режиму роботи підприємств торгівлі, ресторанного господарства та сфери послуг незалежно від форм власності - Координатор:
Колодич Олена - Координатор IGov в Черкаській області. тел.+380674704730,
**дуже велике прохання координатора** -
коли будь-які листи будуть надсилатися на контактних осіб (тестування, питання тощо)
ставити ії в копію elena.kolodich@privatbank.ua
та називати листи **IGov - район/місто -(назва послуги)**
Контактна особа:
Остроглазова Вікторія (0678924701),
Мірошник Іна(0969455819);
economic@zolo.ck.ua
|
process
|
черкаська область м золотоноша встановлення за погодженням з власником режиму роботи підприємств торгівлі ресторанного господарства та сфери послуг незалежно від форм власності координатор колодич олена координатор igov в черкаській області тел дуже велике прохання координатора коли будь які листи будуть надсилатися на контактних осіб тестування питання тощо ставити ії в копію elena kolodich privatbank ua та називати листи igov район місто назва послуги контактна особа остроглазова вікторія мірошник іна economic zolo ck ua
| 1
|
406,800
| 27,583,415,360
|
IssuesEvent
|
2023-03-08 17:47:54
|
Fuenfgeld/DMA2023TeamD
|
https://api.github.com/repos/Fuenfgeld/DMA2023TeamD
|
closed
|
Dokumentation der Quelldaten
|
documentation
|
https://github.com/Fuenfgeld/DMA2023TeamD/wiki/Dokumentation-der-Quelldaten
Datenbank Dokumentation:
- Erstellen sie ein Grafisches Datenbank Schema der Datenbank
- Was sind hier Stammdaten und was sind hier Bewegungsdaten Tabellen.
> Erstellen sie eine Seite Pro Tabelle.
> Diese Seite sollte folgende Informationen enthalten.
>> 1. SQL Create Statement
>> 2. Tabelle mit folgenden Spalten
>> 3. Spalten Bezeichnung,
>> 4. Spalten Datentyp,
>> 5. Spalten Inhaltsbeschreibung,
>> 6. Gibt es für diese Spalte ein Index? Wenn ja was ist der Index Name, was ist der Index Typ
>> 7. `Ist` diese Spalte über ein Foreign Key constraint mit einer anderen Tabelle verbunden? Wenn ja Link zu der anderen Tabellen Dokumentation.
|
1.0
|
Dokumentation der Quelldaten - https://github.com/Fuenfgeld/DMA2023TeamD/wiki/Dokumentation-der-Quelldaten
Datenbank Dokumentation:
- Erstellen sie ein Grafisches Datenbank Schema der Datenbank
- Was sind hier Stammdaten und was sind hier Bewegungsdaten Tabellen.
> Erstellen sie eine Seite Pro Tabelle.
> Diese Seite sollte folgende Informationen enthalten.
>> 1. SQL Create Statement
>> 2. Tabelle mit folgenden Spalten
>> 3. Spalten Bezeichnung,
>> 4. Spalten Datentyp,
>> 5. Spalten Inhaltsbeschreibung,
>> 6. Gibt es für diese Spalte ein Index? Wenn ja was ist der Index Name, was ist der Index Typ
>> 7. `Ist` diese Spalte über ein Foreign Key constraint mit einer anderen Tabelle verbunden? Wenn ja Link zu der anderen Tabellen Dokumentation.
|
non_process
|
dokumentation der quelldaten datenbank dokumentation erstellen sie ein grafisches datenbank schema der datenbank was sind hier stammdaten und was sind hier bewegungsdaten tabellen erstellen sie eine seite pro tabelle diese seite sollte folgende informationen enthalten sql create statement tabelle mit folgenden spalten spalten bezeichnung spalten datentyp spalten inhaltsbeschreibung gibt es für diese spalte ein index wenn ja was ist der index name was ist der index typ ist diese spalte über ein foreign key constraint mit einer anderen tabelle verbunden wenn ja link zu der anderen tabellen dokumentation
| 0
|
219,882
| 16,854,198,286
|
IssuesEvent
|
2021-06-21 02:42:11
|
PlaceOS/docs
|
https://api.github.com/repos/PlaceOS/docs
|
closed
|
Cleanup preexisting branches
|
type: documentation
|
Several (placeholder?) branches started by @stakach require curation, editing and merging. PRs for them have been closed until this review takes place, at which point I will open new PRs as each branch is ready.
|
1.0
|
Cleanup preexisting branches - Several (placeholder?) branches started by @stakach require curation, editing and merging. PRs for them have been closed until this review takes place, at which point I will open new PRs as each branch is ready.
|
non_process
|
cleanup preexisting branches several placeholder branches started by stakach require curation editing and merging prs for them have been closed until this review takes place at which point i will open new prs as each branch is ready
| 0
|
726,760
| 25,010,462,033
|
IssuesEvent
|
2022-11-03 14:54:59
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
qemu_x86: upgrading to q35 breaks networking samples.
|
bug priority: medium area: QEMU area: Ethernet
|
**Describe the bug**
with this commit, samples/net/socket/echo_server fails to run
```
commit a2aa462f7d099bf30c1e054a1f38d34359d83418 (refs/bisect/bad)
Author: Andrei Emeltchenko <andrei.emeltchenko@intel.com>
Date: Thu Oct 27 18:54:13 2022 +0300
qemu_x86: Use Qemu Q35 machine feature (ICH9)
Using very old machine does not make sense anymore. Switch to the new
Q35 (https://wiki.qemu.org/Features/Q35). Among other things it allows
to test SMBus and watchdog.
Signed-off-by: Andrei Emeltchenko <andrei.emeltchenko@intel.com>
diff --git a/boards/x86/qemu_x86/board.cmake b/boards/x86/qemu_x86/board.cmake
index b4f7c03639..934b078154 100644
--- a/boards/x86/qemu_x86/board.cmake
+++ b/boards/x86/qemu_x86/board.cmake
@@ -60,6 +60,7 @@ endif()
set(QEMU_FLAGS_${ARCH}
-m ${QEMU_MEMORY_SIZE_MB}
-cpu ${QEMU_CPU_TYPE_${ARCH}}${QEMU_CPU_FLAGS}
+ -machine q35
-device isa-debug-exit,iobase=0xf4,iosize=0x04
${REBOOT_FLAG}
-nographic
```
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. west build -b qemu_x86 -- -DCONF_FILE=prj.conf -DOVERLAY_CONFIG="overlay-e1000.conf"
3. west build -t run
4. See error
```
-- west build: running target run
[0/1] To exit from QEMU enter: 'CTRL+a, x'[QEMU] CPU: qemu32,+nx,+pae
SeaBIOS (version zephyr-v1.0.0-0-g31d4e0e-dirty-20200714_234759-fv-az50-zephyr)
iPXE (http://ipxe.org) 00:02.0 CA00 PCI2.10 PnP PMM+00392120+002F2120 CA00
Booting from ROM..
FAILED: zephyr/CMakeFiles/run_qemu /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build/zephyr/CMakeFiles/run_qemu
cd /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build && /home/ubuntu/zephyr-sdk/sysroots/x86_64-pokysdk-linux/usr/bin/qemu-system-i386 -m 4 -cpu qemu32,+nx,+pae -machine q35 -device isa-debug-exit,iobase=0xf4,iosize=0x04 -no-reboot -nographic -no-acpi -nic tap,model=e1000,script=no,downscript=no,ifname=zeth -pidfile qemu.pid -chardev stdio,id=con,mux=on -serial chardev:con -mon chardev=con,mode=readline -kernel /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build/zephyr/zephyr.elf
ninja: build stopped: subcommand failed.
FATAL ERROR: command exited with status 1: /usr/local/bin/cmake --build /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build --target run
```
**Expected behavior**
the qemu_x86 can start successfully
**Impact**
showstopper
**Environment (please complete the following information):**
- OS: (e.g. Linux, , )
- Toolchain (e.g Zephyr SDK, )
- Commit SHA or Version used: a2aa462f7d099bf30c1e054a1f38d34359d83418
**Additional context**
this impact the maxpro networking tcpip testing, as we testing tcpip stack on qemu_x86 platforms
|
1.0
|
qemu_x86: upgrading to q35 breaks networking samples. - **Describe the bug**
with this commit, samples/net/socket/echo_server fails to run
```
commit a2aa462f7d099bf30c1e054a1f38d34359d83418 (refs/bisect/bad)
Author: Andrei Emeltchenko <andrei.emeltchenko@intel.com>
Date: Thu Oct 27 18:54:13 2022 +0300
qemu_x86: Use Qemu Q35 machine feature (ICH9)
Using very old machine does not make sense anymore. Switch to the new
Q35 (https://wiki.qemu.org/Features/Q35). Among other things it allows
to test SMBus and watchdog.
Signed-off-by: Andrei Emeltchenko <andrei.emeltchenko@intel.com>
diff --git a/boards/x86/qemu_x86/board.cmake b/boards/x86/qemu_x86/board.cmake
index b4f7c03639..934b078154 100644
--- a/boards/x86/qemu_x86/board.cmake
+++ b/boards/x86/qemu_x86/board.cmake
@@ -60,6 +60,7 @@ endif()
set(QEMU_FLAGS_${ARCH}
-m ${QEMU_MEMORY_SIZE_MB}
-cpu ${QEMU_CPU_TYPE_${ARCH}}${QEMU_CPU_FLAGS}
+ -machine q35
-device isa-debug-exit,iobase=0xf4,iosize=0x04
${REBOOT_FLAG}
-nographic
```
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. west build -b qemu_x86 -- -DCONF_FILE=prj.conf -DOVERLAY_CONFIG="overlay-e1000.conf"
3. west build -t run
4. See error
```
-- west build: running target run
[0/1] To exit from QEMU enter: 'CTRL+a, x'[QEMU] CPU: qemu32,+nx,+pae
SeaBIOS (version zephyr-v1.0.0-0-g31d4e0e-dirty-20200714_234759-fv-az50-zephyr)
iPXE (http://ipxe.org) 00:02.0 CA00 PCI2.10 PnP PMM+00392120+002F2120 CA00
Booting from ROM..
FAILED: zephyr/CMakeFiles/run_qemu /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build/zephyr/CMakeFiles/run_qemu
cd /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build && /home/ubuntu/zephyr-sdk/sysroots/x86_64-pokysdk-linux/usr/bin/qemu-system-i386 -m 4 -cpu qemu32,+nx,+pae -machine q35 -device isa-debug-exit,iobase=0xf4,iosize=0x04 -no-reboot -nographic -no-acpi -nic tap,model=e1000,script=no,downscript=no,ifname=zeth -pidfile qemu.pid -chardev stdio,id=con,mux=on -serial chardev:con -mon chardev=con,mode=readline -kernel /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build/zephyr/zephyr.elf
ninja: build stopped: subcommand failed.
FATAL ERROR: command exited with status 1: /usr/local/bin/cmake --build /home/shared/disk/zephyr_project/zephyr_test/zephyr/samples/net/sockets/echo_server/build --target run
```
**Expected behavior**
the qemu_x86 can start successfully
**Impact**
showstopper
**Environment (please complete the following information):**
- OS: (e.g. Linux, , )
- Toolchain (e.g Zephyr SDK, )
- Commit SHA or Version used: a2aa462f7d099bf30c1e054a1f38d34359d83418
**Additional context**
this impact the maxpro networking tcpip testing, as we testing tcpip stack on qemu_x86 platforms
|
non_process
|
qemu upgrading to breaks networking samples describe the bug with this commit samples net socket echo server fails to run commit refs bisect bad author andrei emeltchenko date thu oct qemu use qemu machine feature using very old machine does not make sense anymore switch to the new among other things it allows to test smbus and watchdog signed off by andrei emeltchenko diff git a boards qemu board cmake b boards qemu board cmake index a boards qemu board cmake b boards qemu board cmake endif set qemu flags arch m qemu memory size mb cpu qemu cpu type arch qemu cpu flags machine device isa debug exit iobase iosize reboot flag nographic to reproduce steps to reproduce the behavior mkdir build cd build west build b qemu dconf file prj conf doverlay config overlay conf west build t run see error west build running target run to exit from qemu enter ctrl a x cpu nx pae seabios version zephyr dirty fv zephyr ipxe pnp pmm booting from rom failed zephyr cmakefiles run qemu home shared disk zephyr project zephyr test zephyr samples net sockets echo server build zephyr cmakefiles run qemu cd home shared disk zephyr project zephyr test zephyr samples net sockets echo server build home ubuntu zephyr sdk sysroots pokysdk linux usr bin qemu system m cpu nx pae machine device isa debug exit iobase iosize no reboot nographic no acpi nic tap model script no downscript no ifname zeth pidfile qemu pid chardev stdio id con mux on serial chardev con mon chardev con mode readline kernel home shared disk zephyr project zephyr test zephyr samples net sockets echo server build zephyr zephyr elf ninja build stopped subcommand failed fatal error command exited with status usr local bin cmake build home shared disk zephyr project zephyr test zephyr samples net sockets echo server build target run expected behavior the qemu can start successfully impact showstopper environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used additional context this impact the maxpro networking tcpip testing as we testing tcpip stack on qemu platforms
| 0
|
147,869
| 13,217,938,872
|
IssuesEvent
|
2020-08-17 07:49:07
|
jesse1412/BOT1412
|
https://api.github.com/repos/jesse1412/BOT1412
|
closed
|
Add an installation guide for development purposes
|
documentation
|
A guide on how to install this project on a local machine would add a bit of polish to the project as a whole, as well reducing the barrier to entry for those newer to python, the detail of the guide would be entirely at your disgression of course.
|
1.0
|
Add an installation guide for development purposes - A guide on how to install this project on a local machine would add a bit of polish to the project as a whole, as well reducing the barrier to entry for those newer to python, the detail of the guide would be entirely at your disgression of course.
|
non_process
|
add an installation guide for development purposes a guide on how to install this project on a local machine would add a bit of polish to the project as a whole as well reducing the barrier to entry for those newer to python the detail of the guide would be entirely at your disgression of course
| 0
|
2,938
| 5,921,045,793
|
IssuesEvent
|
2017-05-22 21:51:16
|
ncbo/bioportal-project
|
https://api.github.com/repos/ncbo/bioportal-project
|
closed
|
HP: BioPortal UI fails to display the class tree
|
ontology processing problem
|
Latest version of the [HP ontology](http://bioportal.bioontology.org/ontologies/HP) (upload date: 04/13/2017) parsed successfully, but the BioPortal UI shows a 404 error if you try to view the class tree. The underlying reason for the 404 is that the [REST call to retrieve the root classes](http://data.bioontology.org/ontologies/HP/classes/roots) returns nothing.
|
1.0
|
HP: BioPortal UI fails to display the class tree - Latest version of the [HP ontology](http://bioportal.bioontology.org/ontologies/HP) (upload date: 04/13/2017) parsed successfully, but the BioPortal UI shows a 404 error if you try to view the class tree. The underlying reason for the 404 is that the [REST call to retrieve the root classes](http://data.bioontology.org/ontologies/HP/classes/roots) returns nothing.
|
process
|
hp bioportal ui fails to display the class tree latest version of the upload date parsed successfully but the bioportal ui shows a error if you try to view the class tree the underlying reason for the is that the returns nothing
| 1
|
13,848
| 16,611,393,495
|
IssuesEvent
|
2021-06-02 11:59:22
|
deepset-ai/haystack
|
https://api.github.com/repos/deepset-ai/haystack
|
closed
|
Document splitting based on word count can cause whitespace normalization
|
topic:preprocessing type:bug
|
When the Preprocessor is used to split Documents by word count, sequences of multiple whitespaces might be normalized as a side effect. This is problematic in evaluation, especially for open domain retrieval (see #983). With #1022, a warning is logged, but the problem is not truly fixed.
|
1.0
|
Document splitting based on word count can cause whitespace normalization - When the Preprocessor is used to split Documents by word count, sequences of multiple whitespaces might be normalized as a side effect. This is problematic in evaluation, especially for open domain retrieval (see #983). With #1022, a warning is logged, but the problem is not truly fixed.
|
process
|
document splitting based on word count can cause whitespace normalization when the preprocessor is used to split documents by word count sequences of multiple whitespaces might be normalized as a side effect this is problematic in evaluation especially for open domain retrieval see with a warning is logged but the problem is not truly fixed
| 1
|
15,964
| 20,177,198,880
|
IssuesEvent
|
2022-02-10 15:24:33
|
ossf/tac
|
https://api.github.com/repos/ossf/tac
|
closed
|
TAC Election: Process & Tools
|
ElectionProcess
|
- Utilize OpaVote with "meek stv" ranked-choice voting
- Election Officials will announce the timeline for nominations and election via the TAC mailing list and Slack channel
- At the conclusion of the nomination period, the Election Officials will publish the complete list of candidates. Publishing options include, email, doc, and/or TAC repo.
- Election Officials will validate the eligibility of voters before the election.
|
1.0
|
TAC Election: Process & Tools - - Utilize OpaVote with "meek stv" ranked-choice voting
- Election Officials will announce the timeline for nominations and election via the TAC mailing list and Slack channel
- At the conclusion of the nomination period, the Election Officials will publish the complete list of candidates. Publishing options include, email, doc, and/or TAC repo.
- Election Officials will validate the eligibility of voters before the election.
|
process
|
tac election process tools utilize opavote with meek stv ranked choice voting election officials will announce the timeline for nominations and election via the tac mailing list and slack channel at the conclusion of the nomination period the election officials will publish the complete list of candidates publishing options include email doc and or tac repo election officials will validate the eligibility of voters before the election
| 1
|
15,419
| 19,605,905,178
|
IssuesEvent
|
2022-01-06 09:26:32
|
plazi/community
|
https://api.github.com/repos/plazi/community
|
opened
|
to be processed https://doi.org/10.1111/mec.15928
|
process request
|
can you process this article and then let's discuss how we create this treatment
https://doi.org/10.1111/mec.15928
[mec.15928.pdf](https://github.com/plazi/community/files/7820785/mec.15928.pdf)

this is part of the CAS press release
|
1.0
|
to be processed https://doi.org/10.1111/mec.15928 - can you process this article and then let's discuss how we create this treatment
https://doi.org/10.1111/mec.15928
[mec.15928.pdf](https://github.com/plazi/community/files/7820785/mec.15928.pdf)

this is part of the CAS press release
|
process
|
to be processed can you process this article and then let s discuss how we create this treatment this is part of the cas press release
| 1
|
1,664
| 4,294,631,825
|
IssuesEvent
|
2016-07-19 01:26:28
|
mitchellh/packer
|
https://api.github.com/repos/mitchellh/packer
|
closed
|
doubled trace in vmware post-processor
|
easy post-processor/vsphere
|
file `post-processor/vsphere/post-processor.go`, function PostProcess
there are two same traces in lines 136 and 146:
```go
ui.Message(fmt.Sprintf("Uploading %s to vSphere", source))
```
|
1.0
|
doubled trace in vmware post-processor - file `post-processor/vsphere/post-processor.go`, function PostProcess
there are two same traces in lines 136 and 146:
```go
ui.Message(fmt.Sprintf("Uploading %s to vSphere", source))
```
|
process
|
doubled trace in vmware post processor file post processor vsphere post processor go function postprocess there are two same traces in lines and go ui message fmt sprintf uploading s to vsphere source
| 1
|
15,351
| 19,521,832,273
|
IssuesEvent
|
2021-12-29 20:09:24
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
Installing cri-o from deb packages causing errors
|
kind/bug priority/important-soon co/runtime/crio kind/process
|
When running a kic image build we get the following error:
`Failed to fetch https://provo-mirror.opensuse.org/repositories/devel:/kubic:/libcontainers:/stable:/cri-o:/1.22/xUbuntu_20.04/amd64/cri-o_1.22.0~0_amd64.deb File has unexpected size (20107660 != 20107360). Mirror sync in progress?`
We should install a new way that doesn't use deb packages so we can continue to build our kic images.
Related https://github.com/cri-o/cri-o/issues/5343
|
1.0
|
Installing cri-o from deb packages causing errors - When running a kic image build we get the following error:
`Failed to fetch https://provo-mirror.opensuse.org/repositories/devel:/kubic:/libcontainers:/stable:/cri-o:/1.22/xUbuntu_20.04/amd64/cri-o_1.22.0~0_amd64.deb File has unexpected size (20107660 != 20107360). Mirror sync in progress?`
We should install a new way that doesn't use deb packages so we can continue to build our kic images.
Related https://github.com/cri-o/cri-o/issues/5343
|
process
|
installing cri o from deb packages causing errors when running a kic image build we get the following error failed to fetch file has unexpected size mirror sync in progress we should install a new way that doesn t use deb packages so we can continue to build our kic images related
| 1
|
54,690
| 30,313,773,142
|
IssuesEvent
|
2023-07-10 14:21:42
|
hydroshare/hydroshare
|
https://api.github.com/repos/hydroshare/hydroshare
|
closed
|
File list endpoint causes timeout error for resource with large number of files
|
Performance
|
Using the file list endpoint (**/hsapi/resource/{id}/files/**) to list files for a resource with large number of files (> ~200 files) throws timeout error.
**Step to produce this error:**
1. Go to www.hydroshare.org/hsapi to see the Swagger UI.
2. Use the GET endpoint /resource/{id}/files/ to list files for a resource
3. For resource ID, enter **038179b73e124d25b270538432e0370d**
4. Click the Execute button
5. You should see **504 timeout error**
6. In the count field, enter 10 to list only 10 files of the resource
7. Click the Execute button
8. You should see **504 timeout error**
**Expected behavior.:**
1. When using this endpoint there should not be timeout error even if the resource has large number of files
2. The response time should be dependent on the number of files returned. Currently, the response time depends on the total number of files the resource has.
|
True
|
File list endpoint causes timeout error for resource with large number of files - Using the file list endpoint (**/hsapi/resource/{id}/files/**) to list files for a resource with large number of files (> ~200 files) throws timeout error.
**Step to produce this error:**
1. Go to www.hydroshare.org/hsapi to see the Swagger UI.
2. Use the GET endpoint /resource/{id}/files/ to list files for a resource
3. For resource ID, enter **038179b73e124d25b270538432e0370d**
4. Click the Execute button
5. You should see **504 timeout error**
6. In the count field, enter 10 to list only 10 files of the resource
7. Click the Execute button
8. You should see **504 timeout error**
**Expected behavior.:**
1. When using this endpoint there should not be timeout error even if the resource has large number of files
2. The response time should be dependent on the number of files returned. Currently, the response time depends on the total number of files the resource has.
|
non_process
|
file list endpoint causes timeout error for resource with large number of files using the file list endpoint hsapi resource id files to list files for a resource with large number of files files throws timeout error step to produce this error go to to see the swagger ui use the get endpoint resource id files to list files for a resource for resource id enter click the execute button you should see timeout error in the count field enter to list only files of the resource click the execute button you should see timeout error expected behavior when using this endpoint there should not be timeout error even if the resource has large number of files the response time should be dependent on the number of files returned currently the response time depends on the total number of files the resource has
| 0
|
49,543
| 6,032,349,483
|
IssuesEvent
|
2017-06-09 03:26:35
|
yahoo/fili
|
https://api.github.com/repos/yahoo/fili
|
opened
|
Ability to override @Timeout annotations when debugging tests
|
EXTENSIBILITY IDEA TESTING
|
Especially for tests that extend existing tests, it's would be great. The tests timing out and shutting down makes debugging extremely painful.
The usecase we ran into is someone extending BaseDataServletSpec, which has a Timeout, and needing to debug with their extension spec.
It would be sufficient to be able to turn off the timeouts.
|
1.0
|
Ability to override @Timeout annotations when debugging tests - Especially for tests that extend existing tests, it's would be great. The tests timing out and shutting down makes debugging extremely painful.
The usecase we ran into is someone extending BaseDataServletSpec, which has a Timeout, and needing to debug with their extension spec.
It would be sufficient to be able to turn off the timeouts.
|
non_process
|
ability to override timeout annotations when debugging tests especially for tests that extend existing tests it s would be great the tests timing out and shutting down makes debugging extremely painful the usecase we ran into is someone extending basedataservletspec which has a timeout and needing to debug with their extension spec it would be sufficient to be able to turn off the timeouts
| 0
|
3,744
| 6,733,148,493
|
IssuesEvent
|
2017-10-18 13:59:41
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Tree planting detail form - request process
|
process workflow
|
Process request and update the status of tree planting detail form.
|
1.0
|
Tree planting detail form - request process - Process request and update the status of tree planting detail form.
|
process
|
tree planting detail form request process process request and update the status of tree planting detail form
| 1
|
334,174
| 10,136,893,281
|
IssuesEvent
|
2019-08-02 14:04:11
|
salesagility/SuiteCRM
|
https://api.github.com/repos/salesagility/SuiteCRM
|
closed
|
Issue Suite P Invoices Screen
|
Medium Priority bug
|
Hi, when you goes to create any document (quote, invoice), details sections not works fine.
i've added screencapture.

#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
- SuiteCRM Version used: 7.7.5
- Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Any
- Environment name and version (e.g. MySQL, PHP 7): MSSQL , PHP7
- Operating System and version (e.g Ubuntu 16.04): WINDOWS SERVER IIS
|
1.0
|
Issue Suite P Invoices Screen - Hi, when you goes to create any document (quote, invoice), details sections not works fine.
i've added screencapture.

#### Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
- SuiteCRM Version used: 7.7.5
- Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Any
- Environment name and version (e.g. MySQL, PHP 7): MSSQL , PHP7
- Operating System and version (e.g Ubuntu 16.04): WINDOWS SERVER IIS
|
non_process
|
issue suite p invoices screen hi when you goes to create any document quote invoice details sections not works fine i ve added screencapture your environment suitecrm version used browser name and version e g chrome version bit any environment name and version e g mysql php mssql operating system and version e g ubuntu windows server iis
| 0
|
123,971
| 4,889,528,689
|
IssuesEvent
|
2016-11-18 10:31:49
|
tardis-sn/tardis
|
https://api.github.com/repos/tardis-sn/tardis
|
closed
|
Path problem in case of debian + virtualenv + astropy (debian package)
|
priority - low
|
Running into a rather pathological problem when using tardis with the following combination:
- debian Jessie (may apply also to other linux distributions and versions)
- a python virtualenv with --system-site-packages (i.e. rely also on system-wide python packages)
- having only the debian astropy package installed
After setting up the vanilla virtualenv, the python path (as in `sys.path`) looks as expected:
```
/home/ulrich/python-virtualenv/test/lib/python2.7
/home/ulrich/python-virtualenv/test/lib/python2.7/plat-x86_64-linux-gnu
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-old
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-dynload
/usr/lib/python2.7
/usr/lib/python2.7/plat-x86_64-linux-gnu
/usr/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages
/home/ulrich/python-virtualenv/test/lib/python2.7/site-packages
/usr/local/lib/python2.7/site-packages
/usr/local/lib/python2.7/dist-packages
/usr/lib/python2.7/dist-packages
/usr/lib/python2.7/dist-packages/PILcompat
/usr/lib/python2.7/dist-packages/gst-0.10
/usr/lib/python2.7/dist-packages/gtk-2.0
/usr/lib/pymodules/python2.7
/usr/lib/python2.7/dist-packages/wx-3.0-gtk2
```
The virtualenv library paths have precedence over the system-wide paths.
If tardis is now installed inside of this virtualenv, using the setup.py script, the python path gets messed up:
```
/home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages/tardis_sn-1.0.1-py2.7-linux-x86_64.egg
/usr/lib/python2.7/dist-packages
/home/ulrich/python-virtualenv/test/lib/python2.7/site-packages/tardis_sn-1.0.1-py2.7-linux-x86_64.egg
/home/ulrich/Work/python-modules
/home/ulrich/python-virtualenv/test/lib/python2.7
/home/ulrich/python-virtualenv/test/lib/python2.7/plat-x86_64-linux-gnu
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-old
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-dynload
/usr/lib/python2.7
/usr/lib/python2.7/plat-x86_64-linux-gnu
/usr/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages
/home/ulrich/python-virtualenv/test/lib/python2.7/site-packages
/usr/local/lib/python2.7/site-packages
/usr/local/lib/python2.7/dist-packages
/usr/lib/python2.7/dist-packages/PILcompat
/usr/lib/python2.7/dist-packages/gst-0.10
/usr/lib/python2.7/dist-packages/gtk-2.0
/usr/lib/pymodules/python2.7
/usr/lib/python2.7/dist-packages/wx-3.0-gtk2
```
Now, the system-wide path /usr/bin/python2.7/dist-packages takes precedence over the virtualenv paths. This is due to the file easy_install.pth in the tardis install directory, /home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages/, which not only adds the tardis egg to the path (which it is supposed to do), but also the global dist-packages path.
This has been confirmed by @tukss, who also identified the cause:
It seems to be connected with the use of the python-setuptools to install/check for astropy. If astropy is not installed in the virtualenv but system-wide through the official debian repositories, the setuptools add the global path to the .pth file and thus to the python path.
|
1.0
|
Path problem in case of debian + virtualenv + astropy (debian package) - Running into a rather pathological problem when using tardis with the following combination:
- debian Jessie (may apply also to other linux distributions and versions)
- a python virtualenv with --system-site-packages (i.e. rely also on system-wide python packages)
- having only the debian astropy package installed
After setting up the vanilla virtualenv, the python path (as in `sys.path`) looks as expected:
```
/home/ulrich/python-virtualenv/test/lib/python2.7
/home/ulrich/python-virtualenv/test/lib/python2.7/plat-x86_64-linux-gnu
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-old
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-dynload
/usr/lib/python2.7
/usr/lib/python2.7/plat-x86_64-linux-gnu
/usr/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages
/home/ulrich/python-virtualenv/test/lib/python2.7/site-packages
/usr/local/lib/python2.7/site-packages
/usr/local/lib/python2.7/dist-packages
/usr/lib/python2.7/dist-packages
/usr/lib/python2.7/dist-packages/PILcompat
/usr/lib/python2.7/dist-packages/gst-0.10
/usr/lib/python2.7/dist-packages/gtk-2.0
/usr/lib/pymodules/python2.7
/usr/lib/python2.7/dist-packages/wx-3.0-gtk2
```
The virtualenv library paths have precedence over the system-wide paths.
If tardis is now installed inside of this virtualenv, using the setup.py script, the python path gets messed up:
```
/home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages/tardis_sn-1.0.1-py2.7-linux-x86_64.egg
/usr/lib/python2.7/dist-packages
/home/ulrich/python-virtualenv/test/lib/python2.7/site-packages/tardis_sn-1.0.1-py2.7-linux-x86_64.egg
/home/ulrich/Work/python-modules
/home/ulrich/python-virtualenv/test/lib/python2.7
/home/ulrich/python-virtualenv/test/lib/python2.7/plat-x86_64-linux-gnu
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-old
/home/ulrich/python-virtualenv/test/lib/python2.7/lib-dynload
/usr/lib/python2.7
/usr/lib/python2.7/plat-x86_64-linux-gnu
/usr/lib/python2.7/lib-tk
/home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages
/home/ulrich/python-virtualenv/test/lib/python2.7/site-packages
/usr/local/lib/python2.7/site-packages
/usr/local/lib/python2.7/dist-packages
/usr/lib/python2.7/dist-packages/PILcompat
/usr/lib/python2.7/dist-packages/gst-0.10
/usr/lib/python2.7/dist-packages/gtk-2.0
/usr/lib/pymodules/python2.7
/usr/lib/python2.7/dist-packages/wx-3.0-gtk2
```
Now, the system-wide path /usr/bin/python2.7/dist-packages takes precedence over the virtualenv paths. This is due to the file easy_install.pth in the tardis install directory, /home/ulrich/python-virtualenv/test/local/lib/python2.7/site-packages/, which not only adds the tardis egg to the path (which it is supposed to do), but also the global dist-packages path.
This has been confirmed by @tukss, who also identified the cause:
It seems to be connected with the use of the python-setuptools to install/check for astropy. If astropy is not installed in the virtualenv but system-wide through the official debian repositories, the setuptools add the global path to the .pth file and thus to the python path.
|
non_process
|
path problem in case of debian virtualenv astropy debian package running into a rather pathological problem when using tardis with the following combination debian jessie may apply also to other linux distributions and versions a python virtualenv with system site packages i e rely also on system wide python packages having only the debian astropy package installed after setting up the vanilla virtualenv the python path as in sys path looks as expected home ulrich python virtualenv test lib home ulrich python virtualenv test lib plat linux gnu home ulrich python virtualenv test lib lib tk home ulrich python virtualenv test lib lib old home ulrich python virtualenv test lib lib dynload usr lib usr lib plat linux gnu usr lib lib tk home ulrich python virtualenv test local lib site packages home ulrich python virtualenv test lib site packages usr local lib site packages usr local lib dist packages usr lib dist packages usr lib dist packages pilcompat usr lib dist packages gst usr lib dist packages gtk usr lib pymodules usr lib dist packages wx the virtualenv library paths have precedence over the system wide paths if tardis is now installed inside of this virtualenv using the setup py script the python path gets messed up home ulrich python virtualenv test local lib site packages tardis sn linux egg usr lib dist packages home ulrich python virtualenv test lib site packages tardis sn linux egg home ulrich work python modules home ulrich python virtualenv test lib home ulrich python virtualenv test lib plat linux gnu home ulrich python virtualenv test lib lib tk home ulrich python virtualenv test lib lib old home ulrich python virtualenv test lib lib dynload usr lib usr lib plat linux gnu usr lib lib tk home ulrich python virtualenv test local lib site packages home ulrich python virtualenv test lib site packages usr local lib site packages usr local lib dist packages usr lib dist packages pilcompat usr lib dist packages gst usr lib dist packages gtk usr lib pymodules usr lib dist packages wx now the system wide path usr bin dist packages takes precedence over the virtualenv paths this is due to the file easy install pth in the tardis install directory home ulrich python virtualenv test local lib site packages which not only adds the tardis egg to the path which it is supposed to do but also the global dist packages path this has been confirmed by tukss who also identified the cause it seems to be connected with the use of the python setuptools to install check for astropy if astropy is not installed in the virtualenv but system wide through the official debian repositories the setuptools add the global path to the pth file and thus to the python path
| 0
|
607,665
| 18,788,284,223
|
IssuesEvent
|
2021-11-08 14:23:52
|
o3de/o3de
|
https://api.github.com/repos/o3de/o3de
|
opened
|
Certain DebugDrawRequestBus Script Canvas nodes are not functioning
|
kind/bug needs-triage sig/core priority/major
|
**Describe the bug**
Some DebugDrawRequestBus Script Canvas nodes are not functioning properly while others work fine.
Examples of functioning nodes:
- DrawTextOnScreen
- DrawTextOnEntity
Examples of nonfunctioning nodes:
- DrawLineEntityToEntity
- DrawRayEntityToDirection
- DrawSphereOnEntity
- DrawAabb
This issue does not occur on the Development (3b1c633) build.
This issue can be related to https://github.com/o3de/o3de/issues/5404 as the functioning nodes seem to be drawing text, and the nonfunctional ones draw lines or shapes.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Create an entity.
2. Add a child entity to the entity from step 1.
3. Move the child entity from the parent.
4. Add a Script Canvas node to the child entity.
5. Create and attach a .scriptcanvas asset as shown in the **Video and Script Canvas graph** section.
6. Enter the Game Mode.
**Expected behavior**
All DebugDrawRequestBus Script Canvas nodes are functional.
**Actual behavior**
Certain DebugDrawRequestBus Script Canvas nodes are nonfunctional.
**Video and Script Canvas graph**
https://user-images.githubusercontent.com/86952082/140755826-8a07f8ed-3313-4b73-b4ed-561b2d227d84.mp4
Script Canvas graph:

**Found in Branch**
Stabilization_2110 (621194e)
**Desktop**
Device: PC
OS: Windows
Version 10
CPU AMD Ryzen 5 3600 6-Core Processor, 3600 MHz, 6 Core(s), 12 Logical Processor(s)
GPU NVIDIA GeForce RTX 2060 SUPER
Memory 16GB
|
1.0
|
Certain DebugDrawRequestBus Script Canvas nodes are not functioning - **Describe the bug**
Some DebugDrawRequestBus Script Canvas nodes are not functioning properly while others work fine.
Examples of functioning nodes:
- DrawTextOnScreen
- DrawTextOnEntity
Examples of nonfunctioning nodes:
- DrawLineEntityToEntity
- DrawRayEntityToDirection
- DrawSphereOnEntity
- DrawAabb
This issue does not occur on the Development (3b1c633) build.
This issue can be related to https://github.com/o3de/o3de/issues/5404 as the functioning nodes seem to be drawing text, and the nonfunctional ones draw lines or shapes.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Create an entity.
2. Add a child entity to the entity from step 1.
3. Move the child entity from the parent.
4. Add a Script Canvas node to the child entity.
5. Create and attach a .scriptcanvas asset as shown in the **Video and Script Canvas graph** section.
6. Enter the Game Mode.
**Expected behavior**
All DebugDrawRequestBus Script Canvas nodes are functional.
**Actual behavior**
Certain DebugDrawRequestBus Script Canvas nodes are nonfunctional.
**Video and Script Canvas graph**
https://user-images.githubusercontent.com/86952082/140755826-8a07f8ed-3313-4b73-b4ed-561b2d227d84.mp4
Script Canvas graph:

**Found in Branch**
Stabilization_2110 (621194e)
**Desktop**
Device: PC
OS: Windows
Version 10
CPU AMD Ryzen 5 3600 6-Core Processor, 3600 MHz, 6 Core(s), 12 Logical Processor(s)
GPU NVIDIA GeForce RTX 2060 SUPER
Memory 16GB
|
non_process
|
certain debugdrawrequestbus script canvas nodes are not functioning describe the bug some debugdrawrequestbus script canvas nodes are not functioning properly while others work fine examples of functioning nodes drawtextonscreen drawtextonentity examples of nonfunctioning nodes drawlineentitytoentity drawrayentitytodirection drawsphereonentity drawaabb this issue does not occur on the development build this issue can be related to as the functioning nodes seem to be drawing text and the nonfunctional ones draw lines or shapes steps to reproduce steps to reproduce the behavior create an entity add a child entity to the entity from step move the child entity from the parent add a script canvas node to the child entity create and attach a scriptcanvas asset as shown in the video and script canvas graph section enter the game mode expected behavior all debugdrawrequestbus script canvas nodes are functional actual behavior certain debugdrawrequestbus script canvas nodes are nonfunctional video and script canvas graph script canvas graph found in branch stabilization desktop device pc os windows version cpu amd ryzen core processor mhz core s logical processor s gpu nvidia geforce rtx super memory
| 0
|
10,934
| 13,748,678,547
|
IssuesEvent
|
2020-10-06 09:23:46
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
closed
|
Log Onboarding May Timeout with Many Log Types
|
bug p0 team:data processing
|
### Describe the bug
Since we changed to adding glue tables if they do not exist to the source api, the run times are longer. This is not an issue if the user has only a few log types if there are large numbers > 10 then the api times out since appsync has a 30sec timeout.
Since having large numbers of sources per bucket is unusual and there is a work around (see below) I don't think this is a pressing issue.
### Work Around
If this is a customer issue, they can onboard the same bucket multiple times with different names for the source and only a few log types per source.
### Possible Solutions
Either make source api optionally async (lots of work), users need to get any errors back so we need a true async api . Possibly move to a socket based api (according to Aggelos) could work with a long lambda timeout. This operation should never take more than a 1 minute. The one case where we could get large variation is that the panther views need to be updated and that is done via Athena which can have long lags before starting. That said, re-trying after any error would fix that ephemeral issue.
|
1.0
|
Log Onboarding May Timeout with Many Log Types - ### Describe the bug
Since we changed to adding glue tables if they do not exist to the source api, the run times are longer. This is not an issue if the user has only a few log types if there are large numbers > 10 then the api times out since appsync has a 30sec timeout.
Since having large numbers of sources per bucket is unusual and there is a work around (see below) I don't think this is a pressing issue.
### Work Around
If this is a customer issue, they can onboard the same bucket multiple times with different names for the source and only a few log types per source.
### Possible Solutions
Either make source api optionally async (lots of work), users need to get any errors back so we need a true async api . Possibly move to a socket based api (according to Aggelos) could work with a long lambda timeout. This operation should never take more than a 1 minute. The one case where we could get large variation is that the panther views need to be updated and that is done via Athena which can have long lags before starting. That said, re-trying after any error would fix that ephemeral issue.
|
process
|
log onboarding may timeout with many log types describe the bug since we changed to adding glue tables if they do not exist to the source api the run times are longer this is not an issue if the user has only a few log types if there are large numbers then the api times out since appsync has a timeout since having large numbers of sources per bucket is unusual and there is a work around see below i don t think this is a pressing issue work around if this is a customer issue they can onboard the same bucket multiple times with different names for the source and only a few log types per source possible solutions either make source api optionally async lots of work users need to get any errors back so we need a true async api possibly move to a socket based api according to aggelos could work with a long lambda timeout this operation should never take more than a minute the one case where we could get large variation is that the panther views need to be updated and that is done via athena which can have long lags before starting that said re trying after any error would fix that ephemeral issue
| 1
|
13,754
| 16,504,306,784
|
IssuesEvent
|
2021-05-25 17:20:26
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
Helm acceptance tests
|
P2 enhancement process
|
**Problem**
During Helm deployments we have no guarantee that the mirror node is healthy end to end.
**Solution**
Run the acceptance tests as a Helm test post deployment
- Add a `charts/hedera-mirror/templates/tests/pod.yaml` that runs the `hedera-mirror-test` image
- Add a `charts/hedera-mirror/templates/tests/secret.yaml` that stores the `application.yaml` config for the test pod
- Add test properties to `values.yaml` that allow to enable/disable, set image, config, etc. Disabled by default.
**Alternatives**
**Additional Context**
|
1.0
|
Helm acceptance tests - **Problem**
During Helm deployments we have no guarantee that the mirror node is healthy end to end.
**Solution**
Run the acceptance tests as a Helm test post deployment
- Add a `charts/hedera-mirror/templates/tests/pod.yaml` that runs the `hedera-mirror-test` image
- Add a `charts/hedera-mirror/templates/tests/secret.yaml` that stores the `application.yaml` config for the test pod
- Add test properties to `values.yaml` that allow to enable/disable, set image, config, etc. Disabled by default.
**Alternatives**
**Additional Context**
|
process
|
helm acceptance tests problem during helm deployments we have no guarantee that the mirror node is healthy end to end solution run the acceptance tests as a helm test post deployment add a charts hedera mirror templates tests pod yaml that runs the hedera mirror test image add a charts hedera mirror templates tests secret yaml that stores the application yaml config for the test pod add test properties to values yaml that allow to enable disable set image config etc disabled by default alternatives additional context
| 1
|
5,823
| 8,657,738,620
|
IssuesEvent
|
2018-11-27 22:13:54
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
[Firestore] Add ARRAY_UNION, ARRAY_REMOVE ARRAY_CONTAINS FieldValues
|
api: firestore triaged for GA type: process
|
There is no support for FieldValue.arrayUnion()/FieldValue.arrayRemove() and no support for array-contains Queries.
|
1.0
|
[Firestore] Add ARRAY_UNION, ARRAY_REMOVE ARRAY_CONTAINS FieldValues - There is no support for FieldValue.arrayUnion()/FieldValue.arrayRemove() and no support for array-contains Queries.
|
process
|
add array union array remove array contains fieldvalues there is no support for fieldvalue arrayunion fieldvalue arrayremove and no support for array contains queries
| 1
|
5,881
| 5,198,309,618
|
IssuesEvent
|
2017-01-23 17:45:38
|
coreos/etcd
|
https://api.github.com/repos/coreos/etcd
|
closed
|
STM: prefetching
|
area/performance kind/enhancement
|
Have `WithPrefetch` for prefetching keys for STM to avoid latency from making multiple `Get` calls when the keys are known a priori.
|
True
|
STM: prefetching - Have `WithPrefetch` for prefetching keys for STM to avoid latency from making multiple `Get` calls when the keys are known a priori.
|
non_process
|
stm prefetching have withprefetch for prefetching keys for stm to avoid latency from making multiple get calls when the keys are known a priori
| 0
|
537,243
| 15,725,915,239
|
IssuesEvent
|
2021-03-29 10:35:25
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
tikv keep panicking at components/engine_rocks/src/engine_iterator.rs:39
|
priority/release-blocker severity/critical type/bug
|
## Bug Report
### What version of TiKV are you using?
Master.
### What operating system and CPU are you using?
Doesn't matter.
### Steps to reproduce
Not sure.
### What did you expect?
TiKV keeps running without panicking.
### What did happened?
TiKV panics:
```
[2021/03/29 12:13:21.004 +08:00] [FATAL] [lib.rs:455] ["assertion failed: cfg!(feature = \\\"nortcheck\\\") || self.valid()?"] [backtrace="stack backtrace:\n 0: t[2/9902]::set_panic_hook::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_util/src/lib.rs:454\n 1: std::panicking::rust_panic_with_hook\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/panicking.rs:595\n 2: std::panicking::begin_panic_handler::{{closure}}\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/panicking.rs:495\n 3: std::sys_common::backtrace::__rust_end_short_backtrace\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/sys_common/backtrace.rs:141\n 4: rust_begin_unwind\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/panicking.rs:493\n 5: core::panicking::panic_fmt\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/core/src/panicking.rs:92\n 6: core::panicking::panic\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/core/src/panicking.rs:50\n 7: <engine_rocks::engine_iterator::RocksEngineIterator as engine_traits::iterable::Iterator>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/engine_rocks/src/engine_iterator.rs:39\n 8: raftstore::store::region_snapshot::RegionIterator<S>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/raftstore/src/store/region_snapshot.rs:339\n tikv_kv::raftstore_impls::<impl tikv_kv::Iterator for raftstore::store::region_snapshot::RegionIterator<S>>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/raftstore_impls.rs:86\n tikv_kv::cursor::Cursor<I>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/cursor.rs:367\n tikv::storage::txn::actions::get_old_value::get_old_value\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/actions/get_old_value.rs:34\n tikv::storage::txn::actions::prewrite::prewrite\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/actions/prewrite.rs:82\n 9: tikv::storage::txn::commands::prewrite::Prewriter<K>::prewrite\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/prewrite.rs:445\n tikv::storage::txn::commands::prewrite::Prewriter<K>::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/prewrite.rs:370\n <tikv::storage::txn::commands::prewrite::Prewrite as tikv::storage::txn::commands::WriteCommand<S,L>>::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/prewrite.rs:198\n 10: tikv::storage::txn::commands::Command::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/mod.rs:539\n 11: tikv::storage::txn::scheduler::Scheduler<E,L>::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/scheduler.rs:635\n tikv::storage::txn::scheduler::Scheduler<E,L>::process_by_worker::{{closure}}::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/scheduler.rs:579\n tikv_kv::with_tls_engine::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/lib.rs:401\n std::thread::local::LocalKey<T>::try_with\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/local.rs:272\n std::thread::local::LocalKey<T>::with\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/local.rs:248\n tikv_kv::with_tls_engine\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/lib.rs:399\n tikv::storage::txn::scheduler::Scheduler<E,L>::process_by_worker::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/scheduler.rs:578\n <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/core/src/future/mod.rs:80\n tikv_util::yatp_pool::future_pool::FuturePool::spawn::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_util/src/yatp_pool/future_pool.rs:113\n <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/core/src/future/mod.rs:80\n 12: <yatp::task::future::Runner as yatp::pool::runner::Runner>::handle\n at /rust/git/checkouts/yatp-e704b73c3ee279b6/6bbea16/src/task/future.rs:261\n 13: <tikv_util::yatp_pool::YatpPoolRunner<T> as yatp::pool::runner::Runner>::handle\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_util/src/yatp_pool/mod.rs:93\n yatp::pool::worker::WorkerThread<T,R>::run\n at /rust/git/checkouts/yatp-e704b73c3ee279b6/6bbea16/src/pool/worker.rs:48\n yatp::pool::builder::LazyBuilder<T>::build::{{closure}}\n at /rust/git/checkouts/yatp-e704b73c3ee279b6/6bbea16/src/pool/builder.rs:91\n std::sys_common::backtrace::__rust_begin_short_backtrace\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/sys_common/backtrace.rs:125\n 14: std::thread::Builder::spawn_unchecked::{{closure}}::{{closure}}\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/mod.rs:474\n <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panic.rs:322\n std::panicking::try::do_call\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panicking.rs:379\n std::panicking::try\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panicking.rs:343\n std::panic::catch_unwind\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panic.rs:396\n std::thread::Builder::spawn_unchecked::{{closure}}\n at
/rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/mod.rs:473\n core::ops::function::FnOnce::call_once{{vtable.shim[2021/03/29 12:13:36.623 +08:00
```
Seems it's introduced in #9630 . However maybe we need to post a branch to get the error returned by `DB::valid` to locate the reason.
|
1.0
|
tikv keep panicking at components/engine_rocks/src/engine_iterator.rs:39 - ## Bug Report
### What version of TiKV are you using?
Master.
### What operating system and CPU are you using?
Doesn't matter.
### Steps to reproduce
Not sure.
### What did you expect?
TiKV keeps running without panicking.
### What did happened?
TiKV panics:
```
[2021/03/29 12:13:21.004 +08:00] [FATAL] [lib.rs:455] ["assertion failed: cfg!(feature = \\\"nortcheck\\\") || self.valid()?"] [backtrace="stack backtrace:\n 0: t[2/9902]::set_panic_hook::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_util/src/lib.rs:454\n 1: std::panicking::rust_panic_with_hook\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/panicking.rs:595\n 2: std::panicking::begin_panic_handler::{{closure}}\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/panicking.rs:495\n 3: std::sys_common::backtrace::__rust_end_short_backtrace\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/sys_common/backtrace.rs:141\n 4: rust_begin_unwind\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/std/src/panicking.rs:493\n 5: core::panicking::panic_fmt\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/core/src/panicking.rs:92\n 6: core::panicking::panic\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35//library/core/src/panicking.rs:50\n 7: <engine_rocks::engine_iterator::RocksEngineIterator as engine_traits::iterable::Iterator>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/engine_rocks/src/engine_iterator.rs:39\n 8: raftstore::store::region_snapshot::RegionIterator<S>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/raftstore/src/store/region_snapshot.rs:339\n tikv_kv::raftstore_impls::<impl tikv_kv::Iterator for raftstore::store::region_snapshot::RegionIterator<S>>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/raftstore_impls.rs:86\n tikv_kv::cursor::Cursor<I>::next\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/cursor.rs:367\n tikv::storage::txn::actions::get_old_value::get_old_value\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/actions/get_old_value.rs:34\n tikv::storage::txn::actions::prewrite::prewrite\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/actions/prewrite.rs:82\n 9: tikv::storage::txn::commands::prewrite::Prewriter<K>::prewrite\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/prewrite.rs:445\n tikv::storage::txn::commands::prewrite::Prewriter<K>::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/prewrite.rs:370\n <tikv::storage::txn::commands::prewrite::Prewrite as tikv::storage::txn::commands::WriteCommand<S,L>>::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/prewrite.rs:198\n 10: tikv::storage::txn::commands::Command::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/commands/mod.rs:539\n 11: tikv::storage::txn::scheduler::Scheduler<E,L>::process_write\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/scheduler.rs:635\n tikv::storage::txn::scheduler::Scheduler<E,L>::process_by_worker::{{closure}}::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/scheduler.rs:579\n tikv_kv::with_tls_engine::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/lib.rs:401\n std::thread::local::LocalKey<T>::try_with\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/local.rs:272\n std::thread::local::LocalKey<T>::with\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/local.rs:248\n tikv_kv::with_tls_engine\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_kv/src/lib.rs:399\n tikv::storage::txn::scheduler::Scheduler<E,L>::process_by_worker::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/src/storage/txn/scheduler.rs:578\n <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/core/src/future/mod.rs:80\n tikv_util::yatp_pool::future_pool::FuturePool::spawn::{{closure}}\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_util/src/yatp_pool/future_pool.rs:113\n <core::future::from_generator::GenFuture<T> as core::future::future::Future>::poll\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/core/src/future/mod.rs:80\n 12: <yatp::task::future::Runner as yatp::pool::runner::Runner>::handle\n at /rust/git/checkouts/yatp-e704b73c3ee279b6/6bbea16/src/task/future.rs:261\n 13: <tikv_util::yatp_pool::YatpPoolRunner<T> as yatp::pool::runner::Runner>::handle\n at /home/jenkins/agent/workspace/tikv_ghpr_build_release/tikv/components/tikv_util/src/yatp_pool/mod.rs:93\n yatp::pool::worker::WorkerThread<T,R>::run\n at /rust/git/checkouts/yatp-e704b73c3ee279b6/6bbea16/src/pool/worker.rs:48\n yatp::pool::builder::LazyBuilder<T>::build::{{closure}}\n at /rust/git/checkouts/yatp-e704b73c3ee279b6/6bbea16/src/pool/builder.rs:91\n std::sys_common::backtrace::__rust_begin_short_backtrace\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/sys_common/backtrace.rs:125\n 14: std::thread::Builder::spawn_unchecked::{{closure}}::{{closure}}\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/mod.rs:474\n <std::panic::AssertUnwindSafe<F> as core::ops::function::FnOnce<()>>::call_once\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panic.rs:322\n std::panicking::try::do_call\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panicking.rs:379\n std::panicking::try\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panicking.rs:343\n std::panic::catch_unwind\n at /rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/panic.rs:396\n std::thread::Builder::spawn_unchecked::{{closure}}\n at
/rustc/bc39d4d9c514e5fdb40a5782e6ca08924f979c35/library/std/src/thread/mod.rs:473\n core::ops::function::FnOnce::call_once{{vtable.shim[2021/03/29 12:13:36.623 +08:00
```
Seems it's introduced in #9630 . However maybe we need to post a branch to get the error returned by `DB::valid` to locate the reason.
|
non_process
|
tikv keep panicking at components engine rocks src engine iterator rs bug report what version of tikv are you using master what operating system and cpu are you using doesn t matter steps to reproduce not sure what did you expect tikv keeps running without panicking what did happened tikv panics set panic hook closure n at home jenkins agent workspace tikv ghpr build release tikv components tikv util src lib rs n std panicking rust panic with hook n at rustc library std src panicking rs n std panicking begin panic handler closure n at rustc library std src panicking rs n std sys common backtrace rust end short backtrace n at rustc library std src sys common backtrace rs n rust begin unwind n at rustc library std src panicking rs n core panicking panic fmt n at rustc library core src panicking rs n core panicking panic n at rustc library core src panicking rs n next n at home jenkins agent workspace tikv ghpr build release tikv components engine rocks src engine iterator rs n raftstore store region snapshot regioniterator next n at home jenkins agent workspace tikv ghpr build release tikv components raftstore src store region snapshot rs n tikv kv raftstore impls next n at home jenkins agent workspace tikv ghpr build release tikv components tikv kv src raftstore impls rs n tikv kv cursor cursor next n at home jenkins agent workspace tikv ghpr build release tikv components tikv kv src cursor rs n tikv storage txn actions get old value get old value n at home jenkins agent workspace tikv ghpr build release tikv src storage txn actions get old value rs n tikv storage txn actions prewrite prewrite n at home jenkins agent workspace tikv ghpr build release tikv src storage txn actions prewrite rs n tikv storage txn commands prewrite prewriter prewrite n at home jenkins agent workspace tikv ghpr build release tikv src storage txn commands prewrite rs n tikv storage txn commands prewrite prewriter process write n at home jenkins agent workspace tikv ghpr build release tikv src storage txn commands prewrite rs n process write n at home jenkins agent workspace tikv ghpr build release tikv src storage txn commands prewrite rs n tikv storage txn commands command process write n at home jenkins agent workspace tikv ghpr build release tikv src storage txn commands mod rs n tikv storage txn scheduler scheduler process write n at home jenkins agent workspace tikv ghpr build release tikv src storage txn scheduler rs n tikv storage txn scheduler scheduler process by worker closure closure n at home jenkins agent workspace tikv ghpr build release tikv src storage txn scheduler rs n tikv kv with tls engine closure n at home jenkins agent workspace tikv ghpr build release tikv components tikv kv src lib rs n std thread local localkey try with n at rustc library std src thread local rs n std thread local localkey with n at rustc library std src thread local rs n tikv kv with tls engine n at home jenkins agent workspace tikv ghpr build release tikv components tikv kv src lib rs n tikv storage txn scheduler scheduler process by worker closure n at home jenkins agent workspace tikv ghpr build release tikv src storage txn scheduler rs n as core future future future poll n at rustc library core src future mod rs n tikv util yatp pool future pool futurepool spawn closure n at home jenkins agent workspace tikv ghpr build release tikv components tikv util src yatp pool future pool rs n as core future future future poll n at rustc library core src future mod rs n handle n at rust git checkouts yatp src task future rs n as yatp pool runner runner handle n at home jenkins agent workspace tikv ghpr build release tikv components tikv util src yatp pool mod rs n yatp pool worker workerthread run n at rust git checkouts yatp src pool worker rs n yatp pool builder lazybuilder build closure n at rust git checkouts yatp src pool builder rs n std sys common backtrace rust begin short backtrace n at rustc library std src sys common backtrace rs n std thread builder spawn unchecked closure closure n at rustc library std src thread mod rs n as core ops function fnonce call once n at rustc library std src panic rs n std panicking try do call n at rustc library std src panicking rs n std panicking try n at rustc library std src panicking rs n std panic catch unwind n at rustc library std src panic rs n std thread builder spawn unchecked closure n at rustc library std src thread mod rs n core ops function fnonce call once vtable shim seems it s introduced in however maybe we need to post a branch to get the error returned by db valid to locate the reason
| 0
|
431
| 2,859,772,006
|
IssuesEvent
|
2015-06-03 12:45:28
|
genomizer/genomizer-server
|
https://api.github.com/repos/genomizer/genomizer-server
|
closed
|
Binary file resources/bowtie/bowtie-buildc tracked
|
Low priority Processing question
|
Should this be tracked? It changes all the time so it pollutes the git statuses
|
1.0
|
Binary file resources/bowtie/bowtie-buildc tracked - Should this be tracked? It changes all the time so it pollutes the git statuses
|
process
|
binary file resources bowtie bowtie buildc tracked should this be tracked it changes all the time so it pollutes the git statuses
| 1
|
644,836
| 20,988,810,047
|
IssuesEvent
|
2022-03-29 07:22:04
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
business.facebook.com - site is not usable
|
priority-critical browser-fenix engine-gecko
|
<!-- @browser: Firefox Mobile 99.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:99.0) Gecko/99.0 Firefox/99.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/101514 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://business.facebook.com/commerce/catalogs/313576712905425/products/add-manual?business_id=200974490871568#_=_
**Browser / Version**: Firefox Mobile 99.0
**Operating System**: Android 9
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
The screen doesn't fit so i cant use the accept button because its hidden. Tried desktop view, mobile view and rotated landscape view
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220320185956</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/3/6daf7cb0-aab6-4713-8d58-4ce55432aca2)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
business.facebook.com - site is not usable - <!-- @browser: Firefox Mobile 99.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:99.0) Gecko/99.0 Firefox/99.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/101514 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://business.facebook.com/commerce/catalogs/313576712905425/products/add-manual?business_id=200974490871568#_=_
**Browser / Version**: Firefox Mobile 99.0
**Operating System**: Android 9
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
The screen doesn't fit so i cant use the accept button because its hidden. Tried desktop view, mobile view and rotated landscape view
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220320185956</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2022/3/6daf7cb0-aab6-4713-8d58-4ce55432aca2)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
business facebook com site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description page not loading correctly steps to reproduce the screen doesn t fit so i cant use the accept button because its hidden tried desktop view mobile view and rotated landscape view browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
2,919
| 5,914,486,635
|
IssuesEvent
|
2017-05-22 03:07:21
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
process.hrtime() unreliable?
|
benchmark libuv process test
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v8.0.0-pre
* **Platform**: ubuntu1604-32
* **Subsystem**: process
<!-- Enter your issue details below this comment. -->
<summary>
From https://ci.nodejs.org/job/node-test-commit-linux/9926/nodes=ubuntu1604-32/consoleText:
</summary>
<details>
```console
not ok 1431 sequential/test-benchmark-http
---
duration_ms: 5.128
severity: fail
stack: |-
http/bench-parser.js
http/bench-parser.js n=1 len=1: 2,785.3291144881678
http/check_invalid_header_char.js
http/check_invalid_header_char.js n=1 key="": 2,725.8503290101344
http/check_invalid_header_char.js n=1 key="1": 3,342.2013074691513
http/check_invalid_header_char.js n=1 key="\t\t\t\t\t\t\t\t\t\tFoo bar baz": 2,301.1676124465553
http/check_invalid_header_char.js n=1 key="keep-alive": 1,714.8099561866056
http/check_invalid_header_char.js n=1 key="close": 1,836.9656266991933
http/check_invalid_header_char.js n=1 key="gzip": 2,570.79321824749
http/check_invalid_header_char.js n=1 key="20091": 2,455.1023163890354
http/check_invalid_header_char.js n=1 key="private": 3,152.505611459989
http/check_invalid_header_char.js n=1 key="text/html; charset=utf-8": 2,783.9023629763255
http/check_invalid_header_char.js n=1 key="text/plain": 1,932.2065992584191
http/check_invalid_header_char.js n=1 key="Sat, 07 May 2016 16:54:48 GMT": 2,396.380506882405
http/check_invalid_header_char.js n=1 key="SAMEORIGIN": 1,960.0387303653122
http/check_invalid_header_char.js n=1 key="en-US": 468.65297418893743
http/check_invalid_header_char.js n=1 key="Here is a value that is really a folded header value\r\n this should be supported, but it is not currently": 2,780.4655611535595
http/check_invalid_header_char.js n=1 key="中文呢": 2,113.7403692704424
http/check_invalid_header_char.js n=1 key="foo\nbar": 2,560.6424139688165
http/check_invalid_header_char.js n=1 key="": 2,448.3100539852367
http/check_is_http_token.js
http/check_is_http_token.js n=1 key="TCN": 2,627.61677785865
http/check_is_http_token.js n=1 key="ETag": 2,592.439926685799
http/check_is_http_token.js n=1 key="date": 3,146.6628067602906
http/check_is_http_token.js n=1 key="Vary": 2,676.5090827335725
http/check_is_http_token.js n=1 key="server": 1,933.1665654976161
http/check_is_http_token.js n=1 key="Server": 2,698.0431093328007
http/check_is_http_token.js n=1 key="status": 3,234.8536067000286
http/check_is_http_token.js n=1 key="version": 3,105.6575764069403
http/check_is_http_token.js n=1 key="Expires": 3,122.765271102867
http/check_is_http_token.js n=1 key="alt-svc": 2,156.3156328570753
http/check_is_http_token.js n=1 key="location": 1,954.7093835823962
/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/common.js:200
throw new Error('insufficient time precision for short benchmark');
^
Error: insufficient time precision for short benchmark
at Benchmark.end (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/common.js:200:11)
at main (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/http/check_is_http_token.js:51:9)
at Benchmark.process.nextTick (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/common.js:34:28)
at _combinedTickCallback (internal/process/next_tick.js:95:7)
at process._tickCallback (internal/process/next_tick.js:161:9)
at Function.Module.runMain (module.js:607:11)
at startup (bootstrap_node.js:144:16)
at bootstrap_node.js:561:3
assert.js:92
throw new AssertionError({
^
AssertionError [ERR_ASSERTION]: 1 === 0
at ChildProcess.child.on (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/test/sequential/test-benchmark-http.js:32:10)
at emitTwo (events.js:125:13)
at ChildProcess.emit (events.js:213:7)
at Process.ChildProcess._handle.onexit (internal/child_process.js:197:12)
...
```
</details>
## Relevant code
### Quick and easy
* `test/benchmark/check_is_http_token.js`: Pretty straightforward, but worth checking for anything funky.
* `lib/_http_common.js`: The `_checkIsHttpToken()` function is also straightforward, but worth looking at just in case...
### Slightly more involved
* `benchmark/common.js`: `Benchmark.prototype.start()` and `Benchmark.prototype.end()` are pretty straightforward and seem unlikely sources of the problem. The `throw` in the stack trace above is on line 200 (currently). This is the "what in the world is going on?!?!?!" part. `process.hrtime()` is called in one function and saved to `this._time`. Then at a later time, `process.hrtime(this._time)` is called and the result is `[0, 0]`. ???? (I don't see anything funky going on with `this` but maybe there's a subtlety I'm missing?)
### Deeper dive
* `lib/internal/process.js`: `setup_hrtime()`
* `src/node.cc`: `Hrtime()`
* `deps/uv/src/unix/linux-core.c`: `uv__hrtime()`
Anything at all going on in the above that could somehow result in this odd behavior?
## Possible causes
I'm having trouble coming up with an explanation that seems likely, but it's gotta be *something*, so...
* Somehow, this fairly fast machine running Ubuntu 16.04 nonetheless lacks nanosecond precision.
* Some subtle bug in the benchmark `common.js` that causes `this._time` to be set incorrectly.
* Small but infrequently triggered bug in the bit-shifting arithmetic in the `Deeper dive` items listed above?
* Bug in Linux
* Expected behavior in Linux unexpected in our benchmark/core code. (Maybe calling `clock_gettime()` too many times in fast succession might return the same time even though it's supposed to have nanosecond precision in this situation?)
* Bug/expected behavior specific to some container/virtual machine magic Digital Ocean is using?
Any other ideas?
/cc @nodejs/benchmarking @mscdex @joyeecheung @jasnell @nodejs/build @addaleax @saghul
Refs: https://github.com/nodejs/node/pull/12934
|
1.0
|
process.hrtime() unreliable? - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Version**: v8.0.0-pre
* **Platform**: ubuntu1604-32
* **Subsystem**: process
<!-- Enter your issue details below this comment. -->
<summary>
From https://ci.nodejs.org/job/node-test-commit-linux/9926/nodes=ubuntu1604-32/consoleText:
</summary>
<details>
```console
not ok 1431 sequential/test-benchmark-http
---
duration_ms: 5.128
severity: fail
stack: |-
http/bench-parser.js
http/bench-parser.js n=1 len=1: 2,785.3291144881678
http/check_invalid_header_char.js
http/check_invalid_header_char.js n=1 key="": 2,725.8503290101344
http/check_invalid_header_char.js n=1 key="1": 3,342.2013074691513
http/check_invalid_header_char.js n=1 key="\t\t\t\t\t\t\t\t\t\tFoo bar baz": 2,301.1676124465553
http/check_invalid_header_char.js n=1 key="keep-alive": 1,714.8099561866056
http/check_invalid_header_char.js n=1 key="close": 1,836.9656266991933
http/check_invalid_header_char.js n=1 key="gzip": 2,570.79321824749
http/check_invalid_header_char.js n=1 key="20091": 2,455.1023163890354
http/check_invalid_header_char.js n=1 key="private": 3,152.505611459989
http/check_invalid_header_char.js n=1 key="text/html; charset=utf-8": 2,783.9023629763255
http/check_invalid_header_char.js n=1 key="text/plain": 1,932.2065992584191
http/check_invalid_header_char.js n=1 key="Sat, 07 May 2016 16:54:48 GMT": 2,396.380506882405
http/check_invalid_header_char.js n=1 key="SAMEORIGIN": 1,960.0387303653122
http/check_invalid_header_char.js n=1 key="en-US": 468.65297418893743
http/check_invalid_header_char.js n=1 key="Here is a value that is really a folded header value\r\n this should be supported, but it is not currently": 2,780.4655611535595
http/check_invalid_header_char.js n=1 key="中文呢": 2,113.7403692704424
http/check_invalid_header_char.js n=1 key="foo\nbar": 2,560.6424139688165
http/check_invalid_header_char.js n=1 key="": 2,448.3100539852367
http/check_is_http_token.js
http/check_is_http_token.js n=1 key="TCN": 2,627.61677785865
http/check_is_http_token.js n=1 key="ETag": 2,592.439926685799
http/check_is_http_token.js n=1 key="date": 3,146.6628067602906
http/check_is_http_token.js n=1 key="Vary": 2,676.5090827335725
http/check_is_http_token.js n=1 key="server": 1,933.1665654976161
http/check_is_http_token.js n=1 key="Server": 2,698.0431093328007
http/check_is_http_token.js n=1 key="status": 3,234.8536067000286
http/check_is_http_token.js n=1 key="version": 3,105.6575764069403
http/check_is_http_token.js n=1 key="Expires": 3,122.765271102867
http/check_is_http_token.js n=1 key="alt-svc": 2,156.3156328570753
http/check_is_http_token.js n=1 key="location": 1,954.7093835823962
/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/common.js:200
throw new Error('insufficient time precision for short benchmark');
^
Error: insufficient time precision for short benchmark
at Benchmark.end (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/common.js:200:11)
at main (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/http/check_is_http_token.js:51:9)
at Benchmark.process.nextTick (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/benchmark/common.js:34:28)
at _combinedTickCallback (internal/process/next_tick.js:95:7)
at process._tickCallback (internal/process/next_tick.js:161:9)
at Function.Module.runMain (module.js:607:11)
at startup (bootstrap_node.js:144:16)
at bootstrap_node.js:561:3
assert.js:92
throw new AssertionError({
^
AssertionError [ERR_ASSERTION]: 1 === 0
at ChildProcess.child.on (/home/iojs/build/workspace/node-test-commit-linux/nodes/ubuntu1604-32/test/sequential/test-benchmark-http.js:32:10)
at emitTwo (events.js:125:13)
at ChildProcess.emit (events.js:213:7)
at Process.ChildProcess._handle.onexit (internal/child_process.js:197:12)
...
```
</details>
## Relevant code
### Quick and easy
* `test/benchmark/check_is_http_token.js`: Pretty straightforward, but worth checking for anything funky.
* `lib/_http_common.js`: The `_checkIsHttpToken()` function is also straightforward, but worth looking at just in case...
### Slightly more involved
* `benchmark/common.js`: `Benchmark.prototype.start()` and `Benchmark.prototype.end()` are pretty straightforward and seem unlikely sources of the problem. The `throw` in the stack trace above is on line 200 (currently). This is the "what in the world is going on?!?!?!" part. `process.hrtime()` is called in one function and saved to `this._time`. Then at a later time, `process.hrtime(this._time)` is called and the result is `[0, 0]`. ???? (I don't see anything funky going on with `this` but maybe there's a subtlety I'm missing?)
### Deeper dive
* `lib/internal/process.js`: `setup_hrtime()`
* `src/node.cc`: `Hrtime()`
* `deps/uv/src/unix/linux-core.c`: `uv__hrtime()`
Anything at all going on in the above that could somehow result in this odd behavior?
## Possible causes
I'm having trouble coming up with an explanation that seems likely, but it's gotta be *something*, so...
* Somehow, this fairly fast machine running Ubuntu 16.04 nonetheless lacks nanosecond precision.
* Some subtle bug in the benchmark `common.js` that causes `this._time` to be set incorrectly.
* Small but infrequently triggered bug in the bit-shifting arithmetic in the `Deeper dive` items listed above?
* Bug in Linux
* Expected behavior in Linux unexpected in our benchmark/core code. (Maybe calling `clock_gettime()` too many times in fast succession might return the same time even though it's supposed to have nanosecond precision in this situation?)
* Bug/expected behavior specific to some container/virtual machine magic Digital Ocean is using?
Any other ideas?
/cc @nodejs/benchmarking @mscdex @joyeecheung @jasnell @nodejs/build @addaleax @saghul
Refs: https://github.com/nodejs/node/pull/12934
|
process
|
process hrtime unreliable thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able version pre platform subsystem process from console not ok sequential test benchmark http duration ms severity fail stack http bench parser js http bench parser js n len http check invalid header char js http check invalid header char js n key http check invalid header char js n key http check invalid header char js n key t t t t t t t t t tfoo bar baz http check invalid header char js n key keep alive http check invalid header char js n key close http check invalid header char js n key gzip http check invalid header char js n key http check invalid header char js n key private http check invalid header char js n key text html charset utf http check invalid header char js n key text plain http check invalid header char js n key sat may gmt http check invalid header char js n key sameorigin http check invalid header char js n key en us http check invalid header char js n key here is a value that is really a folded header value r n this should be supported but it is not currently http check invalid header char js n key 中文呢 http check invalid header char js n key foo nbar http check invalid header char js n key http check is http token js http check is http token js n key tcn http check is http token js n key etag http check is http token js n key date http check is http token js n key vary http check is http token js n key server http check is http token js n key server http check is http token js n key status http check is http token js n key version http check is http token js n key expires http check is http token js n key alt svc http check is http token js n key location home iojs build workspace node test commit linux nodes benchmark common js throw new error insufficient time precision for short benchmark error insufficient time precision for short benchmark at benchmark end home iojs build workspace node test commit linux nodes benchmark common js at main home iojs build workspace node test commit linux nodes benchmark http check is http token js at benchmark process nexttick home iojs build workspace node test commit linux nodes benchmark common js at combinedtickcallback internal process next tick js at process tickcallback internal process next tick js at function module runmain module js at startup bootstrap node js at bootstrap node js assert js throw new assertionerror assertionerror at childprocess child on home iojs build workspace node test commit linux nodes test sequential test benchmark http js at emittwo events js at childprocess emit events js at process childprocess handle onexit internal child process js relevant code quick and easy test benchmark check is http token js pretty straightforward but worth checking for anything funky lib http common js the checkishttptoken function is also straightforward but worth looking at just in case slightly more involved benchmark common js benchmark prototype start and benchmark prototype end are pretty straightforward and seem unlikely sources of the problem the throw in the stack trace above is on line currently this is the what in the world is going on part process hrtime is called in one function and saved to this time then at a later time process hrtime this time is called and the result is i don t see anything funky going on with this but maybe there s a subtlety i m missing deeper dive lib internal process js setup hrtime src node cc hrtime deps uv src unix linux core c uv hrtime anything at all going on in the above that could somehow result in this odd behavior possible causes i m having trouble coming up with an explanation that seems likely but it s gotta be something so somehow this fairly fast machine running ubuntu nonetheless lacks nanosecond precision some subtle bug in the benchmark common js that causes this time to be set incorrectly small but infrequently triggered bug in the bit shifting arithmetic in the deeper dive items listed above bug in linux expected behavior in linux unexpected in our benchmark core code maybe calling clock gettime too many times in fast succession might return the same time even though it s supposed to have nanosecond precision in this situation bug expected behavior specific to some container virtual machine magic digital ocean is using any other ideas cc nodejs benchmarking mscdex joyeecheung jasnell nodejs build addaleax saghul refs
| 1
|
8,038
| 11,214,970,433
|
IssuesEvent
|
2020-01-07 00:20:11
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Bad performance: launching processes in parallel on Win32 / Windows
|
area-System.Diagnostics.Process tenet-performance
|
I would really like to launch multiple worker-processes in parallel with minimal blocking.
My code is heavily using async / await. Unfortunately the overall perfomance is not as good as I would expect.
For me the perfomance killer seems to be line [479 `lock (s_createProcessLock)`](https://github.com/dotnet/corefx/blob/master/src/System.Diagnostics.Process/src/System/Diagnostics/Process.Windows.cs) which disallows effectively parallel process launching.
Why is the lock static?
|
1.0
|
Bad performance: launching processes in parallel on Win32 / Windows - I would really like to launch multiple worker-processes in parallel with minimal blocking.
My code is heavily using async / await. Unfortunately the overall perfomance is not as good as I would expect.
For me the perfomance killer seems to be line [479 `lock (s_createProcessLock)`](https://github.com/dotnet/corefx/blob/master/src/System.Diagnostics.Process/src/System/Diagnostics/Process.Windows.cs) which disallows effectively parallel process launching.
Why is the lock static?
|
process
|
bad performance launching processes in parallel on windows i would really like to launch multiple worker processes in parallel with minimal blocking my code is heavily using async await unfortunately the overall perfomance is not as good as i would expect for me the perfomance killer seems to be line which disallows effectively parallel process launching why is the lock static
| 1
|
354,519
| 25,169,707,207
|
IssuesEvent
|
2022-11-11 01:18:52
|
arungudelli/go
|
https://api.github.com/repos/arungudelli/go
|
closed
|
Contains Duplicate
|
documentation good first issue ds-algo-problems hacktoberfest Stale
|
Given an integer array of numbers, return `true` if any number appears **at least twice** in the array, and return `false` if every element is distinct.
**Assumptions:**
- 1 <= nums.length <= 105
- -109 <= nums[i] <= 109
```
Input: nums = [1,2,3,1]
Output: true
```
```
Input: nums = [1,2,3,4]
Output: false
```
|
1.0
|
Contains Duplicate - Given an integer array of numbers, return `true` if any number appears **at least twice** in the array, and return `false` if every element is distinct.
**Assumptions:**
- 1 <= nums.length <= 105
- -109 <= nums[i] <= 109
```
Input: nums = [1,2,3,1]
Output: true
```
```
Input: nums = [1,2,3,4]
Output: false
```
|
non_process
|
contains duplicate given an integer array of numbers return true if any number appears at least twice in the array and return false if every element is distinct assumptions nums length nums input nums output true input nums output false
| 0
|
58,303
| 24,408,209,236
|
IssuesEvent
|
2022-10-05 09:52:54
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Search/Search Sessions] Stabilization Stage I
|
Feature:Search loe:week Team:AppServicesSv impact:medium Feature:Search Sessions
|
Change search service/search session infrastructure that improves performance, stability, and resiliency by ensuring that search sessions don’t add additional load on a cluster when the feature is not used (see https://github.com/elastic/kibana/issues/125384)
More details on reasoning and implementation in the [RFC](https://docs.google.com/document/d/1xpvW0xHmdm6w6EyYzYd01JLhiPFWF1Ij-tcJxCRLO7U/edit#heading=h.5p0n2eoe511g)
|
1.0
|
[Search/Search Sessions] Stabilization Stage I - Change search service/search session infrastructure that improves performance, stability, and resiliency by ensuring that search sessions don’t add additional load on a cluster when the feature is not used (see https://github.com/elastic/kibana/issues/125384)
More details on reasoning and implementation in the [RFC](https://docs.google.com/document/d/1xpvW0xHmdm6w6EyYzYd01JLhiPFWF1Ij-tcJxCRLO7U/edit#heading=h.5p0n2eoe511g)
|
non_process
|
stabilization stage i change search service search session infrastructure that improves performance stability and resiliency by ensuring that search sessions don’t add additional load on a cluster when the feature is not used see more details on reasoning and implementation in the
| 0
|
32,001
| 13,727,043,853
|
IssuesEvent
|
2020-10-04 03:52:55
|
invertase/react-native-firebase
|
https://api.github.com/repos/invertase/react-native-firebase
|
closed
|
The description InterstitialAd provokes a violation of admob rules
|
Service: AdMob Type: Docs Type: Stale
|
## Documentation Feedback
Hello!
Your [example of using inter-screen advertising](https://rnfirebase.io/reference/admob/interstitialad) provokes violation of [admob rules](https://support.google.com/admob/answer/6201362?hl=en).
```
import { AdEventType } from '@react-native-firebase/admob';
interstitial.onAdEvent((type) => {
if (type === AdEventType.LOADED) {
interstitial.show();
}
});
interstitial.load();
```
This is "Disallowed example: Interstitial launches after page load" in its pure form.

Or at least put a warning against showing ads immediately after loading.
You'd better duplicate this [example with tracking the load event](https://rnfirebase.io/admob/displaying-ads).
---
- 👉 Check out [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter for updates on the library.
|
1.0
|
The description InterstitialAd provokes a violation of admob rules - ## Documentation Feedback
Hello!
Your [example of using inter-screen advertising](https://rnfirebase.io/reference/admob/interstitialad) provokes violation of [admob rules](https://support.google.com/admob/answer/6201362?hl=en).
```
import { AdEventType } from '@react-native-firebase/admob';
interstitial.onAdEvent((type) => {
if (type === AdEventType.LOADED) {
interstitial.show();
}
});
interstitial.load();
```
This is "Disallowed example: Interstitial launches after page load" in its pure form.

Or at least put a warning against showing ads immediately after loading.
You'd better duplicate this [example with tracking the load event](https://rnfirebase.io/admob/displaying-ads).
---
- 👉 Check out [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter for updates on the library.
|
non_process
|
the description interstitialad provokes a violation of admob rules documentation feedback hello your provokes violation of import adeventtype from react native firebase admob interstitial onadevent type if type adeventtype loaded interstitial show interstitial load this is disallowed example interstitial launches after page load in its pure form or at least put a warning against showing ads immediately after loading you d better duplicate this 👉 check out and on twitter for updates on the library
| 0
|
803,394
| 29,175,383,401
|
IssuesEvent
|
2023-05-19 07:27:27
|
paulscherrerinstitute/scilog
|
https://api.github.com/repos/paulscherrerinstitute/scilog
|
opened
|
Display who is logged in top bar
|
frontend UX priority discussion easy
|
Especially in a shared environment or when service accounts are used it is often important to know who is currently logged in. One idea would be that in addition to the accounts thumbnail picture in the top right corner the name is given in text form.
|
1.0
|
Display who is logged in top bar - Especially in a shared environment or when service accounts are used it is often important to know who is currently logged in. One idea would be that in addition to the accounts thumbnail picture in the top right corner the name is given in text form.
|
non_process
|
display who is logged in top bar especially in a shared environment or when service accounts are used it is often important to know who is currently logged in one idea would be that in addition to the accounts thumbnail picture in the top right corner the name is given in text form
| 0
|
12,255
| 14,786,503,486
|
IssuesEvent
|
2021-01-12 05:40:50
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Sites and Apps tab > Load more icon is displaying at end of tabs even though there are only 10 set of list available and due to this continuous loading animator
|
Bug P1 Participant manager Process: Dev Process: Tested QA Process: Tested dev
|
**Steps:**
1. Navigate to sites/apps tab
2. When there are only 10 sets of data in each set, observe the load more icon
**A/R:** Load more icon is displaying at end of tabs even though there are only 10 set of list available and due to this continuous loading animator
**E/R:** Load more icon should not be displayed when there only 10 sets of data
Instance: Dev
Refer attached video:
**Sites tab:**
https://user-images.githubusercontent.com/60386291/102751445-c6d17e80-438d-11eb-8b3a-43cf3b520be1.mp4
**Apps tab:**
https://user-images.githubusercontent.com/60386291/102751489-d94bb800-438d-11eb-897e-8cbde1ab7020.mp4
|
3.0
|
[PM] Sites and Apps tab > Load more icon is displaying at end of tabs even though there are only 10 set of list available and due to this continuous loading animator - **Steps:**
1. Navigate to sites/apps tab
2. When there are only 10 sets of data in each set, observe the load more icon
**A/R:** Load more icon is displaying at end of tabs even though there are only 10 set of list available and due to this continuous loading animator
**E/R:** Load more icon should not be displayed when there only 10 sets of data
Instance: Dev
Refer attached video:
**Sites tab:**
https://user-images.githubusercontent.com/60386291/102751445-c6d17e80-438d-11eb-8b3a-43cf3b520be1.mp4
**Apps tab:**
https://user-images.githubusercontent.com/60386291/102751489-d94bb800-438d-11eb-897e-8cbde1ab7020.mp4
|
process
|
sites and apps tab load more icon is displaying at end of tabs even though there are only set of list available and due to this continuous loading animator steps navigate to sites apps tab when there are only sets of data in each set observe the load more icon a r load more icon is displaying at end of tabs even though there are only set of list available and due to this continuous loading animator e r load more icon should not be displayed when there only sets of data instance dev refer attached video sites tab apps tab
| 1
|
503,548
| 14,594,091,758
|
IssuesEvent
|
2020-12-20 03:14:50
|
DannyGlover/Solar2DTux
|
https://api.github.com/repos/DannyGlover/Solar2DTux
|
opened
|
Move releases to GitHub
|
High Priority enhancement
|
Releases are currently hosted on my server, as we were in beta testing phase up until now.
There is a build bot created (not in this repo) that I've used for all releases thus far that works well. It also handles versioning and automated changelogs based on commit history between releases.
Initially, builds will still be triggered manually. We'll get it integrated into GitHub actions at a later date.
|
1.0
|
Move releases to GitHub - Releases are currently hosted on my server, as we were in beta testing phase up until now.
There is a build bot created (not in this repo) that I've used for all releases thus far that works well. It also handles versioning and automated changelogs based on commit history between releases.
Initially, builds will still be triggered manually. We'll get it integrated into GitHub actions at a later date.
|
non_process
|
move releases to github releases are currently hosted on my server as we were in beta testing phase up until now there is a build bot created not in this repo that i ve used for all releases thus far that works well it also handles versioning and automated changelogs based on commit history between releases initially builds will still be triggered manually we ll get it integrated into github actions at a later date
| 0
|
780,518
| 27,398,526,515
|
IssuesEvent
|
2023-02-28 21:52:46
|
GoogleCloudPlatform/cloud-sql-python-connector
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-sql-python-connector
|
closed
|
system.test_asyncpg_iam_auth: test_connection_with_asyncpg_iam_auth failed
|
type: bug priority: p2 flakybot: issue flakybot: flaky
|
Note: #514 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 8bd9f8a9e376b3cbff421cf3b267be9a7701d707
buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/3579767196
status: failed
<details><summary>Test output</summary><br><pre>@pytest.fixture(name="conn")
async def setup() -> AsyncGenerator:
# initialize Cloud SQL Python Connector object
connector = await create_async_connector()
> conn: asyncpg.Connection = await connector.connect_async(
os.environ["POSTGRES_IAM_CONNECTION_NAME"],
"asyncpg",
user=os.environ["POSTGRES_IAM_USER"],
db=os.environ["POSTGRES_DB"],
enable_iam_auth=True,
)
tests\system\test_asyncpg_iam_auth.py:31:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google\cloud\sql\connector\connector.py:248: in connect_async
return await asyncio.wait_for(get_connection(), timeout)
c:\hostedtoolcache\windows\python\3.8.10\x64\lib\asyncio\tasks.py:494: in wait_for
return fut.result()
google\cloud\sql\connector\connector.py:239: in get_connection
return await connector(ip_address, instance_data.context, **kwargs)
google\cloud\sql\connector\asyncpg.py:55: in connect
return await asyncpg.connect(
.nox\system-3-8\lib\site-packages\asyncpg\connection.py:2092: in connect
return await connect_utils._connect(
.nox\system-3-8\lib\site-packages\asyncpg\connect_utils.py:881: in _connect
return await _connect_addr(
.nox\system-3-8\lib\site-packages\asyncpg\connect_utils.py:768: in _connect_addr
return await __connect_addr(params, timeout, False, *args)
.nox\system-3-8\lib\site-packages\asyncpg\connect_utils.py:831: in __connect_addr
await compat.wait_for(connected, timeout=timeout)
.nox\system-3-8\lib\site-packages\asyncpg\compat.py:56: in wait_for
return await asyncio.wait_for(fut, timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
fut = <Future finished exception=InvalidAuthorizationSpecificationError('Cloud SQL IAM service account authentication failed for user "cloud-sql-python-connector@gh-13a715-cloud-sql-pyt-dd1c5f.iam"')>
timeout = 59.92199999999997
async def wait_for(fut, timeout, *, loop=None):
"""Wait for the single Future or coroutine to complete, with timeout.
Coroutine will be wrapped in Task.
Returns result of the Future or coroutine. When a timeout occurs,
it cancels the task and raises TimeoutError. To avoid the task
cancellation, wrap it in shield().
If the wait is cancelled, the task is also cancelled.
This function is a coroutine.
"""
if loop is None:
loop = events.get_running_loop()
else:
warnings.warn("The loop argument is deprecated since Python 3.8, "
"and scheduled for removal in Python 3.10.",
DeprecationWarning, stacklevel=2)
if timeout is None:
return await fut
if timeout <= 0:
fut = ensure_future(fut, loop=loop)
if fut.done():
return fut.result()
await _cancel_and_wait(fut, loop=loop)
try:
fut.result()
except exceptions.CancelledError as exc:
raise exceptions.TimeoutError() from exc
else:
raise exceptions.TimeoutError()
waiter = loop.create_future()
timeout_handle = loop.call_later(timeout, _release_waiter, waiter)
cb = functools.partial(_release_waiter, waiter)
fut = ensure_future(fut, loop=loop)
fut.add_done_callback(cb)
try:
# wait until the future completes or the timeout
try:
await waiter
except exceptions.CancelledError:
if fut.done():
return fut.result()
else:
fut.remove_done_callback(cb)
# We must ensure that the task is not running
# after wait_for() returns.
# See https://bugs.python.org/issue32751
await _cancel_and_wait(fut, loop=loop)
raise
if fut.done():
> return fut.result()
E asyncpg.exceptions.InvalidAuthorizationSpecificationError: Cloud SQL IAM service account authentication failed for user "cloud-sql-python-connector@gh-13a715-cloud-sql-pyt-dd1c5f.iam"
c:\hostedtoolcache\windows\python\3.8.10\x64\lib\asyncio\tasks.py:494: InvalidAuthorizationSpecificationError</pre></details>
|
1.0
|
system.test_asyncpg_iam_auth: test_connection_with_asyncpg_iam_auth failed - Note: #514 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 8bd9f8a9e376b3cbff421cf3b267be9a7701d707
buildURL: https://github.com/GoogleCloudPlatform/cloud-sql-python-connector/actions/runs/3579767196
status: failed
<details><summary>Test output</summary><br><pre>@pytest.fixture(name="conn")
async def setup() -> AsyncGenerator:
# initialize Cloud SQL Python Connector object
connector = await create_async_connector()
> conn: asyncpg.Connection = await connector.connect_async(
os.environ["POSTGRES_IAM_CONNECTION_NAME"],
"asyncpg",
user=os.environ["POSTGRES_IAM_USER"],
db=os.environ["POSTGRES_DB"],
enable_iam_auth=True,
)
tests\system\test_asyncpg_iam_auth.py:31:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google\cloud\sql\connector\connector.py:248: in connect_async
return await asyncio.wait_for(get_connection(), timeout)
c:\hostedtoolcache\windows\python\3.8.10\x64\lib\asyncio\tasks.py:494: in wait_for
return fut.result()
google\cloud\sql\connector\connector.py:239: in get_connection
return await connector(ip_address, instance_data.context, **kwargs)
google\cloud\sql\connector\asyncpg.py:55: in connect
return await asyncpg.connect(
.nox\system-3-8\lib\site-packages\asyncpg\connection.py:2092: in connect
return await connect_utils._connect(
.nox\system-3-8\lib\site-packages\asyncpg\connect_utils.py:881: in _connect
return await _connect_addr(
.nox\system-3-8\lib\site-packages\asyncpg\connect_utils.py:768: in _connect_addr
return await __connect_addr(params, timeout, False, *args)
.nox\system-3-8\lib\site-packages\asyncpg\connect_utils.py:831: in __connect_addr
await compat.wait_for(connected, timeout=timeout)
.nox\system-3-8\lib\site-packages\asyncpg\compat.py:56: in wait_for
return await asyncio.wait_for(fut, timeout)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
fut = <Future finished exception=InvalidAuthorizationSpecificationError('Cloud SQL IAM service account authentication failed for user "cloud-sql-python-connector@gh-13a715-cloud-sql-pyt-dd1c5f.iam"')>
timeout = 59.92199999999997
async def wait_for(fut, timeout, *, loop=None):
"""Wait for the single Future or coroutine to complete, with timeout.
Coroutine will be wrapped in Task.
Returns result of the Future or coroutine. When a timeout occurs,
it cancels the task and raises TimeoutError. To avoid the task
cancellation, wrap it in shield().
If the wait is cancelled, the task is also cancelled.
This function is a coroutine.
"""
if loop is None:
loop = events.get_running_loop()
else:
warnings.warn("The loop argument is deprecated since Python 3.8, "
"and scheduled for removal in Python 3.10.",
DeprecationWarning, stacklevel=2)
if timeout is None:
return await fut
if timeout <= 0:
fut = ensure_future(fut, loop=loop)
if fut.done():
return fut.result()
await _cancel_and_wait(fut, loop=loop)
try:
fut.result()
except exceptions.CancelledError as exc:
raise exceptions.TimeoutError() from exc
else:
raise exceptions.TimeoutError()
waiter = loop.create_future()
timeout_handle = loop.call_later(timeout, _release_waiter, waiter)
cb = functools.partial(_release_waiter, waiter)
fut = ensure_future(fut, loop=loop)
fut.add_done_callback(cb)
try:
# wait until the future completes or the timeout
try:
await waiter
except exceptions.CancelledError:
if fut.done():
return fut.result()
else:
fut.remove_done_callback(cb)
# We must ensure that the task is not running
# after wait_for() returns.
# See https://bugs.python.org/issue32751
await _cancel_and_wait(fut, loop=loop)
raise
if fut.done():
> return fut.result()
E asyncpg.exceptions.InvalidAuthorizationSpecificationError: Cloud SQL IAM service account authentication failed for user "cloud-sql-python-connector@gh-13a715-cloud-sql-pyt-dd1c5f.iam"
c:\hostedtoolcache\windows\python\3.8.10\x64\lib\asyncio\tasks.py:494: InvalidAuthorizationSpecificationError</pre></details>
|
non_process
|
system test asyncpg iam auth test connection with asyncpg iam auth failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output pytest fixture name conn async def setup asyncgenerator initialize cloud sql python connector object connector await create async connector conn asyncpg connection await connector connect async os environ asyncpg user os environ db os environ enable iam auth true tests system test asyncpg iam auth py google cloud sql connector connector py in connect async return await asyncio wait for get connection timeout c hostedtoolcache windows python lib asyncio tasks py in wait for return fut result google cloud sql connector connector py in get connection return await connector ip address instance data context kwargs google cloud sql connector asyncpg py in connect return await asyncpg connect nox system lib site packages asyncpg connection py in connect return await connect utils connect nox system lib site packages asyncpg connect utils py in connect return await connect addr nox system lib site packages asyncpg connect utils py in connect addr return await connect addr params timeout false args nox system lib site packages asyncpg connect utils py in connect addr await compat wait for connected timeout timeout nox system lib site packages asyncpg compat py in wait for return await asyncio wait for fut timeout fut timeout async def wait for fut timeout loop none wait for the single future or coroutine to complete with timeout coroutine will be wrapped in task returns result of the future or coroutine when a timeout occurs it cancels the task and raises timeouterror to avoid the task cancellation wrap it in shield if the wait is cancelled the task is also cancelled this function is a coroutine if loop is none loop events get running loop else warnings warn the loop argument is deprecated since python and scheduled for removal in python deprecationwarning stacklevel if timeout is none return await fut if timeout fut ensure future fut loop loop if fut done return fut result await cancel and wait fut loop loop try fut result except exceptions cancellederror as exc raise exceptions timeouterror from exc else raise exceptions timeouterror waiter loop create future timeout handle loop call later timeout release waiter waiter cb functools partial release waiter waiter fut ensure future fut loop loop fut add done callback cb try wait until the future completes or the timeout try await waiter except exceptions cancellederror if fut done return fut result else fut remove done callback cb we must ensure that the task is not running after wait for returns see await cancel and wait fut loop loop raise if fut done return fut result e asyncpg exceptions invalidauthorizationspecificationerror cloud sql iam service account authentication failed for user cloud sql python connector gh cloud sql pyt iam c hostedtoolcache windows python lib asyncio tasks py invalidauthorizationspecificationerror
| 0
|
288,800
| 21,721,385,402
|
IssuesEvent
|
2022-05-11 00:45:59
|
Mod-Sim/Image-Bot
|
https://api.github.com/repos/Mod-Sim/Image-Bot
|
closed
|
Final Report
|
documentation
|
Prepare a report that summarizes:
- The problem your bot solved
- Primary features and screenshots.
- Your reflection on the development process and project.
- Any limitations and future work.
|
1.0
|
Final Report - Prepare a report that summarizes:
- The problem your bot solved
- Primary features and screenshots.
- Your reflection on the development process and project.
- Any limitations and future work.
|
non_process
|
final report prepare a report that summarizes the problem your bot solved primary features and screenshots your reflection on the development process and project any limitations and future work
| 0
|
21,690
| 30,186,594,811
|
IssuesEvent
|
2023-07-04 12:32:58
|
chipsalliance/verible
|
https://api.github.com/repos/chipsalliance/verible
|
closed
|
Ugly output for type declarations with macro + argument
|
formatter preprocessor
|
**Test case**
A type declaration like the following:
```
logic [WIDTH-1:0] data[
`NONNEGATIVE(COUNT-1)
:0];
```
gets formatted with newlines. (Until recently, this crashed -- thanks for fixing that!) I would prefer:
```
logic [WIDTH-1:0] data[`NONNEGATIVE(COUNT-1):0];
```
Is there an option to control this that I have missed?
|
1.0
|
Ugly output for type declarations with macro + argument - **Test case**
A type declaration like the following:
```
logic [WIDTH-1:0] data[
`NONNEGATIVE(COUNT-1)
:0];
```
gets formatted with newlines. (Until recently, this crashed -- thanks for fixing that!) I would prefer:
```
logic [WIDTH-1:0] data[`NONNEGATIVE(COUNT-1):0];
```
Is there an option to control this that I have missed?
|
process
|
ugly output for type declarations with macro argument test case a type declaration like the following logic data nonnegative count gets formatted with newlines until recently this crashed thanks for fixing that i would prefer logic data is there an option to control this that i have missed
| 1
|
20,151
| 28,150,429,583
|
IssuesEvent
|
2023-04-03 00:04:35
|
Cheos137/ArmorpointsPlusplus
|
https://api.github.com/repos/Cheos137/ArmorpointsPlusplus
|
closed
|
[Compatibility Request]: Dhydration mod compativiity (Fabric)
|
fabric compatibility
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Are you using the latest version currently available?
- [X] I am using the latest version currently available
### Mod name/version/id
Dehydration 1.3.3
### Link
https://www.curseforge.com/minecraft/mc-mods/dehydration
### Alternatives
Armor Points ++ still best solution for modded Minecraft. Also for "RPG-like" modpacks. But game become annoying without Dehydratation HUD.
It's duplicate of old request. But I want notice about this compability issue before AP++ v3.1.0 release
### Additional Context
_No response_
|
True
|
[Compatibility Request]: Dhydration mod compativiity (Fabric) - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Are you using the latest version currently available?
- [X] I am using the latest version currently available
### Mod name/version/id
Dehydration 1.3.3
### Link
https://www.curseforge.com/minecraft/mc-mods/dehydration
### Alternatives
Armor Points ++ still best solution for modded Minecraft. Also for "RPG-like" modpacks. But game become annoying without Dehydratation HUD.
It's duplicate of old request. But I want notice about this compability issue before AP++ v3.1.0 release
### Additional Context
_No response_
|
non_process
|
dhydration mod compativiity fabric is there an existing issue for this i have searched the existing issues are you using the latest version currently available i am using the latest version currently available mod name version id dehydration link alternatives armor points still best solution for modded minecraft also for rpg like modpacks but game become annoying without dehydratation hud it s duplicate of old request but i want notice about this compability issue before ap release additional context no response
| 0
|
10,898
| 13,675,297,234
|
IssuesEvent
|
2020-09-29 12:29:58
|
aiidateam/aiida-core
|
https://api.github.com/repos/aiidateam/aiida-core
|
closed
|
Use process type and sort by creation time for caller and called links in `verdi process show`
|
priority/nice-to-have topic/processes topic/verdi type/enhancement
|
Currently the node type is shown, which is often not very informative and the sorting is random. Showing the process type and sorting by creation time shows a more natural overview of, especially, what sub processes were called in what order.
Example, currently:
```
Called PK Type
-------- ----- ----------------
CALL 19012 WorkChainNode
CALL 19003 WorkChainNode
CALL 18999 CalcFunctionNode
CALL 18988 WorkChainNode
CALL 18979 WorkChainNode
CALL 18969 WorkChainNode
```
but which could become
```
Called PK Type
-------- ----- -----------------------
CALL 18969 PwBaseWorkChain
CALL 18979 PwBaseWorkChain
CALL 18988 HpWorkChain
CALL 18999 structure_relabel_kinds
CALL 19003 PwBaseWorkChain
CALL 19012 HpWorkChain
```
|
1.0
|
Use process type and sort by creation time for caller and called links in `verdi process show` - Currently the node type is shown, which is often not very informative and the sorting is random. Showing the process type and sorting by creation time shows a more natural overview of, especially, what sub processes were called in what order.
Example, currently:
```
Called PK Type
-------- ----- ----------------
CALL 19012 WorkChainNode
CALL 19003 WorkChainNode
CALL 18999 CalcFunctionNode
CALL 18988 WorkChainNode
CALL 18979 WorkChainNode
CALL 18969 WorkChainNode
```
but which could become
```
Called PK Type
-------- ----- -----------------------
CALL 18969 PwBaseWorkChain
CALL 18979 PwBaseWorkChain
CALL 18988 HpWorkChain
CALL 18999 structure_relabel_kinds
CALL 19003 PwBaseWorkChain
CALL 19012 HpWorkChain
```
|
process
|
use process type and sort by creation time for caller and called links in verdi process show currently the node type is shown which is often not very informative and the sorting is random showing the process type and sorting by creation time shows a more natural overview of especially what sub processes were called in what order example currently called pk type call workchainnode call workchainnode call calcfunctionnode call workchainnode call workchainnode call workchainnode but which could become called pk type call pwbaseworkchain call pwbaseworkchain call hpworkchain call structure relabel kinds call pwbaseworkchain call hpworkchain
| 1
|
110,710
| 11,708,373,223
|
IssuesEvent
|
2020-03-08 12:56:10
|
MeeatNow/MeeatNow
|
https://api.github.com/repos/MeeatNow/MeeatNow
|
opened
|
[doc] 서버통신 할 url
|
documentation
|
## 공통
- 모든 데이터는 json으로 유지하며 송수신 할때만 string으로 바꿈.
- json명세는 도메인(user, loby, location) 별 가장 아래에 있음.
- 서버포트는 8080 으로 개발 -> `localhost:8080`으로 시작하면된다.
## 유저(User)
**1. insert** `/user/insert` + request json (아래 참고)
**2. update** `/user/update/숫자` + request json
> ex) `/user/update/2`
**3. find** `/user/find/숫자`
**4. findAll** `/user/findAll`
**5. delete** `/user/delete/숫자`
※ request json 명세( 안드 -> 서버 )
```
명세
String email -> string형은 " " 안에 값 넣어주면 된다.
String name
ex)
{
"email":"2kw@n",
"name":"yhhun"
}
```
## ~로비(Loby)~ 아직 참고X
**1. insert** `/loby/insert` + request json
**2. update** `/loby/update/숫자` + request json
> ex) `/loby/update/2`
**3. find** `/loby/find/숫자`
**4. findAll** `/loby/findAll`
**5. delete** `/loby/delete/숫자`
※ request json 명세( 안드 -> 서버 )
```
String title;
String hostName;
String openLink;
Location location; -> latitude, longitude으로 값 채워준다. (ex참고)
LocalDate meetingDate; -> 패턴 : yyyy-MM-dd (ex참고)
ex)
{
"title": "chicken party",
"hostName": "kim",
"location": {
"latitude": 14143,
"longitude": 53535
},
"meetingDate": "2020-03-05"
}
```
|
1.0
|
[doc] 서버통신 할 url - ## 공통
- 모든 데이터는 json으로 유지하며 송수신 할때만 string으로 바꿈.
- json명세는 도메인(user, loby, location) 별 가장 아래에 있음.
- 서버포트는 8080 으로 개발 -> `localhost:8080`으로 시작하면된다.
## 유저(User)
**1. insert** `/user/insert` + request json (아래 참고)
**2. update** `/user/update/숫자` + request json
> ex) `/user/update/2`
**3. find** `/user/find/숫자`
**4. findAll** `/user/findAll`
**5. delete** `/user/delete/숫자`
※ request json 명세( 안드 -> 서버 )
```
명세
String email -> string형은 " " 안에 값 넣어주면 된다.
String name
ex)
{
"email":"2kw@n",
"name":"yhhun"
}
```
## ~로비(Loby)~ 아직 참고X
**1. insert** `/loby/insert` + request json
**2. update** `/loby/update/숫자` + request json
> ex) `/loby/update/2`
**3. find** `/loby/find/숫자`
**4. findAll** `/loby/findAll`
**5. delete** `/loby/delete/숫자`
※ request json 명세( 안드 -> 서버 )
```
String title;
String hostName;
String openLink;
Location location; -> latitude, longitude으로 값 채워준다. (ex참고)
LocalDate meetingDate; -> 패턴 : yyyy-MM-dd (ex참고)
ex)
{
"title": "chicken party",
"hostName": "kim",
"location": {
"latitude": 14143,
"longitude": 53535
},
"meetingDate": "2020-03-05"
}
```
|
non_process
|
서버통신 할 url 공통 모든 데이터는 json으로 유지하며 송수신 할때만 string으로 바꿈 json명세는 도메인 user loby location 별 가장 아래에 있음 서버포트는 으로 개발 localhost 으로 시작하면된다 유저 user insert user insert request json 아래 참고 update user update 숫자 request json ex user update find user find 숫자 findall user findall delete user delete 숫자 ※ request json 명세 안드 서버 명세 string email string형은 안에 값 넣어주면 된다 string name ex email n name yhhun 로비 loby 아직 참고x insert loby insert request json update loby update 숫자 request json ex loby update find loby find 숫자 findall loby findall delete loby delete 숫자 ※ request json 명세 안드 서버 string title string hostname string openlink location location latitude longitude으로 값 채워준다 ex참고 localdate meetingdate 패턴 yyyy mm dd ex참고 ex title chicken party hostname kim location latitude longitude meetingdate
| 0
|
559,033
| 16,548,518,615
|
IssuesEvent
|
2021-05-28 05:02:07
|
wso2/product-apim
|
https://api.github.com/repos/wso2/product-apim
|
opened
|
BackEnd Password of endpoint automatically get changed in publisher
|
Priority/Normal Type/Bug
|
Hi Team,
We have noticed one bug in wso2 publisher when we are trying to check suspension time out configuration in end point section it seems back end password overridden to something else due to browser cache. Ideally we have not changed back end password so it should not be 

please advise on this bug in publisher.
|
1.0
|
BackEnd Password of endpoint automatically get changed in publisher - Hi Team,
We have noticed one bug in wso2 publisher when we are trying to check suspension time out configuration in end point section it seems back end password overridden to something else due to browser cache. Ideally we have not changed back end password so it should not be 

please advise on this bug in publisher.
|
non_process
|
backend password of endpoint automatically get changed in publisher hi team we have noticed one bug in publisher when we are trying to check suspension time out configuration in end point section it seems back end password overridden to something else due to browser cache ideally we have not changed back end password so it should not be please advise on this bug in publisher
| 0
|
11,219
| 13,999,068,690
|
IssuesEvent
|
2020-10-28 10:19:07
|
cetic/tsorage
|
https://api.github.com/repos/cetic/tsorage
|
opened
|
Provide an alerting system for TSorage
|
enhancement processing
|
Grafana offers some features for sending alerts when a basic condition is verified. However
- Not all TSorage deployments come with Grafana, and for those deployments an alerting system may remain useful.
- The Grafana alerting system is based on the periodic evaluation of the time series, while we are considering a stream based approach, where a condition is only evaluated when new values are ingested for the time series of interest.
- An independent alerting system paves the way for a more generic approach, based on Complex Event Processing, that includes alerting for anomaly detection, state machine management, etc.
The purpose of this issue is to develop a platform, based on which alerts can be managed. Extensions for supporting messaging systems will be added later.
|
1.0
|
Provide an alerting system for TSorage - Grafana offers some features for sending alerts when a basic condition is verified. However
- Not all TSorage deployments come with Grafana, and for those deployments an alerting system may remain useful.
- The Grafana alerting system is based on the periodic evaluation of the time series, while we are considering a stream based approach, where a condition is only evaluated when new values are ingested for the time series of interest.
- An independent alerting system paves the way for a more generic approach, based on Complex Event Processing, that includes alerting for anomaly detection, state machine management, etc.
The purpose of this issue is to develop a platform, based on which alerts can be managed. Extensions for supporting messaging systems will be added later.
|
process
|
provide an alerting system for tsorage grafana offers some features for sending alerts when a basic condition is verified however not all tsorage deployments come with grafana and for those deployments an alerting system may remain useful the grafana alerting system is based on the periodic evaluation of the time series while we are considering a stream based approach where a condition is only evaluated when new values are ingested for the time series of interest an independent alerting system paves the way for a more generic approach based on complex event processing that includes alerting for anomaly detection state machine management etc the purpose of this issue is to develop a platform based on which alerts can be managed extensions for supporting messaging systems will be added later
| 1
|
18,982
| 24,971,402,557
|
IssuesEvent
|
2022-11-02 01:45:37
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
Documenting the process for treewide changes
|
Process
|
How to deal with tree wide changed and changes to platforms and SoCs on a wide scale, for example in Kconfig, device tree, device structures, etc.
The process should cover:
- Documentation of changes and rationale behind changes and if needed, the decision or vote leading to this.
- Announcement of changes to the wider community
- Dealing with tree wide changes in PRs and how we make sure PRs are reviewed by the right people and that "owners" of platforms or subsystems changed are given enough time to review and adapt to the changes.
----------
Final proposed language for the TSC:
- This process covers how to make "treewide" changes. The project GitHub shall add a 'treewide' label for use on issues and PRs that are treewide changes.
- The person proposing a treewide change shall write an RFC issue describing the change.
- The architecture WG shall include the issue on the agenda and discuss whether the project will accept or reject the change, with escalation to the TSC if consensus is not reached.
- The person proposing a treewide change shall email devel about the RFC if it is accepted by the architecture WG.
- The architecture WG shall specify the procedure for merging any PRs associated with each treewide change.
|
1.0
|
Documenting the process for treewide changes - How to deal with tree wide changed and changes to platforms and SoCs on a wide scale, for example in Kconfig, device tree, device structures, etc.
The process should cover:
- Documentation of changes and rationale behind changes and if needed, the decision or vote leading to this.
- Announcement of changes to the wider community
- Dealing with tree wide changes in PRs and how we make sure PRs are reviewed by the right people and that "owners" of platforms or subsystems changed are given enough time to review and adapt to the changes.
----------
Final proposed language for the TSC:
- This process covers how to make "treewide" changes. The project GitHub shall add a 'treewide' label for use on issues and PRs that are treewide changes.
- The person proposing a treewide change shall write an RFC issue describing the change.
- The architecture WG shall include the issue on the agenda and discuss whether the project will accept or reject the change, with escalation to the TSC if consensus is not reached.
- The person proposing a treewide change shall email devel about the RFC if it is accepted by the architecture WG.
- The architecture WG shall specify the procedure for merging any PRs associated with each treewide change.
|
process
|
documenting the process for treewide changes how to deal with tree wide changed and changes to platforms and socs on a wide scale for example in kconfig device tree device structures etc the process should cover documentation of changes and rationale behind changes and if needed the decision or vote leading to this announcement of changes to the wider community dealing with tree wide changes in prs and how we make sure prs are reviewed by the right people and that owners of platforms or subsystems changed are given enough time to review and adapt to the changes final proposed language for the tsc this process covers how to make treewide changes the project github shall add a treewide label for use on issues and prs that are treewide changes the person proposing a treewide change shall write an rfc issue describing the change the architecture wg shall include the issue on the agenda and discuss whether the project will accept or reject the change with escalation to the tsc if consensus is not reached the person proposing a treewide change shall email devel about the rfc if it is accepted by the architecture wg the architecture wg shall specify the procedure for merging any prs associated with each treewide change
| 1
|
13,673
| 16,419,386,653
|
IssuesEvent
|
2021-05-19 10:40:35
|
Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
|
https://api.github.com/repos/Bedrohung-der-Bienen/Transformationsfelder-Digitalisierung
|
closed
|
Tabelle Benutzer in der Datenbank anlegen
|
backend datenbank login process register process
|
# Szenario: Der Benutzer muss sich anmelden können
- **Gegeben** Der Benutzer will sich anmelden können
- **Wenn** der Benutzer sich registriert hat
- **Dann** müssen die Daten in der Datenbank abgespeichert werden
- **Und** diese beim Loginvorgang verglichen werden
-----
__Als__ Benutzer,
__möchte ich__ mich registrieren können,
__damit__ ich mich anmelden kann und alle Vorteile nutzen kann.
|
2.0
|
Tabelle Benutzer in der Datenbank anlegen - # Szenario: Der Benutzer muss sich anmelden können
- **Gegeben** Der Benutzer will sich anmelden können
- **Wenn** der Benutzer sich registriert hat
- **Dann** müssen die Daten in der Datenbank abgespeichert werden
- **Und** diese beim Loginvorgang verglichen werden
-----
__Als__ Benutzer,
__möchte ich__ mich registrieren können,
__damit__ ich mich anmelden kann und alle Vorteile nutzen kann.
|
process
|
tabelle benutzer in der datenbank anlegen szenario der benutzer muss sich anmelden können gegeben der benutzer will sich anmelden können wenn der benutzer sich registriert hat dann müssen die daten in der datenbank abgespeichert werden und diese beim loginvorgang verglichen werden als benutzer möchte ich mich registrieren können damit ich mich anmelden kann und alle vorteile nutzen kann
| 1
|
105,016
| 16,623,618,937
|
IssuesEvent
|
2021-06-03 06:44:56
|
Thanraj/OpenSSL_1.0.1
|
https://api.github.com/repos/Thanraj/OpenSSL_1.0.1
|
opened
|
CVE-2016-2109 (High) detected in opensslOpenSSL_1_0_1
|
security vulnerability
|
## CVE-2016-2109 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_1</b></p></summary>
<p>
<p>Akamai fork of openssl master.</p>
<p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_1.0.1/commit/f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc">f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/crypto/asn1/a_d2i_fp.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/crypto/asn1/a_d2i_fp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The asn1_d2i_read_bio function in crypto/asn1/a_d2i_fp.c in the ASN.1 BIO implementation in OpenSSL before 1.0.1t and 1.0.2 before 1.0.2h allows remote attackers to cause a denial of service (memory consumption) via a short invalid encoding.
<p>Publish Date: 2016-05-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2109>CVE-2016-2109</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2109">https://nvd.nist.gov/vuln/detail/CVE-2016-2109</a></p>
<p>Release Date: 2016-05-05</p>
<p>Fix Resolution: 1.0.1t,1.0.2h</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-2109 (High) detected in opensslOpenSSL_1_0_1 - ## CVE-2016-2109 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_1</b></p></summary>
<p>
<p>Akamai fork of openssl master.</p>
<p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/OpenSSL_1.0.1/commit/f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc">f1fe40536a9d3c961cc1415e9dd6d4fd002b61dc</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/crypto/asn1/a_d2i_fp.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>OpenSSL_1.0.1/crypto/asn1/a_d2i_fp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The asn1_d2i_read_bio function in crypto/asn1/a_d2i_fp.c in the ASN.1 BIO implementation in OpenSSL before 1.0.1t and 1.0.2 before 1.0.2h allows remote attackers to cause a denial of service (memory consumption) via a short invalid encoding.
<p>Publish Date: 2016-05-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2109>CVE-2016-2109</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2109">https://nvd.nist.gov/vuln/detail/CVE-2016-2109</a></p>
<p>Release Date: 2016-05-05</p>
<p>Fix Resolution: 1.0.1t,1.0.2h</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in opensslopenssl cve high severity vulnerability vulnerable library opensslopenssl akamai fork of openssl master library home page a href found in head commit a href found in base branch master vulnerable source files openssl crypto a fp c openssl crypto a fp c vulnerability details the read bio function in crypto a fp c in the asn bio implementation in openssl before and before allows remote attackers to cause a denial of service memory consumption via a short invalid encoding publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
318
| 2,764,360,441
|
IssuesEvent
|
2015-04-29 15:05:45
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
mapref's href with external scope is not resolved correctly
|
bug P2 platform-dependent preprocess
|
How to reproduce:
In Windows:
1. Create a ditamap with a mapref with an external scope such as:
```<mapref href="../generic/connectors.ditamap" format="ditamap"/>```
2. Build the ditamap with the dita.bat binary, for example:
>```DITA-OT\bin\dita.bat -f myFormat -i myDitamap.ditamap -o output```
Result:
Fails with error:
8<-----------------
Failed to run pipeline: Failed to create output directory C:..
-------------------->8
The output directory directory is resolved a merged string of a temp directory and the resolved output directory, in my case: C:\Users\EKO~\AppData\Local\Temp\temp20150427140200179\C:\Users\eko.RSD\doc\src\generic\topics
I am able to build the document under Linux with the dita binary just fine.
|
1.0
|
mapref's href with external scope is not resolved correctly - How to reproduce:
In Windows:
1. Create a ditamap with a mapref with an external scope such as:
```<mapref href="../generic/connectors.ditamap" format="ditamap"/>```
2. Build the ditamap with the dita.bat binary, for example:
>```DITA-OT\bin\dita.bat -f myFormat -i myDitamap.ditamap -o output```
Result:
Fails with error:
8<-----------------
Failed to run pipeline: Failed to create output directory C:..
-------------------->8
The output directory directory is resolved a merged string of a temp directory and the resolved output directory, in my case: C:\Users\EKO~\AppData\Local\Temp\temp20150427140200179\C:\Users\eko.RSD\doc\src\generic\topics
I am able to build the document under Linux with the dita binary just fine.
|
process
|
mapref s href with external scope is not resolved correctly how to reproduce in windows create a ditamap with a mapref with an external scope such as build the ditamap with the dita bat binary for example dita ot bin dita bat f myformat i myditamap ditamap o output result fails with error failed to run pipeline failed to create output directory c the output directory directory is resolved a merged string of a temp directory and the resolved output directory in my case c users eko appdata local temp c users eko rsd doc src generic topics i am able to build the document under linux with the dita binary just fine
| 1
|
19,538
| 25,851,483,702
|
IssuesEvent
|
2022-12-13 10:41:23
|
huggingface/datasets-server
|
https://api.github.com/repos/huggingface/datasets-server
|
closed
|
Publish a parquet file for every dataset on the Hub
|
feature request new processing step
|
To be able to apply specific processes as: stats, or random access, we need to first download the datasets on the disk.
Possibly in the parquet format.
One part will be implemented in the `datasets` library, but we also have challenges in the datasets-server project: infrastructure, workers
|
1.0
|
Publish a parquet file for every dataset on the Hub - To be able to apply specific processes as: stats, or random access, we need to first download the datasets on the disk.
Possibly in the parquet format.
One part will be implemented in the `datasets` library, but we also have challenges in the datasets-server project: infrastructure, workers
|
process
|
publish a parquet file for every dataset on the hub to be able to apply specific processes as stats or random access we need to first download the datasets on the disk possibly in the parquet format one part will be implemented in the datasets library but we also have challenges in the datasets server project infrastructure workers
| 1
|
9,238
| 12,268,824,345
|
IssuesEvent
|
2020-05-07 13:09:55
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[Processing R provider] Add section to describe porting of old scripts
|
Processing
|
## Description
A reminder to add a section on the migration of R scripts from 2.x to 3.x once the plugin is released. Till now issues are R related rather than QGIS, but users that will port their scripts can freak out when they are not working anymore.
Related to https://github.com/north-road/qgis-processing-r/issues/12
|
1.0
|
[Processing R provider] Add section to describe porting of old scripts - ## Description
A reminder to add a section on the migration of R scripts from 2.x to 3.x once the plugin is released. Till now issues are R related rather than QGIS, but users that will port their scripts can freak out when they are not working anymore.
Related to https://github.com/north-road/qgis-processing-r/issues/12
|
process
|
add section to describe porting of old scripts description a reminder to add a section on the migration of r scripts from x to x once the plugin is released till now issues are r related rather than qgis but users that will port their scripts can freak out when they are not working anymore related to
| 1
|
16,010
| 20,188,223,937
|
IssuesEvent
|
2022-02-11 01:19:27
|
savitamittalmsft/WAS-SEC-TEST
|
https://api.github.com/repos/savitamittalmsft/WAS-SEC-TEST
|
opened
|
Conduct periodic access reviews for the workload
|
WARP-Import WAF FEB 2021 Security Performance and Scalability Capacity Management Processes Security & Compliance Control-plane RBAC
|
<a href="https://docs.microsoft.com/azure/architecture/framework/security/monitor-audit#enforce-policy-compliance">Conduct periodic access reviews for the workload</a>
<p><b>Why Consider This?</b></p>
As people in the organization and on the project change, it is crucial to make sure that only the right people have access to the application infrastructure. Auditing and reviewing the access control reduces the attack vector to the application. Azure control plane depends on Azure AD and access reviews are often centrally performed often as part of internal or external audit activities.
<p><b>Context</b></p>
<p><b>Suggested Actions</b></p>
<p><span>Consider using Azure Access Reviews or Entitlement Management to periodically review access to the workload</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/active-directory/governance/access-reviews-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/active-directory/governance/access-reviews-overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/active-directory/governance/entitlement-management-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/active-directory/governance/entitlement-management-overview</span></a><span /></p>
|
1.0
|
Conduct periodic access reviews for the workload - <a href="https://docs.microsoft.com/azure/architecture/framework/security/monitor-audit#enforce-policy-compliance">Conduct periodic access reviews for the workload</a>
<p><b>Why Consider This?</b></p>
As people in the organization and on the project change, it is crucial to make sure that only the right people have access to the application infrastructure. Auditing and reviewing the access control reduces the attack vector to the application. Azure control plane depends on Azure AD and access reviews are often centrally performed often as part of internal or external audit activities.
<p><b>Context</b></p>
<p><b>Suggested Actions</b></p>
<p><span>Consider using Azure Access Reviews or Entitlement Management to periodically review access to the workload</span></p>
<p><b>Learn More</b></p>
<p><a href="https://docs.microsoft.com/en-us/azure/active-directory/governance/access-reviews-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/active-directory/governance/access-reviews-overview</span></a><span /></p><p><a href="https://docs.microsoft.com/en-us/azure/active-directory/governance/entitlement-management-overview" target="_blank"><span>https://docs.microsoft.com/en-us/azure/active-directory/governance/entitlement-management-overview</span></a><span /></p>
|
process
|
conduct periodic access reviews for the workload why consider this as people in the organization and on the project change it is crucial to make sure that only the right people have access to the application infrastructure auditing and reviewing the access control reduces the attack vector to the application azure control plane depends on azure ad and access reviews are often centrally performed often as part of internal or external audit activities context suggested actions consider using azure access reviews or entitlement management to periodically review access to the workload learn more
| 1
|
1,893
| 4,724,801,572
|
IssuesEvent
|
2016-10-18 02:31:53
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Investigate flaky tests on FreeBSD
|
child_process freebsd http test timers
|
* **Version**: master
* **Platform**: freebsd
* **Subsystem**: child_process, http, timers
I saw these three unrelated tests fail recently on CI: https://ci.nodejs.org/job/node-test-commit-freebsd/4650/nodes=freebsd10-64/console
* **parallel/test-child-process-fork-dgram**
```
not ok 129 parallel/test-child-process-fork-dgram
# TIMEOUT
---
duration_ms: 60.55
```
* **parallel/test-http-server-consumed-timeout**
```
not ok 546 parallel/test-http-server-consumed-timeout
#
# assert.js:85
# throw new assert.AssertionError({
# ^
# AssertionError: Request timeout should not fire
# at Object.exports.fail (/usr/home/iojs/build/workspace/node-test-commit-freebsd/nodes/freebsd10-64/test/common.js:443:10)
# at IncomingMessage.req.setTimeout (/usr/home/iojs/build/workspace/node-test-commit-freebsd/nodes/freebsd10-64/test/parallel/test-http-server-consumed-timeout.js:13:12)
# at emitOne (events.js:96:13)
# at IncomingMessage.emit (events.js:188:7)
# at Socket.<anonymous> (_http_server.js:310:50)
# at emitNone (events.js:86:13)
# at Socket.emit (events.js:185:7)
# at Socket._onTimeout (net.js:342:8)
# at ontimeout (timers.js:365:14)
# at tryOnTimeout (timers.js:237:5)
---
duration_ms: 1.943
```
* **parallel/test-timers-same-timeout-wrong-list-deleted**
```
not ok 970 parallel/test-timers-same-timeout-wrong-list-deleted
#
# assert.js:85
# throw new assert.AssertionError({
# ^
# AssertionError: Elapsed time does not include second timer's timeout.
# at process.<anonymous> (/usr/home/iojs/build/workspace/node-test-commit-freebsd/nodes/freebsd10-64/test/parallel/test-timers-same-timeout-wrong-list-deleted.js:26:10)
# at emitOne (events.js:101:20)
# at process.emit (events.js:188:7)
---
duration_ms: 1.69
```
|
1.0
|
Investigate flaky tests on FreeBSD - * **Version**: master
* **Platform**: freebsd
* **Subsystem**: child_process, http, timers
I saw these three unrelated tests fail recently on CI: https://ci.nodejs.org/job/node-test-commit-freebsd/4650/nodes=freebsd10-64/console
* **parallel/test-child-process-fork-dgram**
```
not ok 129 parallel/test-child-process-fork-dgram
# TIMEOUT
---
duration_ms: 60.55
```
* **parallel/test-http-server-consumed-timeout**
```
not ok 546 parallel/test-http-server-consumed-timeout
#
# assert.js:85
# throw new assert.AssertionError({
# ^
# AssertionError: Request timeout should not fire
# at Object.exports.fail (/usr/home/iojs/build/workspace/node-test-commit-freebsd/nodes/freebsd10-64/test/common.js:443:10)
# at IncomingMessage.req.setTimeout (/usr/home/iojs/build/workspace/node-test-commit-freebsd/nodes/freebsd10-64/test/parallel/test-http-server-consumed-timeout.js:13:12)
# at emitOne (events.js:96:13)
# at IncomingMessage.emit (events.js:188:7)
# at Socket.<anonymous> (_http_server.js:310:50)
# at emitNone (events.js:86:13)
# at Socket.emit (events.js:185:7)
# at Socket._onTimeout (net.js:342:8)
# at ontimeout (timers.js:365:14)
# at tryOnTimeout (timers.js:237:5)
---
duration_ms: 1.943
```
* **parallel/test-timers-same-timeout-wrong-list-deleted**
```
not ok 970 parallel/test-timers-same-timeout-wrong-list-deleted
#
# assert.js:85
# throw new assert.AssertionError({
# ^
# AssertionError: Elapsed time does not include second timer's timeout.
# at process.<anonymous> (/usr/home/iojs/build/workspace/node-test-commit-freebsd/nodes/freebsd10-64/test/parallel/test-timers-same-timeout-wrong-list-deleted.js:26:10)
# at emitOne (events.js:101:20)
# at process.emit (events.js:188:7)
---
duration_ms: 1.69
```
|
process
|
investigate flaky tests on freebsd version master platform freebsd subsystem child process http timers i saw these three unrelated tests fail recently on ci parallel test child process fork dgram not ok parallel test child process fork dgram timeout duration ms parallel test http server consumed timeout not ok parallel test http server consumed timeout assert js throw new assert assertionerror assertionerror request timeout should not fire at object exports fail usr home iojs build workspace node test commit freebsd nodes test common js at incomingmessage req settimeout usr home iojs build workspace node test commit freebsd nodes test parallel test http server consumed timeout js at emitone events js at incomingmessage emit events js at socket http server js at emitnone events js at socket emit events js at socket ontimeout net js at ontimeout timers js at tryontimeout timers js duration ms parallel test timers same timeout wrong list deleted not ok parallel test timers same timeout wrong list deleted assert js throw new assert assertionerror assertionerror elapsed time does not include second timer s timeout at process usr home iojs build workspace node test commit freebsd nodes test parallel test timers same timeout wrong list deleted js at emitone events js at process emit events js duration ms
| 1
|
155,123
| 5,949,483,575
|
IssuesEvent
|
2017-05-26 14:24:20
|
3Blades/3blades
|
https://api.github.com/repos/3Blades/3blades
|
closed
|
Manage server life cycle states with triggers
|
docs epic feature priority 1
|
This issue deals with how to manage server life cycle states.
This is a general specification for triggers to manage server states. Servers are managed as a service, which is a natural definition since we are using Docker Swarm Mode services to manage user servers. We need to define specific commands for [API](https://github.com/3blades/app-backend), [CLI Tools](https://github.com/3blades/cli-tools) as sub-issues.
Triggers are URLs that will start an action of the service whenever a POST request is sent to them. No authentication tokens should be required. The ability to revoke triggers should exist for security reasons. Example use cases could be:
- A spark job finishes and that event is sent as a POST to a 3Blades unique URL (a.k.a. webhook) which represents the resource uri to start a specific Jupyter Notebook server
- An updated image is available in DockerHub and the user decides to redeploy server based on new image update automatically.
Triggers should have UUIDs associated with them. Services have their own UUIDs. Thus, in some cases the `resource_uri` contains both the service UUID and the trigger UUID in the `resource_uri`.
Attributes:
- url: address to be used to call the trigger with a POST request
- name: a user-provided name for the trigger
- operation: the operation that the trigger call performs (see table Operations below)
- resource_uri: a unique API endpoint that represents the trigger. Otherwise known as webhook. The resource_uri should
Operations:
START: performs a redeploy service operation.
STOP: performs a scale up service operation.
REDEPLOY: restarts a service restart operation.
TERMINATE: terminates a service operation.
SCALEUP: scales up a service operation.
> Note: restarting a service task is a flag that is set when creating service with Swarm Mode.
**List Triggers**
- Lists all current triggers the service has associated to. Returns a list of Service Trigger objects.
HTTP Request
- GET /{namespace}/service/{service_uuid}/triggers/
Path Parameters
- service_uuid: the UUID of the service the triggers are associated to. As mentioned above, service = server service, since this is inline with Docker Swarm Mode concepts.
**Create a new trigger**
- Creates a new service trigger.
HTTP Request
- POST /{namespace}/service/{service_uuid}/trigger/
JSON Parameters
- name: (optional) A user provided name for the trigger
- operation: (optional) The operation to be performed by the trigger (default: “REDEPLOY”)
Get and Existing Trigger
- Creates a new trigger
HTTP Request
- GET /{namespace}/service/{service_uuid}/trigger/{trigger_uuid}
Path Parameters
- service_uuid: the UUID of the service the triggers are associated to
- trigger_uuid: the UUID of the trigger to retrieve
**Delete a Trigger**
- Deletes a trigger
HTTP Request
- DELETE /{namespace}/service/{service_uuid}/trigger/{trigger_uuid}
Path Parameters
uuid: the UUID of the associated service
trigger_uuid: the UUID of the trigger to delete
**Call a Trigger**
Executes the trigger. For SCALEUP triggers, the number of containers to scale up can be passed at the end of the trigger call url.
HTTP Request
- POST /{namespace}/service/{service_uuid}/trigger/{trigger_uuid}/call
Path Parameters
- uuid: the UUID of the associated service
- trigger_uuid: the UUID of the trigger to call
/cc @johngriebel @jkp85
#68
|
1.0
|
Manage server life cycle states with triggers - This issue deals with how to manage server life cycle states.
This is a general specification for triggers to manage server states. Servers are managed as a service, which is a natural definition since we are using Docker Swarm Mode services to manage user servers. We need to define specific commands for [API](https://github.com/3blades/app-backend), [CLI Tools](https://github.com/3blades/cli-tools) as sub-issues.
Triggers are URLs that will start an action of the service whenever a POST request is sent to them. No authentication tokens should be required. The ability to revoke triggers should exist for security reasons. Example use cases could be:
- A spark job finishes and that event is sent as a POST to a 3Blades unique URL (a.k.a. webhook) which represents the resource uri to start a specific Jupyter Notebook server
- An updated image is available in DockerHub and the user decides to redeploy server based on new image update automatically.
Triggers should have UUIDs associated with them. Services have their own UUIDs. Thus, in some cases the `resource_uri` contains both the service UUID and the trigger UUID in the `resource_uri`.
Attributes:
- url: address to be used to call the trigger with a POST request
- name: a user-provided name for the trigger
- operation: the operation that the trigger call performs (see table Operations below)
- resource_uri: a unique API endpoint that represents the trigger. Otherwise known as webhook. The resource_uri should
Operations:
START: performs a redeploy service operation.
STOP: performs a scale up service operation.
REDEPLOY: restarts a service restart operation.
TERMINATE: terminates a service operation.
SCALEUP: scales up a service operation.
> Note: restarting a service task is a flag that is set when creating service with Swarm Mode.
**List Triggers**
- Lists all current triggers the service has associated to. Returns a list of Service Trigger objects.
HTTP Request
- GET /{namespace}/service/{service_uuid}/triggers/
Path Parameters
- service_uuid: the UUID of the service the triggers are associated to. As mentioned above, service = server service, since this is inline with Docker Swarm Mode concepts.
**Create a new trigger**
- Creates a new service trigger.
HTTP Request
- POST /{namespace}/service/{service_uuid}/trigger/
JSON Parameters
- name: (optional) A user provided name for the trigger
- operation: (optional) The operation to be performed by the trigger (default: “REDEPLOY”)
Get and Existing Trigger
- Creates a new trigger
HTTP Request
- GET /{namespace}/service/{service_uuid}/trigger/{trigger_uuid}
Path Parameters
- service_uuid: the UUID of the service the triggers are associated to
- trigger_uuid: the UUID of the trigger to retrieve
**Delete a Trigger**
- Deletes a trigger
HTTP Request
- DELETE /{namespace}/service/{service_uuid}/trigger/{trigger_uuid}
Path Parameters
uuid: the UUID of the associated service
trigger_uuid: the UUID of the trigger to delete
**Call a Trigger**
Executes the trigger. For SCALEUP triggers, the number of containers to scale up can be passed at the end of the trigger call url.
HTTP Request
- POST /{namespace}/service/{service_uuid}/trigger/{trigger_uuid}/call
Path Parameters
- uuid: the UUID of the associated service
- trigger_uuid: the UUID of the trigger to call
/cc @johngriebel @jkp85
#68
|
non_process
|
manage server life cycle states with triggers this issue deals with how to manage server life cycle states this is a general specification for triggers to manage server states servers are managed as a service which is a natural definition since we are using docker swarm mode services to manage user servers we need to define specific commands for as sub issues triggers are urls that will start an action of the service whenever a post request is sent to them no authentication tokens should be required the ability to revoke triggers should exist for security reasons example use cases could be a spark job finishes and that event is sent as a post to a unique url a k a webhook which represents the resource uri to start a specific jupyter notebook server an updated image is available in dockerhub and the user decides to redeploy server based on new image update automatically triggers should have uuids associated with them services have their own uuids thus in some cases the resource uri contains both the service uuid and the trigger uuid in the resource uri attributes url address to be used to call the trigger with a post request name a user provided name for the trigger operation the operation that the trigger call performs see table operations below resource uri a unique api endpoint that represents the trigger otherwise known as webhook the resource uri should operations start performs a redeploy service operation stop performs a scale up service operation redeploy restarts a service restart operation terminate terminates a service operation scaleup scales up a service operation note restarting a service task is a flag that is set when creating service with swarm mode list triggers lists all current triggers the service has associated to returns a list of service trigger objects http request get namespace service service uuid triggers path parameters service uuid the uuid of the service the triggers are associated to as mentioned above service server service since this is inline with docker swarm mode concepts create a new trigger creates a new service trigger http request post namespace service service uuid trigger json parameters name optional a user provided name for the trigger operation optional the operation to be performed by the trigger default “redeploy” get and existing trigger creates a new trigger http request get namespace service service uuid trigger trigger uuid path parameters service uuid the uuid of the service the triggers are associated to trigger uuid the uuid of the trigger to retrieve delete a trigger deletes a trigger http request delete namespace service service uuid trigger trigger uuid path parameters uuid the uuid of the associated service trigger uuid the uuid of the trigger to delete call a trigger executes the trigger for scaleup triggers the number of containers to scale up can be passed at the end of the trigger call url http request post namespace service service uuid trigger trigger uuid call path parameters uuid the uuid of the associated service trigger uuid the uuid of the trigger to call cc johngriebel
| 0
|
524,058
| 15,195,255,301
|
IssuesEvent
|
2021-02-16 05:55:53
|
LujainKhalaf/Soengram
|
https://api.github.com/repos/LujainKhalaf/Soengram
|
closed
|
Posts on profile page should be sorted in descending order
|
.5 points bug low-risk priority-high
|
### Version
- Python Version: [e.g. 3.8.6]
### Describe the bug
The posts on the profile page are sorted in an ascending order, from oldest to newest
### Steps to reproduce
1. Go to 'profile page
2. Observe post ordering
### Expected behavior
Post should be ordered from newest -> oldest
### Actual behavior
Post should be ordered from oldest -> newest
|
1.0
|
Posts on profile page should be sorted in descending order - ### Version
- Python Version: [e.g. 3.8.6]
### Describe the bug
The posts on the profile page are sorted in an ascending order, from oldest to newest
### Steps to reproduce
1. Go to 'profile page
2. Observe post ordering
### Expected behavior
Post should be ordered from newest -> oldest
### Actual behavior
Post should be ordered from oldest -> newest
|
non_process
|
posts on profile page should be sorted in descending order version python version describe the bug the posts on the profile page are sorted in an ascending order from oldest to newest steps to reproduce go to profile page observe post ordering expected behavior post should be ordered from newest oldest actual behavior post should be ordered from oldest newest
| 0
|
20,166
| 26,720,024,114
|
IssuesEvent
|
2023-01-29 02:00:06
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Fri, 27 Jan 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
There is no result
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Shape Reconstruction from Thoracoscopic Images using Self-supervised Virtual Learning
- **Authors:** Tomoki Oya, Megumi Nakao, Tetsuya Matsuda
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2301.10863
- **Pdf link:** https://arxiv.org/pdf/2301.10863
- **Abstract**
Intraoperative shape reconstruction of organs from endoscopic camera images is a complex yet indispensable technique for image-guided surgery. To address the uncertainty in reconstructing entire shapes from single-viewpoint occluded images, we propose a framework for generative virtual learning of shape reconstruction using image translation with common latent variables between simulated and real images. As it is difficult to prepare sufficient amount of data to learn the relationship between endoscopic images and organ shapes, self-supervised virtual learning is performed using simulated images generated from statistical shape models. However, small differences between virtual and real images can degrade the estimation performance even if the simulated images are regarded as equivalent by humans. To address this issue, a Variational Autoencoder is used to convert real and simulated images into identical synthetic images. In this study, we targeted the shape reconstruction of collapsed lungs from thoracoscopic images and confirmed that virtual learning could improve the similarity between real and simulated images. Furthermore, shape reconstruction error could be improved by 16.9%.
### Graph Contrastive Learning for Skeleton-based Action Recognition
- **Authors:** Xiaohu Huang, Hao Zhou, Bin Feng, Xinggang Wang, Wenyu Liu, Jian Wang, Haocheng Feng, Junyu Han, Errui Ding, Jingdong Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2301.10900
- **Pdf link:** https://arxiv.org/pdf/2301.10900
- **Abstract**
In the field of skeleton-based action recognition, current top-performing graph convolutional networks (GCNs) exploit intra-sequence context to construct adaptive graphs for feature aggregation. However, we argue that such context is still \textit{local} since the rich cross-sequence relations have not been explicitly investigated. In this paper, we propose a graph contrastive learning framework for skeleton-based action recognition (\textit{SkeletonGCL}) to explore the \textit{global} context across all sequences. In specific, SkeletonGCL associates graph learning across sequences by enforcing graphs to be class-discriminative, \emph{i.e.,} intra-class compact and inter-class dispersed, which improves the GCN capacity to distinguish various action patterns. Besides, two memory banks are designed to enrich cross-sequence context from two complementary levels, \emph{i.e.,} instance and semantic levels, enabling graph contrastive learning in multiple context scales. Consequently, SkeletonGCL establishes a new training paradigm, and it can be seamlessly incorporated into current GCNs. Without loss of generality, we combine SkeletonGCL with three GCNs (2S-ACGN, CTR-GCN, and InfoGCN), and achieve consistent improvements on NTU60, NTU120, and NW-UCLA benchmarks. The source code will be available at \url{https://github.com/OliverHxh/SkeletonGCL}.
### Detecting Building Changes with Off-Nadir Aerial Images
- **Authors:** Chao Pang, Jiang Wu, Jian Ding, Can Song, Gui-Song Xia
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2301.10922
- **Pdf link:** https://arxiv.org/pdf/2301.10922
- **Abstract**
The tilted viewing nature of the off-nadir aerial images brings severe challenges to the building change detection (BCD) problem: the mismatch of the nearby buildings and the semantic ambiguity of the building facades. To tackle these challenges, we present a multi-task guided change detection network model, named as MTGCD-Net. The proposed model approaches the specific BCD problem by designing three auxiliary tasks, including: (1) a pixel-wise classification task to predict the roofs and facades of buildings; (2) an auxiliary task for learning the roof-to-footprint offsets of each building to account for the misalignment between building roof instances; and (3) an auxiliary task for learning the identical roof matching flow between bi-temporal aerial images to tackle the building roof mismatch problem. These auxiliary tasks provide indispensable and complementary building parsing and matching information. The predictions of the auxiliary tasks are finally fused to the main building change detection branch with a multi-modal distillation module. To train and test models for the BCD problem with off-nadir aerial images, we create a new benchmark dataset, named BANDON. Extensive experiments demonstrate that our model achieves superior performance over the previous state-of-the-art competitors.
### Vision-Language Models Performing Zero-Shot Tasks Exhibit Gender-based Disparities
- **Authors:** Melissa Hall, Laura Gustafson, Aaron Adcock, Ishan Misra, Candace Ross
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computers and Society (cs.CY); Human-Computer Interaction (cs.HC)
- **Arxiv link:** https://arxiv.org/abs/2301.11100
- **Pdf link:** https://arxiv.org/pdf/2301.11100
- **Abstract**
We explore the extent to which zero-shot vision-language models exhibit gender bias for different vision tasks. Vision models traditionally required task-specific labels for representing concepts, as well as finetuning; zero-shot models like CLIP instead perform tasks with an open-vocabulary, meaning they do not need a fixed set of labels, by using text embeddings to represent concepts. With these capabilities in mind, we ask: Do vision-language models exhibit gender bias when performing zero-shot image classification, object detection and semantic segmentation? We evaluate different vision-language models with multiple datasets across a set of concepts and find (i) all models evaluated show distinct performance differences based on the perceived gender of the person co-occurring with a given concept in the image and that aggregating analyses over all concepts can mask these concerns; (ii) model calibration (i.e. the relationship between accuracy and confidence) also differs distinctly by perceived gender, even when evaluating on similar representations of concepts; and (iii) these observed disparities align with existing gender biases in word embeddings from language models. These findings suggest that, while language greatly expands the capability of vision tasks, it can also contribute to social biases in zero-shot vision settings. Furthermore, biases can further propagate when foundational models like CLIP are used by other models to enable zero-shot capabilities.
### Multitemporal and multispectral data fusion for super-resolution of Sentinel-2 images
- **Authors:** Tomasz Tarasiewicz, Jakub Nalepa, Reuben A. Farrugia, Gianluca Valentino, Mang Chen, Johann A. Briffa, Michal Kawulok
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2301.11154
- **Pdf link:** https://arxiv.org/pdf/2301.11154
- **Abstract**
Multispectral Sentinel-2 images are a valuable source of Earth observation data, however spatial resolution of their spectral bands limited to 10 m, 20 m, and 60 m ground sampling distance remains insufficient in many cases. This problem can be addressed with super-resolution, aimed at reconstructing a high-resolution image from a low-resolution observation. For Sentinel-2, spectral information fusion allows for enhancing the 20 m and 60 m bands to the 10 m resolution. Also, there were attempts to combine multitemporal stacks of individual Sentinel-2 bands, however these two approaches have not been combined so far. In this paper, we introduce DeepSent -- a new deep network for super-resolving multitemporal series of multispectral Sentinel-2 images. It is underpinned with information fusion performed simultaneously in the spectral and temporal dimensions to generate an enlarged multispectral image. In our extensive experimental study, we demonstrate that our solution outperforms other state-of-the-art techniques that realize either multitemporal or multispectral data fusion. Furthermore, we show that the advantage of DeepSent results from how these two fusion types are combined in a single architecture, which is superior to performing such fusion in a sequential manner. Importantly, we have applied our method to super-resolve real-world Sentinel-2 images, enhancing the spatial resolution of all the spectral bands to 3.3 m nominal ground sampling distance, and we compare the outcome with very high-resolution WorldView-2 images. We will publish our implementation upon paper acceptance, and we expect it will increase the possibilities of exploiting super-resolved Sentinel-2 images in real-life applications.
### Evaluate underdiagnosis and overdiagnosis bias of deep learning model on primary open-angle glaucoma diagnosis in under-served patient populations
- **Authors:** Mingquan Lin, Yuyun Xiao, Bojian Hou, Tingyi Wanyan, Mohit Manoj Sharma, Zhangyang Wang, Fei Wang, Sarah Van Tassel, Yifan Peng
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2301.11315
- **Pdf link:** https://arxiv.org/pdf/2301.11315
- **Abstract**
In the United States, primary open-angle glaucoma (POAG) is the leading cause of blindness, especially among African American and Hispanic individuals. Deep learning has been widely used to detect POAG using fundus images as its performance is comparable to or even surpasses diagnosis by clinicians. However, human bias in clinical diagnosis may be reflected and amplified in the widely-used deep learning models, thus impacting their performance. Biases may cause (1) underdiagnosis, increasing the risks of delayed or inadequate treatment, and (2) overdiagnosis, which may increase individuals' stress, fear, well-being, and unnecessary/costly treatment. In this study, we examined the underdiagnosis and overdiagnosis when applying deep learning in POAG detection based on the Ocular Hypertension Treatment Study (OHTS) from 22 centers across 16 states in the United States. Our results show that the widely-used deep learning model can underdiagnose or overdiagnose underserved populations. The most underdiagnosed group is female younger (< 60 yrs) group, and the most overdiagnosed group is Black older (>=60 yrs) group. Biased diagnosis through traditional deep learning methods may delay disease detection, treatment and create burdens among under-served populations, thereby, raising ethical concerns about using deep learning models in ophthalmology clinics.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### The Projection-Enhancement Network (PEN)
- **Authors:** Christopher Z. Eddy, Austin Naylor, Bo Sun
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Quantitative Methods (q-bio.QM)
- **Arxiv link:** https://arxiv.org/abs/2301.10877
- **Pdf link:** https://arxiv.org/pdf/2301.10877
- **Abstract**
Contemporary approaches to instance segmentation in cell science use 2D or 3D convolutional networks depending on the experiment and data structures. However, limitations in microscopy systems or efforts to prevent phototoxicity commonly require recording sub-optimally sampled data regimes that greatly reduces the utility of such 3D data, especially in crowded environments with significant axial overlap between objects. In such regimes, 2D segmentations are both more reliable for cell morphology and easier to annotate. In this work, we propose the Projection Enhancement Network (PEN), a novel convolutional module which processes the sub-sampled 3D data and produces a 2D RGB semantic compression, and is trained in conjunction with an instance segmentation network of choice to produce 2D segmentations. Our approach combines augmentation to increase cell density using a low-density cell image dataset to train PEN, and curated datasets to evaluate PEN. We show that with PEN, the learned semantic representation in CellPose encodes depth and greatly improves segmentation performance in comparison to maximum intensity projection images as input, but does not similarly aid segmentation in region-based networks like Mask-RCNN. Finally, we dissect the segmentation strength against cell density of PEN with CellPose on disseminated cells from side-by-side spheroids. We present PEN as a data-driven solution to form compressed representations of 3D data that improve 2D segmentations from instance segmentation networks.
### BiBench: Benchmarking and Analyzing Network Binarization
- **Authors:** Haotong Qin, Mingyuan Zhang, Yifu Ding, Aoyu Li, Zhongang Cai, Ziwei Liu, Fisher Yu, Xianglong Liu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2301.11233
- **Pdf link:** https://arxiv.org/pdf/2301.11233
- **Abstract**
Network binarization emerges as one of the most promising compression approaches offering extraordinary computation and memory savings by minimizing the bit-width. However, recent research has shown that applying existing binarization algorithms to diverse tasks, architectures, and hardware in realistic scenarios is still not straightforward. Common challenges of binarization, such as accuracy degradation and efficiency limitation, suggest that its attributes are not fully understood. To close this gap, we present BiBench, a rigorously designed benchmark with in-depth analysis for network binarization. We first carefully scrutinize the requirements of binarization in the actual production and define evaluation tracks and metrics for a comprehensive and fair investigation. Then, we evaluate and analyze a series of milestone binarization algorithms that function at the operator level and with extensive influence. Our benchmark reveals that 1) the binarized operator has a crucial impact on the performance and deployability of binarized networks; 2) the accuracy of binarization varies significantly across different learning tasks and neural architectures; 3) binarization has demonstrated promising efficiency potential on edge devices despite the limited hardware support. The results and analysis also lead to a promising paradigm for accurate and efficient binarization. We believe that BiBench will contribute to the broader adoption of binarization and serve as a foundation for future research.
## Keyword: RAW
### Learning from Mistakes: Self-Regularizing Hierarchical Semantic Representations in Point Cloud Segmentation
- **Authors:** Elena Camuffo, Umberto Michieli, Simone Milani
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM); Machine Learning (stat.ML)
- **Arxiv link:** https://arxiv.org/abs/2301.11145
- **Pdf link:** https://arxiv.org/pdf/2301.11145
- **Abstract**
Recent advances in autonomous robotic technologies have highlighted the growing need for precise environmental analysis. LiDAR semantic segmentation has gained attention to accomplish fine-grained scene understanding by acting directly on raw content provided by sensors. Recent solutions showed how different learning techniques can be used to improve the performance of the model, without any architectural or dataset change. Following this trend, we present a coarse-to-fine setup that LEArns from classification mistaKes (LEAK) derived from a standard model. First, classes are clustered into macro groups according to mutual prediction errors; then, the learning process is regularized by: (1) aligning class-conditional prototypical feature representation for both fine and coarse classes, (2) weighting instances with a per-class fairness index. Our LEAK approach is very general and can be seamlessly applied on top of any segmentation architecture; indeed, experimental results showed that it enables state-of-the-art performances on different architectures, datasets and tasks, while ensuring more balanced class-wise results and faster convergence.
### Semi-Supervised Image Captioning by Adversarially Propagating Labeled Data
- **Authors:** Dong-Jin Kim, Tae-Hyun Oh, Jinsoo Choi, In So Kweon
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2301.11174
- **Pdf link:** https://arxiv.org/pdf/2301.11174
- **Abstract**
We present a novel data-efficient semi-supervised framework to improve the generalization of image captioning models. Constructing a large-scale labeled image captioning dataset is an expensive task in terms of labor, time, and cost. In contrast to manually annotating all the training samples, separately collecting uni-modal datasets is immensely easier, e.g., a large-scale image dataset and a sentence dataset. We leverage such massive unpaired image and caption data upon standard paired data by learning to associate them. To this end, our proposed semi-supervised learning method assigns pseudo-labels to unpaired samples in an adversarial learning fashion, where the joint distribution of image and caption is learned. Our method trains a captioner to learn from a paired data and to progressively associate unpaired data. This approach shows noticeable performance improvement even in challenging scenarios including out-of-task data (i.e., relational captioning, where the target task is different from the unpaired data) and web-crawled data. We also show that our proposed method is theoretically well-motivated and has a favorable global optimal property. Our extensive and comprehensive empirical results both on (1) image-based and (2) dense region-based captioning datasets followed by comprehensive analysis on the scarcely-paired COCO dataset demonstrate the consistent effectiveness of our semisupervised learning method with unpaired data compared to competing methods.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Fri, 27 Jan 23 - ## Keyword: events
There is no result
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### Shape Reconstruction from Thoracoscopic Images using Self-supervised Virtual Learning
- **Authors:** Tomoki Oya, Megumi Nakao, Tetsuya Matsuda
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2301.10863
- **Pdf link:** https://arxiv.org/pdf/2301.10863
- **Abstract**
Intraoperative shape reconstruction of organs from endoscopic camera images is a complex yet indispensable technique for image-guided surgery. To address the uncertainty in reconstructing entire shapes from single-viewpoint occluded images, we propose a framework for generative virtual learning of shape reconstruction using image translation with common latent variables between simulated and real images. As it is difficult to prepare sufficient amount of data to learn the relationship between endoscopic images and organ shapes, self-supervised virtual learning is performed using simulated images generated from statistical shape models. However, small differences between virtual and real images can degrade the estimation performance even if the simulated images are regarded as equivalent by humans. To address this issue, a Variational Autoencoder is used to convert real and simulated images into identical synthetic images. In this study, we targeted the shape reconstruction of collapsed lungs from thoracoscopic images and confirmed that virtual learning could improve the similarity between real and simulated images. Furthermore, shape reconstruction error could be improved by 16.9%.
### Graph Contrastive Learning for Skeleton-based Action Recognition
- **Authors:** Xiaohu Huang, Hao Zhou, Bin Feng, Xinggang Wang, Wenyu Liu, Jian Wang, Haocheng Feng, Junyu Han, Errui Ding, Jingdong Wang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2301.10900
- **Pdf link:** https://arxiv.org/pdf/2301.10900
- **Abstract**
In the field of skeleton-based action recognition, current top-performing graph convolutional networks (GCNs) exploit intra-sequence context to construct adaptive graphs for feature aggregation. However, we argue that such context is still \textit{local} since the rich cross-sequence relations have not been explicitly investigated. In this paper, we propose a graph contrastive learning framework for skeleton-based action recognition (\textit{SkeletonGCL}) to explore the \textit{global} context across all sequences. In specific, SkeletonGCL associates graph learning across sequences by enforcing graphs to be class-discriminative, \emph{i.e.,} intra-class compact and inter-class dispersed, which improves the GCN capacity to distinguish various action patterns. Besides, two memory banks are designed to enrich cross-sequence context from two complementary levels, \emph{i.e.,} instance and semantic levels, enabling graph contrastive learning in multiple context scales. Consequently, SkeletonGCL establishes a new training paradigm, and it can be seamlessly incorporated into current GCNs. Without loss of generality, we combine SkeletonGCL with three GCNs (2S-ACGN, CTR-GCN, and InfoGCN), and achieve consistent improvements on NTU60, NTU120, and NW-UCLA benchmarks. The source code will be available at \url{https://github.com/OliverHxh/SkeletonGCL}.
### Detecting Building Changes with Off-Nadir Aerial Images
- **Authors:** Chao Pang, Jiang Wu, Jian Ding, Can Song, Gui-Song Xia
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2301.10922
- **Pdf link:** https://arxiv.org/pdf/2301.10922
- **Abstract**
The tilted viewing nature of the off-nadir aerial images brings severe challenges to the building change detection (BCD) problem: the mismatch of the nearby buildings and the semantic ambiguity of the building facades. To tackle these challenges, we present a multi-task guided change detection network model, named as MTGCD-Net. The proposed model approaches the specific BCD problem by designing three auxiliary tasks, including: (1) a pixel-wise classification task to predict the roofs and facades of buildings; (2) an auxiliary task for learning the roof-to-footprint offsets of each building to account for the misalignment between building roof instances; and (3) an auxiliary task for learning the identical roof matching flow between bi-temporal aerial images to tackle the building roof mismatch problem. These auxiliary tasks provide indispensable and complementary building parsing and matching information. The predictions of the auxiliary tasks are finally fused to the main building change detection branch with a multi-modal distillation module. To train and test models for the BCD problem with off-nadir aerial images, we create a new benchmark dataset, named BANDON. Extensive experiments demonstrate that our model achieves superior performance over the previous state-of-the-art competitors.
### Vision-Language Models Performing Zero-Shot Tasks Exhibit Gender-based Disparities
- **Authors:** Melissa Hall, Laura Gustafson, Aaron Adcock, Ishan Misra, Candace Ross
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Computers and Society (cs.CY); Human-Computer Interaction (cs.HC)
- **Arxiv link:** https://arxiv.org/abs/2301.11100
- **Pdf link:** https://arxiv.org/pdf/2301.11100
- **Abstract**
We explore the extent to which zero-shot vision-language models exhibit gender bias for different vision tasks. Vision models traditionally required task-specific labels for representing concepts, as well as finetuning; zero-shot models like CLIP instead perform tasks with an open-vocabulary, meaning they do not need a fixed set of labels, by using text embeddings to represent concepts. With these capabilities in mind, we ask: Do vision-language models exhibit gender bias when performing zero-shot image classification, object detection and semantic segmentation? We evaluate different vision-language models with multiple datasets across a set of concepts and find (i) all models evaluated show distinct performance differences based on the perceived gender of the person co-occurring with a given concept in the image and that aggregating analyses over all concepts can mask these concerns; (ii) model calibration (i.e. the relationship between accuracy and confidence) also differs distinctly by perceived gender, even when evaluating on similar representations of concepts; and (iii) these observed disparities align with existing gender biases in word embeddings from language models. These findings suggest that, while language greatly expands the capability of vision tasks, it can also contribute to social biases in zero-shot vision settings. Furthermore, biases can further propagate when foundational models like CLIP are used by other models to enable zero-shot capabilities.
### Multitemporal and multispectral data fusion for super-resolution of Sentinel-2 images
- **Authors:** Tomasz Tarasiewicz, Jakub Nalepa, Reuben A. Farrugia, Gianluca Valentino, Mang Chen, Johann A. Briffa, Michal Kawulok
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV)
- **Arxiv link:** https://arxiv.org/abs/2301.11154
- **Pdf link:** https://arxiv.org/pdf/2301.11154
- **Abstract**
Multispectral Sentinel-2 images are a valuable source of Earth observation data, however spatial resolution of their spectral bands limited to 10 m, 20 m, and 60 m ground sampling distance remains insufficient in many cases. This problem can be addressed with super-resolution, aimed at reconstructing a high-resolution image from a low-resolution observation. For Sentinel-2, spectral information fusion allows for enhancing the 20 m and 60 m bands to the 10 m resolution. Also, there were attempts to combine multitemporal stacks of individual Sentinel-2 bands, however these two approaches have not been combined so far. In this paper, we introduce DeepSent -- a new deep network for super-resolving multitemporal series of multispectral Sentinel-2 images. It is underpinned with information fusion performed simultaneously in the spectral and temporal dimensions to generate an enlarged multispectral image. In our extensive experimental study, we demonstrate that our solution outperforms other state-of-the-art techniques that realize either multitemporal or multispectral data fusion. Furthermore, we show that the advantage of DeepSent results from how these two fusion types are combined in a single architecture, which is superior to performing such fusion in a sequential manner. Importantly, we have applied our method to super-resolve real-world Sentinel-2 images, enhancing the spatial resolution of all the spectral bands to 3.3 m nominal ground sampling distance, and we compare the outcome with very high-resolution WorldView-2 images. We will publish our implementation upon paper acceptance, and we expect it will increase the possibilities of exploiting super-resolved Sentinel-2 images in real-life applications.
### Evaluate underdiagnosis and overdiagnosis bias of deep learning model on primary open-angle glaucoma diagnosis in under-served patient populations
- **Authors:** Mingquan Lin, Yuyun Xiao, Bojian Hou, Tingyi Wanyan, Mohit Manoj Sharma, Zhangyang Wang, Fei Wang, Sarah Van Tassel, Yifan Peng
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI)
- **Arxiv link:** https://arxiv.org/abs/2301.11315
- **Pdf link:** https://arxiv.org/pdf/2301.11315
- **Abstract**
In the United States, primary open-angle glaucoma (POAG) is the leading cause of blindness, especially among African American and Hispanic individuals. Deep learning has been widely used to detect POAG using fundus images as its performance is comparable to or even surpasses diagnosis by clinicians. However, human bias in clinical diagnosis may be reflected and amplified in the widely-used deep learning models, thus impacting their performance. Biases may cause (1) underdiagnosis, increasing the risks of delayed or inadequate treatment, and (2) overdiagnosis, which may increase individuals' stress, fear, well-being, and unnecessary/costly treatment. In this study, we examined the underdiagnosis and overdiagnosis when applying deep learning in POAG detection based on the Ocular Hypertension Treatment Study (OHTS) from 22 centers across 16 states in the United States. Our results show that the widely-used deep learning model can underdiagnose or overdiagnose underserved populations. The most underdiagnosed group is female younger (< 60 yrs) group, and the most overdiagnosed group is Black older (>=60 yrs) group. Biased diagnosis through traditional deep learning methods may delay disease detection, treatment and create burdens among under-served populations, thereby, raising ethical concerns about using deep learning models in ophthalmology clinics.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### The Projection-Enhancement Network (PEN)
- **Authors:** Christopher Z. Eddy, Austin Naylor, Bo Sun
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Quantitative Methods (q-bio.QM)
- **Arxiv link:** https://arxiv.org/abs/2301.10877
- **Pdf link:** https://arxiv.org/pdf/2301.10877
- **Abstract**
Contemporary approaches to instance segmentation in cell science use 2D or 3D convolutional networks depending on the experiment and data structures. However, limitations in microscopy systems or efforts to prevent phototoxicity commonly require recording sub-optimally sampled data regimes that greatly reduces the utility of such 3D data, especially in crowded environments with significant axial overlap between objects. In such regimes, 2D segmentations are both more reliable for cell morphology and easier to annotate. In this work, we propose the Projection Enhancement Network (PEN), a novel convolutional module which processes the sub-sampled 3D data and produces a 2D RGB semantic compression, and is trained in conjunction with an instance segmentation network of choice to produce 2D segmentations. Our approach combines augmentation to increase cell density using a low-density cell image dataset to train PEN, and curated datasets to evaluate PEN. We show that with PEN, the learned semantic representation in CellPose encodes depth and greatly improves segmentation performance in comparison to maximum intensity projection images as input, but does not similarly aid segmentation in region-based networks like Mask-RCNN. Finally, we dissect the segmentation strength against cell density of PEN with CellPose on disseminated cells from side-by-side spheroids. We present PEN as a data-driven solution to form compressed representations of 3D data that improve 2D segmentations from instance segmentation networks.
### BiBench: Benchmarking and Analyzing Network Binarization
- **Authors:** Haotong Qin, Mingyuan Zhang, Yifu Ding, Aoyu Li, Zhongang Cai, Ziwei Liu, Fisher Yu, Xianglong Liu
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2301.11233
- **Pdf link:** https://arxiv.org/pdf/2301.11233
- **Abstract**
Network binarization emerges as one of the most promising compression approaches offering extraordinary computation and memory savings by minimizing the bit-width. However, recent research has shown that applying existing binarization algorithms to diverse tasks, architectures, and hardware in realistic scenarios is still not straightforward. Common challenges of binarization, such as accuracy degradation and efficiency limitation, suggest that its attributes are not fully understood. To close this gap, we present BiBench, a rigorously designed benchmark with in-depth analysis for network binarization. We first carefully scrutinize the requirements of binarization in the actual production and define evaluation tracks and metrics for a comprehensive and fair investigation. Then, we evaluate and analyze a series of milestone binarization algorithms that function at the operator level and with extensive influence. Our benchmark reveals that 1) the binarized operator has a crucial impact on the performance and deployability of binarized networks; 2) the accuracy of binarization varies significantly across different learning tasks and neural architectures; 3) binarization has demonstrated promising efficiency potential on edge devices despite the limited hardware support. The results and analysis also lead to a promising paradigm for accurate and efficient binarization. We believe that BiBench will contribute to the broader adoption of binarization and serve as a foundation for future research.
## Keyword: RAW
### Learning from Mistakes: Self-Regularizing Hierarchical Semantic Representations in Point Cloud Segmentation
- **Authors:** Elena Camuffo, Umberto Michieli, Simone Milani
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM); Machine Learning (stat.ML)
- **Arxiv link:** https://arxiv.org/abs/2301.11145
- **Pdf link:** https://arxiv.org/pdf/2301.11145
- **Abstract**
Recent advances in autonomous robotic technologies have highlighted the growing need for precise environmental analysis. LiDAR semantic segmentation has gained attention to accomplish fine-grained scene understanding by acting directly on raw content provided by sensors. Recent solutions showed how different learning techniques can be used to improve the performance of the model, without any architectural or dataset change. Following this trend, we present a coarse-to-fine setup that LEArns from classification mistaKes (LEAK) derived from a standard model. First, classes are clustered into macro groups according to mutual prediction errors; then, the learning process is regularized by: (1) aligning class-conditional prototypical feature representation for both fine and coarse classes, (2) weighting instances with a per-class fairness index. Our LEAK approach is very general and can be seamlessly applied on top of any segmentation architecture; indeed, experimental results showed that it enables state-of-the-art performances on different architectures, datasets and tasks, while ensuring more balanced class-wise results and faster convergence.
### Semi-Supervised Image Captioning by Adversarially Propagating Labeled Data
- **Authors:** Dong-Jin Kim, Tae-Hyun Oh, Jinsoo Choi, In So Kweon
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2301.11174
- **Pdf link:** https://arxiv.org/pdf/2301.11174
- **Abstract**
We present a novel data-efficient semi-supervised framework to improve the generalization of image captioning models. Constructing a large-scale labeled image captioning dataset is an expensive task in terms of labor, time, and cost. In contrast to manually annotating all the training samples, separately collecting uni-modal datasets is immensely easier, e.g., a large-scale image dataset and a sentence dataset. We leverage such massive unpaired image and caption data upon standard paired data by learning to associate them. To this end, our proposed semi-supervised learning method assigns pseudo-labels to unpaired samples in an adversarial learning fashion, where the joint distribution of image and caption is learned. Our method trains a captioner to learn from a paired data and to progressively associate unpaired data. This approach shows noticeable performance improvement even in challenging scenarios including out-of-task data (i.e., relational captioning, where the target task is different from the unpaired data) and web-crawled data. We also show that our proposed method is theoretically well-motivated and has a favorable global optimal property. Our extensive and comprehensive empirical results both on (1) image-based and (2) dense region-based captioning datasets followed by comprehensive analysis on the scarcely-paired COCO dataset demonstrate the consistent effectiveness of our semisupervised learning method with unpaired data compared to competing methods.
## Keyword: raw image
There is no result
|
process
|
new submissions for fri jan keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp shape reconstruction from thoracoscopic images using self supervised virtual learning authors tomoki oya megumi nakao tetsuya matsuda subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract intraoperative shape reconstruction of organs from endoscopic camera images is a complex yet indispensable technique for image guided surgery to address the uncertainty in reconstructing entire shapes from single viewpoint occluded images we propose a framework for generative virtual learning of shape reconstruction using image translation with common latent variables between simulated and real images as it is difficult to prepare sufficient amount of data to learn the relationship between endoscopic images and organ shapes self supervised virtual learning is performed using simulated images generated from statistical shape models however small differences between virtual and real images can degrade the estimation performance even if the simulated images are regarded as equivalent by humans to address this issue a variational autoencoder is used to convert real and simulated images into identical synthetic images in this study we targeted the shape reconstruction of collapsed lungs from thoracoscopic images and confirmed that virtual learning could improve the similarity between real and simulated images furthermore shape reconstruction error could be improved by graph contrastive learning for skeleton based action recognition authors xiaohu huang hao zhou bin feng xinggang wang wenyu liu jian wang haocheng feng junyu han errui ding jingdong wang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in the field of skeleton based action recognition current top performing graph convolutional networks gcns exploit intra sequence context to construct adaptive graphs for feature aggregation however we argue that such context is still textit local since the rich cross sequence relations have not been explicitly investigated in this paper we propose a graph contrastive learning framework for skeleton based action recognition textit skeletongcl to explore the textit global context across all sequences in specific skeletongcl associates graph learning across sequences by enforcing graphs to be class discriminative emph i e intra class compact and inter class dispersed which improves the gcn capacity to distinguish various action patterns besides two memory banks are designed to enrich cross sequence context from two complementary levels emph i e instance and semantic levels enabling graph contrastive learning in multiple context scales consequently skeletongcl establishes a new training paradigm and it can be seamlessly incorporated into current gcns without loss of generality we combine skeletongcl with three gcns acgn ctr gcn and infogcn and achieve consistent improvements on and nw ucla benchmarks the source code will be available at url detecting building changes with off nadir aerial images authors chao pang jiang wu jian ding can song gui song xia subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract the tilted viewing nature of the off nadir aerial images brings severe challenges to the building change detection bcd problem the mismatch of the nearby buildings and the semantic ambiguity of the building facades to tackle these challenges we present a multi task guided change detection network model named as mtgcd net the proposed model approaches the specific bcd problem by designing three auxiliary tasks including a pixel wise classification task to predict the roofs and facades of buildings an auxiliary task for learning the roof to footprint offsets of each building to account for the misalignment between building roof instances and an auxiliary task for learning the identical roof matching flow between bi temporal aerial images to tackle the building roof mismatch problem these auxiliary tasks provide indispensable and complementary building parsing and matching information the predictions of the auxiliary tasks are finally fused to the main building change detection branch with a multi modal distillation module to train and test models for the bcd problem with off nadir aerial images we create a new benchmark dataset named bandon extensive experiments demonstrate that our model achieves superior performance over the previous state of the art competitors vision language models performing zero shot tasks exhibit gender based disparities authors melissa hall laura gustafson aaron adcock ishan misra candace ross subjects computer vision and pattern recognition cs cv computers and society cs cy human computer interaction cs hc arxiv link pdf link abstract we explore the extent to which zero shot vision language models exhibit gender bias for different vision tasks vision models traditionally required task specific labels for representing concepts as well as finetuning zero shot models like clip instead perform tasks with an open vocabulary meaning they do not need a fixed set of labels by using text embeddings to represent concepts with these capabilities in mind we ask do vision language models exhibit gender bias when performing zero shot image classification object detection and semantic segmentation we evaluate different vision language models with multiple datasets across a set of concepts and find i all models evaluated show distinct performance differences based on the perceived gender of the person co occurring with a given concept in the image and that aggregating analyses over all concepts can mask these concerns ii model calibration i e the relationship between accuracy and confidence also differs distinctly by perceived gender even when evaluating on similar representations of concepts and iii these observed disparities align with existing gender biases in word embeddings from language models these findings suggest that while language greatly expands the capability of vision tasks it can also contribute to social biases in zero shot vision settings furthermore biases can further propagate when foundational models like clip are used by other models to enable zero shot capabilities multitemporal and multispectral data fusion for super resolution of sentinel images authors tomasz tarasiewicz jakub nalepa reuben a farrugia gianluca valentino mang chen johann a briffa michal kawulok subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract multispectral sentinel images are a valuable source of earth observation data however spatial resolution of their spectral bands limited to m m and m ground sampling distance remains insufficient in many cases this problem can be addressed with super resolution aimed at reconstructing a high resolution image from a low resolution observation for sentinel spectral information fusion allows for enhancing the m and m bands to the m resolution also there were attempts to combine multitemporal stacks of individual sentinel bands however these two approaches have not been combined so far in this paper we introduce deepsent a new deep network for super resolving multitemporal series of multispectral sentinel images it is underpinned with information fusion performed simultaneously in the spectral and temporal dimensions to generate an enlarged multispectral image in our extensive experimental study we demonstrate that our solution outperforms other state of the art techniques that realize either multitemporal or multispectral data fusion furthermore we show that the advantage of deepsent results from how these two fusion types are combined in a single architecture which is superior to performing such fusion in a sequential manner importantly we have applied our method to super resolve real world sentinel images enhancing the spatial resolution of all the spectral bands to m nominal ground sampling distance and we compare the outcome with very high resolution worldview images we will publish our implementation upon paper acceptance and we expect it will increase the possibilities of exploiting super resolved sentinel images in real life applications evaluate underdiagnosis and overdiagnosis bias of deep learning model on primary open angle glaucoma diagnosis in under served patient populations authors mingquan lin yuyun xiao bojian hou tingyi wanyan mohit manoj sharma zhangyang wang fei wang sarah van tassel yifan peng subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract in the united states primary open angle glaucoma poag is the leading cause of blindness especially among african american and hispanic individuals deep learning has been widely used to detect poag using fundus images as its performance is comparable to or even surpasses diagnosis by clinicians however human bias in clinical diagnosis may be reflected and amplified in the widely used deep learning models thus impacting their performance biases may cause underdiagnosis increasing the risks of delayed or inadequate treatment and overdiagnosis which may increase individuals stress fear well being and unnecessary costly treatment in this study we examined the underdiagnosis and overdiagnosis when applying deep learning in poag detection based on the ocular hypertension treatment study ohts from centers across states in the united states our results show that the widely used deep learning model can underdiagnose or overdiagnose underserved populations the most underdiagnosed group is female younger yrs group biased diagnosis through traditional deep learning methods may delay disease detection treatment and create burdens among under served populations thereby raising ethical concerns about using deep learning models in ophthalmology clinics keyword image signal processing there is no result keyword image signal process there is no result keyword compression the projection enhancement network pen authors christopher z eddy austin naylor bo sun subjects computer vision and pattern recognition cs cv quantitative methods q bio qm arxiv link pdf link abstract contemporary approaches to instance segmentation in cell science use or convolutional networks depending on the experiment and data structures however limitations in microscopy systems or efforts to prevent phototoxicity commonly require recording sub optimally sampled data regimes that greatly reduces the utility of such data especially in crowded environments with significant axial overlap between objects in such regimes segmentations are both more reliable for cell morphology and easier to annotate in this work we propose the projection enhancement network pen a novel convolutional module which processes the sub sampled data and produces a rgb semantic compression and is trained in conjunction with an instance segmentation network of choice to produce segmentations our approach combines augmentation to increase cell density using a low density cell image dataset to train pen and curated datasets to evaluate pen we show that with pen the learned semantic representation in cellpose encodes depth and greatly improves segmentation performance in comparison to maximum intensity projection images as input but does not similarly aid segmentation in region based networks like mask rcnn finally we dissect the segmentation strength against cell density of pen with cellpose on disseminated cells from side by side spheroids we present pen as a data driven solution to form compressed representations of data that improve segmentations from instance segmentation networks bibench benchmarking and analyzing network binarization authors haotong qin mingyuan zhang yifu ding aoyu li zhongang cai ziwei liu fisher yu xianglong liu subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract network binarization emerges as one of the most promising compression approaches offering extraordinary computation and memory savings by minimizing the bit width however recent research has shown that applying existing binarization algorithms to diverse tasks architectures and hardware in realistic scenarios is still not straightforward common challenges of binarization such as accuracy degradation and efficiency limitation suggest that its attributes are not fully understood to close this gap we present bibench a rigorously designed benchmark with in depth analysis for network binarization we first carefully scrutinize the requirements of binarization in the actual production and define evaluation tracks and metrics for a comprehensive and fair investigation then we evaluate and analyze a series of milestone binarization algorithms that function at the operator level and with extensive influence our benchmark reveals that the binarized operator has a crucial impact on the performance and deployability of binarized networks the accuracy of binarization varies significantly across different learning tasks and neural architectures binarization has demonstrated promising efficiency potential on edge devices despite the limited hardware support the results and analysis also lead to a promising paradigm for accurate and efficient binarization we believe that bibench will contribute to the broader adoption of binarization and serve as a foundation for future research keyword raw learning from mistakes self regularizing hierarchical semantic representations in point cloud segmentation authors elena camuffo umberto michieli simone milani subjects computer vision and pattern recognition cs cv multimedia cs mm machine learning stat ml arxiv link pdf link abstract recent advances in autonomous robotic technologies have highlighted the growing need for precise environmental analysis lidar semantic segmentation has gained attention to accomplish fine grained scene understanding by acting directly on raw content provided by sensors recent solutions showed how different learning techniques can be used to improve the performance of the model without any architectural or dataset change following this trend we present a coarse to fine setup that learns from classification mistakes leak derived from a standard model first classes are clustered into macro groups according to mutual prediction errors then the learning process is regularized by aligning class conditional prototypical feature representation for both fine and coarse classes weighting instances with a per class fairness index our leak approach is very general and can be seamlessly applied on top of any segmentation architecture indeed experimental results showed that it enables state of the art performances on different architectures datasets and tasks while ensuring more balanced class wise results and faster convergence semi supervised image captioning by adversarially propagating labeled data authors dong jin kim tae hyun oh jinsoo choi in so kweon subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computation and language cs cl machine learning cs lg arxiv link pdf link abstract we present a novel data efficient semi supervised framework to improve the generalization of image captioning models constructing a large scale labeled image captioning dataset is an expensive task in terms of labor time and cost in contrast to manually annotating all the training samples separately collecting uni modal datasets is immensely easier e g a large scale image dataset and a sentence dataset we leverage such massive unpaired image and caption data upon standard paired data by learning to associate them to this end our proposed semi supervised learning method assigns pseudo labels to unpaired samples in an adversarial learning fashion where the joint distribution of image and caption is learned our method trains a captioner to learn from a paired data and to progressively associate unpaired data this approach shows noticeable performance improvement even in challenging scenarios including out of task data i e relational captioning where the target task is different from the unpaired data and web crawled data we also show that our proposed method is theoretically well motivated and has a favorable global optimal property our extensive and comprehensive empirical results both on image based and dense region based captioning datasets followed by comprehensive analysis on the scarcely paired coco dataset demonstrate the consistent effectiveness of our semisupervised learning method with unpaired data compared to competing methods keyword raw image there is no result
| 1
|
273,280
| 29,820,270,668
|
IssuesEvent
|
2023-06-17 01:19:08
|
pazhanivel07/frameworks_base_2021-0970
|
https://api.github.com/repos/pazhanivel07/frameworks_base_2021-0970
|
closed
|
CVE-2023-21089 (High) detected in baseandroid-10.0.0_r44 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2023-21089 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary>
<p>
<p>Android framework classes and services</p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_2021-0970/commit/ad3ed522c9ac4d72ed6d51ed523780ad73330cbe">ad3ed522c9ac4d72ed6d51ed523780ad73330cbe</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/java/com/android/server/am/ActivityManagerService.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In startInstrumentation of ActivityManagerService.java, there is a possible way to keep the foreground service alive while the app is in the background. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-11 Android-12 Android-12L Android-13Android ID: A-237766679
<p>Publish Date: 2023-04-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-21089>CVE-2023-21089</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://android.googlesource.com/platform/frameworks/base/+/2d20bee73d9ba1c56d49041991e141ba3fe68c5a">https://android.googlesource.com/platform/frameworks/base/+/2d20bee73d9ba1c56d49041991e141ba3fe68c5a</a></p>
<p>Release Date: 2023-04-19</p>
<p>Fix Resolution: android-13.0.0_r38</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-21089 (High) detected in baseandroid-10.0.0_r44 - autoclosed - ## CVE-2023-21089 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary>
<p>
<p>Android framework classes and services</p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_2021-0970/commit/ad3ed522c9ac4d72ed6d51ed523780ad73330cbe">ad3ed522c9ac4d72ed6d51ed523780ad73330cbe</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/core/java/com/android/server/am/ActivityManagerService.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In startInstrumentation of ActivityManagerService.java, there is a possible way to keep the foreground service alive while the app is in the background. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-11 Android-12 Android-12L Android-13Android ID: A-237766679
<p>Publish Date: 2023-04-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-21089>CVE-2023-21089</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://android.googlesource.com/platform/frameworks/base/+/2d20bee73d9ba1c56d49041991e141ba3fe68c5a">https://android.googlesource.com/platform/frameworks/base/+/2d20bee73d9ba1c56d49041991e141ba3fe68c5a</a></p>
<p>Release Date: 2023-04-19</p>
<p>Fix Resolution: android-13.0.0_r38</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in baseandroid autoclosed cve high severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files services core java com android server am activitymanagerservice java vulnerability details in startinstrumentation of activitymanagerservice java there is a possible way to keep the foreground service alive while the app is in the background this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
| 0
|
161,273
| 6,111,798,337
|
IssuesEvent
|
2017-06-21 17:52:21
|
TerraFusion/basicFusion
|
https://api.github.com/repos/TerraFusion/basicFusion
|
closed
|
Check MISR missing input granules
|
High Priority
|
Landon,
Again, I assign this to you, you don't need to do this now. You can assign to others. We want to check all MISR missing granules for the TERRA fusion files. At roger, under MISR's radiance data directory, ****/MI1B2E.003/
Under the subdirectories like 2013.05.18, you may see real HDF4 files like MISR_AM1_GRP_ELLIPSOID_GM_P109_O071352_AA_F03_0024.hdf
In this filename, the _O071352 is the orbit number. For this case, orbit 71352. For our input granules,
For the fusion product, the starting orbit should be 1000 and the ending orbit should be 85302.
If there are no missing granules, we should see all the orbit numbers between 1000 and 85302 in the filenames under MI1B2E.003 from
MISR_XX......XX_O001000_XX...X.hdf, to
MISR_XX......XX_O085302_XX...X.hdf
But just in our 2013, we find quite a few missing orbits. For examples, we cannot find files for the orbit 72454.
This ticket is to find all the missing orbits under the MI1B2E.003 by searching the keyword O00???? for the orbit number <10000 and O0????? for the number >=10000 and <=85302
|
1.0
|
Check MISR missing input granules - Landon,
Again, I assign this to you, you don't need to do this now. You can assign to others. We want to check all MISR missing granules for the TERRA fusion files. At roger, under MISR's radiance data directory, ****/MI1B2E.003/
Under the subdirectories like 2013.05.18, you may see real HDF4 files like MISR_AM1_GRP_ELLIPSOID_GM_P109_O071352_AA_F03_0024.hdf
In this filename, the _O071352 is the orbit number. For this case, orbit 71352. For our input granules,
For the fusion product, the starting orbit should be 1000 and the ending orbit should be 85302.
If there are no missing granules, we should see all the orbit numbers between 1000 and 85302 in the filenames under MI1B2E.003 from
MISR_XX......XX_O001000_XX...X.hdf, to
MISR_XX......XX_O085302_XX...X.hdf
But just in our 2013, we find quite a few missing orbits. For examples, we cannot find files for the orbit 72454.
This ticket is to find all the missing orbits under the MI1B2E.003 by searching the keyword O00???? for the orbit number <10000 and O0????? for the number >=10000 and <=85302
|
non_process
|
check misr missing input granules landon again i assign this to you you don t need to do this now you can assign to others we want to check all misr missing granules for the terra fusion files at roger under misr s radiance data directory under the subdirectories like you may see real files like misr grp ellipsoid gm aa hdf in this filename the is the orbit number for this case orbit for our input granules for the fusion product the starting orbit should be and the ending orbit should be if there are no missing granules we should see all the orbit numbers between and in the filenames under from misr xx xx xx x hdf to misr xx xx xx x hdf but just in our we find quite a few missing orbits for examples we cannot find files for the orbit this ticket is to find all the missing orbits under the by searching the keyword for the orbit number and
| 0
|
16,481
| 21,428,259,780
|
IssuesEvent
|
2022-04-23 01:14:28
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
SQLite driver generates incorrect SQL for adding months/quarters to a date
|
Type:Bug Priority:P2 Database/SQLite Querying/Processor .Correctness .Backend
|
Suppose you want to do something like add 3 months to March 15th. When compiling to SQL we'll be executing functions that look something like this:
```clj
(sql.qp/add-interval-honeysql-form
:sqlite
[:absolute-datetime #t "2022-03-15" :day]
3 :month)
```
That incorrectly compiles to
```sql
datetime(date(date(date('2022-03-15')), 'start of month'), '+3 months')
```
which actually gives use **June 1st**.
We're incorrectly truncating the date to the first of the month before adding the 3 month interval to it.
|
1.0
|
SQLite driver generates incorrect SQL for adding months/quarters to a date - Suppose you want to do something like add 3 months to March 15th. When compiling to SQL we'll be executing functions that look something like this:
```clj
(sql.qp/add-interval-honeysql-form
:sqlite
[:absolute-datetime #t "2022-03-15" :day]
3 :month)
```
That incorrectly compiles to
```sql
datetime(date(date(date('2022-03-15')), 'start of month'), '+3 months')
```
which actually gives use **June 1st**.
We're incorrectly truncating the date to the first of the month before adding the 3 month interval to it.
|
process
|
sqlite driver generates incorrect sql for adding months quarters to a date suppose you want to do something like add months to march when compiling to sql we ll be executing functions that look something like this clj sql qp add interval honeysql form sqlite month that incorrectly compiles to sql datetime date date date start of month months which actually gives use june we re incorrectly truncating the date to the first of the month before adding the month interval to it
| 1
|
20,509
| 27,167,378,083
|
IssuesEvent
|
2023-02-17 16:21:38
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Example is confusing and not helpful.
|
devops/prod doc-bug Pri2 devops-cicd-process/tech
|
In the section "Use a template parameter as part of a condition", the example of passing a parameter to a template explains nothing, because both the parameter declarations are identical. If the intent was to inform, then a different parameter value should have been used so that it was clear which parameter declaration took effect and which did not.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 21e5cee4-eaae-3a96-db91-540ac759e83a
* Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92
* Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Example is confusing and not helpful. - In the section "Use a template parameter as part of a condition", the example of passing a parameter to a template explains nothing, because both the parameter declarations are identical. If the intent was to inform, then a different parameter value should have been used so that it was clear which parameter declaration took effect and which did not.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 21e5cee4-eaae-3a96-db91-540ac759e83a
* Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92
* Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
example is confusing and not helpful in the section use a template parameter as part of a condition the example of passing a parameter to a template explains nothing because both the parameter declarations are identical if the intent was to inform then a different parameter value should have been used so that it was clear which parameter declaration took effect and which did not document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id eaae version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
8,335
| 11,495,079,943
|
IssuesEvent
|
2020-02-12 03:34:52
|
fluent/fluent-bit
|
https://api.github.com/repos/fluent/fluent-bit
|
closed
|
[needs #1672] Add option to validate configuration
|
enhancement work-in-process
|
**Is your feature request related to a problem? Please describe.**
<!--- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
It seems fluent-bit is missing the capability to only test its config file (and the files it recursively includes) for valid syntax.
At the moment when I deploy a configuration I have to restart fluent-bit and hope that everything works and if there is an issue with the config fluent-bit could halt with the risk of losing event records.
**Describe the solution you'd like**
<!--- A clear and concise description of what you want to happen. -->
In order to prevent unwanted downtimes I'd like to use the fluent-bit binary to validate the configuration after an update (be it via config management or manually). The service should only be attempted to be restarted (by my orchestration of choice) if the config check is considered successful.
Many applications allow you to check if their configuration contains errors, e.g.:
* `nginx -c /etc/nginx/nginx.conf -t`
* `apachectl configtest`
**Describe alternatives you've considered**
<!--- A clear and concise description of any alternative solutions or features you've considered. -->
The only issue that's somewhat similar is #252 but does not offer a solution.
**Additional context**
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
<!--- Add any other context or screenshots about the feature request here. -->
```
OS:
CentOS 7
[Fluent Bit]
Edition Community Edition
Version 1.3.3
```
Here's an example Ansible playbook:
```yml
- hosts: loadbalancers
become: yes
handlers:
- name: Restart fluent-bit
service:
name: td-agent-bit
state: restarted
enabled: yes
tasks:
- name: Add fluent-bit repo
yum_repository:
name: TD Agent Bit
file: td-agent-bit
baseurl: http://packages.fluentbit.io/centos/7
enabled: true
gpgkey: http://packages.fluentbit.io/fluentbit.key
gpgcheck: true
- name: Install latest version of fluent-bit
yum:
name: td-agent-bit
state: latest
- name: Ensure fluent-bit is running
systemd:
name: td-agent-bit
enabled: true
state: started
- name: Deploy fluent-bit config
template:
src: "templates/{{ item }}.j2"
dest: "/etc/td-agent-bit/{{ item }}"
owner: root
group: root
mode: 0644
validate: "/opt/td-agent-bit/bin/td-agent-bit --check"
loop:
- td-agent-bit.conf
- output-gelf.conf
notify: Restart fluent-bit
```
|
1.0
|
[needs #1672] Add option to validate configuration - **Is your feature request related to a problem? Please describe.**
<!--- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
It seems fluent-bit is missing the capability to only test its config file (and the files it recursively includes) for valid syntax.
At the moment when I deploy a configuration I have to restart fluent-bit and hope that everything works and if there is an issue with the config fluent-bit could halt with the risk of losing event records.
**Describe the solution you'd like**
<!--- A clear and concise description of what you want to happen. -->
In order to prevent unwanted downtimes I'd like to use the fluent-bit binary to validate the configuration after an update (be it via config management or manually). The service should only be attempted to be restarted (by my orchestration of choice) if the config check is considered successful.
Many applications allow you to check if their configuration contains errors, e.g.:
* `nginx -c /etc/nginx/nginx.conf -t`
* `apachectl configtest`
**Describe alternatives you've considered**
<!--- A clear and concise description of any alternative solutions or features you've considered. -->
The only issue that's somewhat similar is #252 but does not offer a solution.
**Additional context**
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
<!--- Add any other context or screenshots about the feature request here. -->
```
OS:
CentOS 7
[Fluent Bit]
Edition Community Edition
Version 1.3.3
```
Here's an example Ansible playbook:
```yml
- hosts: loadbalancers
become: yes
handlers:
- name: Restart fluent-bit
service:
name: td-agent-bit
state: restarted
enabled: yes
tasks:
- name: Add fluent-bit repo
yum_repository:
name: TD Agent Bit
file: td-agent-bit
baseurl: http://packages.fluentbit.io/centos/7
enabled: true
gpgkey: http://packages.fluentbit.io/fluentbit.key
gpgcheck: true
- name: Install latest version of fluent-bit
yum:
name: td-agent-bit
state: latest
- name: Ensure fluent-bit is running
systemd:
name: td-agent-bit
enabled: true
state: started
- name: Deploy fluent-bit config
template:
src: "templates/{{ item }}.j2"
dest: "/etc/td-agent-bit/{{ item }}"
owner: root
group: root
mode: 0644
validate: "/opt/td-agent-bit/bin/td-agent-bit --check"
loop:
- td-agent-bit.conf
- output-gelf.conf
notify: Restart fluent-bit
```
|
process
|
add option to validate configuration is your feature request related to a problem please describe it seems fluent bit is missing the capability to only test its config file and the files it recursively includes for valid syntax at the moment when i deploy a configuration i have to restart fluent bit and hope that everything works and if there is an issue with the config fluent bit could halt with the risk of losing event records describe the solution you d like in order to prevent unwanted downtimes i d like to use the fluent bit binary to validate the configuration after an update be it via config management or manually the service should only be attempted to be restarted by my orchestration of choice if the config check is considered successful many applications allow you to check if their configuration contains errors e g nginx c etc nginx nginx conf t apachectl configtest describe alternatives you ve considered the only issue that s somewhat similar is but does not offer a solution additional context os centos edition community edition version here s an example ansible playbook yml hosts loadbalancers become yes handlers name restart fluent bit service name td agent bit state restarted enabled yes tasks name add fluent bit repo yum repository name td agent bit file td agent bit baseurl enabled true gpgkey gpgcheck true name install latest version of fluent bit yum name td agent bit state latest name ensure fluent bit is running systemd name td agent bit enabled true state started name deploy fluent bit config template src templates item dest etc td agent bit item owner root group root mode validate opt td agent bit bin td agent bit check loop td agent bit conf output gelf conf notify restart fluent bit
| 1
|
365,146
| 10,776,160,450
|
IssuesEvent
|
2019-11-03 18:57:59
|
Hydractify/kanna_kobayashi
|
https://api.github.com/repos/Hydractify/kanna_kobayashi
|
opened
|
[BUG] - Being able to use the `cry` command on users who have blocked you
|
bug priority
|
Currently there is a bug where you can use interactive commands on other users, such as the `cry` and `dance` command. That is caused because they have their own `parseArgs` function.
|
1.0
|
[BUG] - Being able to use the `cry` command on users who have blocked you - Currently there is a bug where you can use interactive commands on other users, such as the `cry` and `dance` command. That is caused because they have their own `parseArgs` function.
|
non_process
|
being able to use the cry command on users who have blocked you currently there is a bug where you can use interactive commands on other users such as the cry and dance command that is caused because they have their own parseargs function
| 0
|
304
| 2,736,197,444
|
IssuesEvent
|
2015-04-19 06:26:00
|
sysown/proxysql-0.2
|
https://api.github.com/repos/sysown/proxysql-0.2
|
opened
|
Make query parsing optional , add variable mysql-commands_stats
|
ADMIN development MYSQL PROTOCOL QUERY PROCESSOR
|
Query parsing was a feature introduced some time ago in issue #168 .
It introduced a very small overhead, but should be configurable in order to completely reduce such overhead.
Suggested variable name: mysql-commands_stats
|
1.0
|
Make query parsing optional , add variable mysql-commands_stats - Query parsing was a feature introduced some time ago in issue #168 .
It introduced a very small overhead, but should be configurable in order to completely reduce such overhead.
Suggested variable name: mysql-commands_stats
|
process
|
make query parsing optional add variable mysql commands stats query parsing was a feature introduced some time ago in issue it introduced a very small overhead but should be configurable in order to completely reduce such overhead suggested variable name mysql commands stats
| 1
|
437,324
| 12,576,778,827
|
IssuesEvent
|
2020-06-09 08:28:36
|
pantheracorp/PantheraIDS_Issues
|
https://api.github.com/repos/pantheracorp/PantheraIDS_Issues
|
closed
|
Data Analyses / SCR / Identifications - Identifications csv still needed
|
priority: HIGH
|
Just wanted to bring attention to the fact that the Identifications csv is still required for the analysis if working from a local dtbs. This file needs to be created manually (Southern Africa hub has been using a R script originally created by Jo for this purpose). From what I understand there was a module within IDS that created this csv, however it was removed.
I believe that asking the user to create this csv manually is clunky and prone to user-error. Perhaps there is a solution that doesn't require this csv? If not, is there a way to automate this (I guess put the old module back into IDS somewhere).
I can provide the R script that we have been using to manually create this csv, if that will help.
|
1.0
|
Data Analyses / SCR / Identifications - Identifications csv still needed - Just wanted to bring attention to the fact that the Identifications csv is still required for the analysis if working from a local dtbs. This file needs to be created manually (Southern Africa hub has been using a R script originally created by Jo for this purpose). From what I understand there was a module within IDS that created this csv, however it was removed.
I believe that asking the user to create this csv manually is clunky and prone to user-error. Perhaps there is a solution that doesn't require this csv? If not, is there a way to automate this (I guess put the old module back into IDS somewhere).
I can provide the R script that we have been using to manually create this csv, if that will help.
|
non_process
|
data analyses scr identifications identifications csv still needed just wanted to bring attention to the fact that the identifications csv is still required for the analysis if working from a local dtbs this file needs to be created manually southern africa hub has been using a r script originally created by jo for this purpose from what i understand there was a module within ids that created this csv however it was removed i believe that asking the user to create this csv manually is clunky and prone to user error perhaps there is a solution that doesn t require this csv if not is there a way to automate this i guess put the old module back into ids somewhere i can provide the r script that we have been using to manually create this csv if that will help
| 0
|
18,495
| 24,550,996,829
|
IssuesEvent
|
2022-10-12 12:36:26
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] [Offline indicator] 'You are offline' error message is not getting displayed when participant clicks on the Notifications
|
Bug P1 iOS Process: Fixed Process: Tested QA Process: Tested dev
|
AR: Blank screen is getting displayed and the 'You are offline error message is not getting displayed when the participant clicks on the Notifications if a participant is offline
ER: You are offline error message should get displayed

|
3.0
|
[iOS] [Offline indicator] 'You are offline' error message is not getting displayed when participant clicks on the Notifications - AR: Blank screen is getting displayed and the 'You are offline error message is not getting displayed when the participant clicks on the Notifications if a participant is offline
ER: You are offline error message should get displayed

|
process
|
you are offline error message is not getting displayed when participant clicks on the notifications ar blank screen is getting displayed and the you are offline error message is not getting displayed when the participant clicks on the notifications if a participant is offline er you are offline error message should get displayed
| 1
|
194,801
| 15,439,363,837
|
IssuesEvent
|
2021-03-07 23:59:48
|
lussierc/StockSwingPredictor
|
https://api.github.com/repos/lussierc/StockSwingPredictor
|
opened
|
Enhance Project Accessibility
|
documentation feature
|
The tool and project as a whole should be accessible and easy to interact with. The README should be detailed with what the project is and what it can do. It should also provide detailed instructions on how to use the tool.
With this, there should be quite a few ways to use the tool:
- Python (locally): install needed packages via pip on your own machine
- Docker
- Pipenv
- Poetry
All of methods to run the tool should be considered or included when the project is made more accessible.
|
1.0
|
Enhance Project Accessibility - The tool and project as a whole should be accessible and easy to interact with. The README should be detailed with what the project is and what it can do. It should also provide detailed instructions on how to use the tool.
With this, there should be quite a few ways to use the tool:
- Python (locally): install needed packages via pip on your own machine
- Docker
- Pipenv
- Poetry
All of methods to run the tool should be considered or included when the project is made more accessible.
|
non_process
|
enhance project accessibility the tool and project as a whole should be accessible and easy to interact with the readme should be detailed with what the project is and what it can do it should also provide detailed instructions on how to use the tool with this there should be quite a few ways to use the tool python locally install needed packages via pip on your own machine docker pipenv poetry all of methods to run the tool should be considered or included when the project is made more accessible
| 0
|
58,292
| 24,399,916,455
|
IssuesEvent
|
2022-10-04 23:43:16
|
microsoft/botbuilder-dotnet
|
https://api.github.com/repos/microsoft/botbuilder-dotnet
|
closed
|
TimerCallback function not getting hit in deployed environment C#
|
bug customer-reported Bot Services customer-replied-to
|
Microsoft.Bot.Builder Version = 4.15.1
I have a Microsoft Chatbot C# code,which has a TimerCallback function, which runs after some inactive time by user. This function gets hit when running code locally, but the same function does not get hit, when deployed on Azure environment.
Appologies for too much logging lines in the code, it was just to verify that function gets hit on deployed environment or not.
Here is the whole code:
```csharp
using System;
using System.Collections.Concurrent;
using System.Collections.Generic;
using System.Threading;
using System.Threading.Tasks;
using Lebara.Crm.Bot.Core.Data;
using Lebara.Crm.Bot.Core.ServiceContracts;
using Microsoft.Bot.Builder;
using Microsoft.Bot.Schema;
using Microsoft.Extensions.Caching.Distributed;
using Microsoft.Extensions.Configuration;
using Microsoft.Extensions.Options;
using Newtonsoft.Json;
using Microsoft.Extensions.DependencyInjection;
using Microsoft.Bot.Builder.Integration.AspNet.Core;
using Microsoft.AspNetCore.Mvc;
using Microsoft.ApplicationInsights;
using Microsoft.ApplicationInsights.DataContracts;
using Microsoft.Extensions.Logging;
namespace Lebara.Crm.Bot.Services
{
public class ChatSessionTimeoutMiddleware : ActivityHandler, IMiddleware
{
private readonly IDistributedCache _distributedCache;
private readonly string _cachekey;
private readonly BotOptions _botOptions;
private readonly ISystemMessageSender _systemMessageSender;
private readonly CustomConversationStateAccessors _customConversationStateAccessors;
private readonly ITelemetryManager _telemetryManager;
private readonly ISessionHandler _sessionHandler;
private readonly IServiceProvider _serviceProvider;
private static ConcurrentDictionary<string, Timer> _sessionTimers = new ConcurrentDictionary<string, Timer>();
private readonly IBotFrameworkHttpAdapter _adapter;
private readonly string _appId;
private readonly ConcurrentDictionary<string, ConversationReference> _conversationReferences;
private string ObjectKey;
private readonly TelemetryClient telemetry = new TelemetryClient();
private readonly ILogger<ChatSessionTimeoutMiddleware> _logger;
// Timer related variables, made global, to avoid collected by Garbage Collector as they were created in memory previously.
private Timer _warningTimer = null;
private Timer _warningTimerGCIssue = null;
private Timer _timer = null;
private Timer _timerGCIssue = null;
public ChatSessionTimeoutMiddleware(
IDistributedCache distributedCache,
IConfiguration configuration,
IOptions<BotOptions> options,
ISystemMessageSender systemMessageSender,
CustomConversationStateAccessors customConversationStateAccessors,
ITelemetryManager telemetryManager,
ISessionHandler sessionHandler,
IServiceProvider serviceProvider,
IBotFrameworkHttpAdapter adapter,
ConcurrentDictionary<string, ConversationReference> conversationReferences,
ILogger<ChatSessionTimeoutMiddleware> logger)
{
_cachekey = $"{configuration["RedisCachingRoot"]}session-timeout:";
_distributedCache = distributedCache;
_botOptions = options.Value;
_systemMessageSender = systemMessageSender;
_customConversationStateAccessors = customConversationStateAccessors;
_telemetryManager = telemetryManager;
_sessionHandler = sessionHandler;
_serviceProvider = serviceProvider;
_adapter = adapter;
_conversationReferences = conversationReferences;
_appId = configuration["MicrosoftAppId"] ?? string.Empty;
_logger = logger;
}
private void AddConversationReference(Activity activity)
{
var conversationReference = activity.GetConversationReference();
_conversationReferences.AddOrUpdate(conversationReference.User.Id, conversationReference, (key, newValue) => conversationReference);
}
public async Task OnTurnAsync(ITurnContext turnContext, NextDelegate nextTurn, CancellationToken cancellationToken)
{
_logger.LogDebug("chat session timeout middleware");
//telemetry.TrackEvent("ChatSessionMiddleware - OnTurnAsync");
//telemetry.TrackTrace("ChatSessionMiddleware - OnTurnAsync", SeverityLevel.Warning, null);
var customConversationState = await _customConversationStateAccessors.CustomConversationState.GetAsync(turnContext, () => new CustomConversationState());
var hasChatSessionRunning = await _sessionHandler.HasRunningSessionAsync(turnContext.Activity.Conversation.Id);
if (turnContext.Activity.Type == ActivityTypes.Message
&& !string.IsNullOrEmpty(turnContext.Activity.Text)
&& !hasChatSessionRunning)
{
_logger.LogDebug("chatsessiontimeout OnTurnAsync if statement");
var key = _cachekey + turnContext.Activity.Conversation.Id;
_logger.LogDebug($"Key {key}");
var warningKey = "warning_" + key;
_logger.LogDebug($"WarningKey {warningKey}");
var period = _botOptions.InactivityPeriod;
_logger.LogDebug($"chatsessiontimeout period {period}");
var warningPeriod = _botOptions.WarningInactivityPeriod;
_logger.LogDebug($"chatsessiontimeout warningPeriod {warningPeriod}");
var cacheOptions = new DistributedCacheEntryOptions
{
AbsoluteExpirationRelativeToNow = period.Add(period)
};
_logger.LogDebug($"cacheOptions {cacheOptions}");
AddConversationReference(turnContext.Activity as Activity);
await _distributedCache.SetStringAsync(key, JsonConvert.SerializeObject(DateTime.Now.Add(period)), cacheOptions);
await _distributedCache.SetStringAsync(warningKey, JsonConvert.SerializeObject(DateTime.Now.Add(warningPeriod)), cacheOptions);
var timerPeriod = period.Add(TimeSpan.FromSeconds(1));
var warningTimePeriod = warningPeriod.Add(TimeSpan.FromSeconds(1));
_warningTimer = null;
_warningTimerGCIssue = null;
_sessionTimers.TryGetValue(warningKey, out _warningTimer);
_logger.LogDebug($"warningTimer {_warningTimer}");
if (_warningTimer == null)
{
_logger.LogDebug("warningTimer is null");
_warningTimerGCIssue = new Timer(new TimerCallback(WarningCallback), (warningKey, turnContext), warningTimePeriod, warningTimePeriod);
_warningTimer = _sessionTimers.GetOrAdd(warningKey, _warningTimerGCIssue);
}
_timer = null;
_timerGCIssue = null;
_sessionTimers.TryGetValue(key, out _timer);
_logger.LogDebug($"timer {_timer}");
if (_timer == null)
{
_logger.LogDebug("timer is null");
_logger.LogDebug($"key {key}");
_timerGCIssue = new Timer(new TimerCallback(Callback), (key, turnContext), timerPeriod, timerPeriod);
_timer = _sessionTimers.GetOrAdd(key, _timerGCIssue);
}
_warningTimer.Change(warningTimePeriod, warningTimePeriod);
_logger.LogDebug("chatSessionTimeoutMiddleware timer change");
_timer.Change(timerPeriod, timerPeriod);
}
_logger.LogDebug("chatSessionTimeoutMiddleware nextturn");
await nextTurn(cancellationToken).ConfigureAwait(false);
}
private async void Callback(object target)
{
//telemetry.TrackEvent("ChatSessionMiddleware - InactivityCallback");
_logger.LogDebug("ChatSessionMiddleware InactivityCallback");
var tuple = ((string, ITurnContext))target;
ObjectKey = tuple.Item1;
var turnContext = tuple.Item2;
foreach (var conversationReference in _conversationReferences.Values)
{
_logger.LogDebug("ChatSessionMiddleware InactivityCallback for loop");
await ((BotAdapter)_adapter).ContinueConversationAsync(_appId, conversationReference, EndOfChatCallback, default(CancellationToken));
}
}
private async void WarningCallback(object target)
{
//telemetry.TrackEvent("ChatSessionMiddleware - WarningCallback");
_logger.LogDebug("ChatSessionMiddleware WarningCallback");
var tuple = ((string, ITurnContext))target;
ObjectKey = tuple.Item1;
var turnContext = tuple.Item2;
foreach (var conversationReference in _conversationReferences.Values)
{
_logger.LogDebug("ChatSessionMiddleware WarningCallback for loop");
await ((BotAdapter)_adapter).ContinueConversationAsync(_appId, conversationReference, WarningMessageCallback, default(CancellationToken));
}
}
private async Task WarningMessageCallback(ITurnContext turnContext, CancellationToken cancellationToken)
{
//telemetry.TrackEvent("ChatSessionMiddleware - WarningMessageCallback");
_logger.LogDebug("ChatSessionMiddleware WarningMessageCallback");
var customConversationState = await _customConversationStateAccessors.CustomConversationState.GetAsync(turnContext, () => new CustomConversationState());
void DisposeTimer()
{
bool found = _sessionTimers.TryRemove(ObjectKey, out var timer);
if (found)
{
timer.Dispose();
timer = null;
}
}
var json = await _distributedCache.GetStringAsync(ObjectKey);
var hasChatSessionRunning = await _sessionHandler.HasRunningSessionAsync(turnContext.Activity.Conversation.Id);
if (hasChatSessionRunning)
{
DisposeTimer();
return;
}
if (!string.IsNullOrEmpty(json))
{
var sessionEnd = JsonConvert.DeserializeObject<DateTime>(json);
if (DateTime.Now >= sessionEnd)
{
//telemetry.TrackEvent("ChatSessionMiddleware - SendingWarningMessage");
_logger.LogDebug("ChatSessionMiddleware SendingWarningMessage");
await _systemMessageSender.SendSystemMessage(turnContext, customConversationState, turnContext.Activity, ResourceIds.BotWarningEndOfChat);
}
}
DisposeTimer();
}
private async Task EndOfChatCallback(ITurnContext turnContext, CancellationToken cancellationToken)
{
//telemetry.TrackEvent("ChatSessionMiddleware - EndOfChatCallback");
_logger.LogDebug("ChatSessionMiddleware EndOfChatCallback");
var chatSdk = (IChatProvider)_serviceProvider.GetService(typeof(IChatProvider));
var customConversationState = await _customConversationStateAccessors.CustomConversationState.GetAsync(turnContext, () => new CustomConversationState());
void DisposeTimer()
{
bool found = _sessionTimers.TryRemove(ObjectKey, out var timer);
if (found)
{
timer.Dispose();
timer = null;
}
}
var json = await _distributedCache.GetStringAsync(ObjectKey);
var hasChatSessionRunning = await _sessionHandler.HasRunningSessionAsync(turnContext.Activity.Conversation.Id);
if (hasChatSessionRunning)
{
DisposeTimer();
return;
}
if (!string.IsNullOrEmpty(json))
{
var sessionEnd = JsonConvert.DeserializeObject<DateTime>(json);
if (DateTime.Now >= sessionEnd)
{
var parts = ObjectKey.Split(new char[] { ':' });
var dict = new Dictionary<string, string>
{
{"EndTime", json },
{"State", JsonConvert.SerializeObject(customConversationState) }
};
_telemetryManager.TrackEvent("AutomaticChatClosing", parts[parts.Length - 1], dict);
DisposeTimer();
//telemetry.TrackEvent("ChatSessionMiddleware - SendingEndOfChatMessage");
_logger.LogDebug("ChatSessionMiddleware SendingEndOfChatMessage");
await _systemMessageSender.SendSystemMessage(turnContext, customConversationState, turnContext.Activity, ResourceIds.BotAutomaticEndOfChat);
await Task.Delay(2000);
await chatSdk.EndChat(customConversationState.ChatContext, turnContext);
}
}
else
{
DisposeTimer();
}
}
}
}
```
|
1.0
|
TimerCallback function not getting hit in deployed environment C# - Microsoft.Bot.Builder Version = 4.15.1
I have a Microsoft Chatbot C# code,which has a TimerCallback function, which runs after some inactive time by user. This function gets hit when running code locally, but the same function does not get hit, when deployed on Azure environment.
Appologies for too much logging lines in the code, it was just to verify that function gets hit on deployed environment or not.
Here is the whole code:
```csharp
using System;
using System.Collections.Concurrent;
using System.Collections.Generic;
using System.Threading;
using System.Threading.Tasks;
using Lebara.Crm.Bot.Core.Data;
using Lebara.Crm.Bot.Core.ServiceContracts;
using Microsoft.Bot.Builder;
using Microsoft.Bot.Schema;
using Microsoft.Extensions.Caching.Distributed;
using Microsoft.Extensions.Configuration;
using Microsoft.Extensions.Options;
using Newtonsoft.Json;
using Microsoft.Extensions.DependencyInjection;
using Microsoft.Bot.Builder.Integration.AspNet.Core;
using Microsoft.AspNetCore.Mvc;
using Microsoft.ApplicationInsights;
using Microsoft.ApplicationInsights.DataContracts;
using Microsoft.Extensions.Logging;
namespace Lebara.Crm.Bot.Services
{
public class ChatSessionTimeoutMiddleware : ActivityHandler, IMiddleware
{
private readonly IDistributedCache _distributedCache;
private readonly string _cachekey;
private readonly BotOptions _botOptions;
private readonly ISystemMessageSender _systemMessageSender;
private readonly CustomConversationStateAccessors _customConversationStateAccessors;
private readonly ITelemetryManager _telemetryManager;
private readonly ISessionHandler _sessionHandler;
private readonly IServiceProvider _serviceProvider;
private static ConcurrentDictionary<string, Timer> _sessionTimers = new ConcurrentDictionary<string, Timer>();
private readonly IBotFrameworkHttpAdapter _adapter;
private readonly string _appId;
private readonly ConcurrentDictionary<string, ConversationReference> _conversationReferences;
private string ObjectKey;
private readonly TelemetryClient telemetry = new TelemetryClient();
private readonly ILogger<ChatSessionTimeoutMiddleware> _logger;
// Timer related variables, made global, to avoid collected by Garbage Collector as they were created in memory previously.
private Timer _warningTimer = null;
private Timer _warningTimerGCIssue = null;
private Timer _timer = null;
private Timer _timerGCIssue = null;
public ChatSessionTimeoutMiddleware(
IDistributedCache distributedCache,
IConfiguration configuration,
IOptions<BotOptions> options,
ISystemMessageSender systemMessageSender,
CustomConversationStateAccessors customConversationStateAccessors,
ITelemetryManager telemetryManager,
ISessionHandler sessionHandler,
IServiceProvider serviceProvider,
IBotFrameworkHttpAdapter adapter,
ConcurrentDictionary<string, ConversationReference> conversationReferences,
ILogger<ChatSessionTimeoutMiddleware> logger)
{
_cachekey = $"{configuration["RedisCachingRoot"]}session-timeout:";
_distributedCache = distributedCache;
_botOptions = options.Value;
_systemMessageSender = systemMessageSender;
_customConversationStateAccessors = customConversationStateAccessors;
_telemetryManager = telemetryManager;
_sessionHandler = sessionHandler;
_serviceProvider = serviceProvider;
_adapter = adapter;
_conversationReferences = conversationReferences;
_appId = configuration["MicrosoftAppId"] ?? string.Empty;
_logger = logger;
}
private void AddConversationReference(Activity activity)
{
var conversationReference = activity.GetConversationReference();
_conversationReferences.AddOrUpdate(conversationReference.User.Id, conversationReference, (key, newValue) => conversationReference);
}
public async Task OnTurnAsync(ITurnContext turnContext, NextDelegate nextTurn, CancellationToken cancellationToken)
{
_logger.LogDebug("chat session timeout middleware");
//telemetry.TrackEvent("ChatSessionMiddleware - OnTurnAsync");
//telemetry.TrackTrace("ChatSessionMiddleware - OnTurnAsync", SeverityLevel.Warning, null);
var customConversationState = await _customConversationStateAccessors.CustomConversationState.GetAsync(turnContext, () => new CustomConversationState());
var hasChatSessionRunning = await _sessionHandler.HasRunningSessionAsync(turnContext.Activity.Conversation.Id);
if (turnContext.Activity.Type == ActivityTypes.Message
&& !string.IsNullOrEmpty(turnContext.Activity.Text)
&& !hasChatSessionRunning)
{
_logger.LogDebug("chatsessiontimeout OnTurnAsync if statement");
var key = _cachekey + turnContext.Activity.Conversation.Id;
_logger.LogDebug($"Key {key}");
var warningKey = "warning_" + key;
_logger.LogDebug($"WarningKey {warningKey}");
var period = _botOptions.InactivityPeriod;
_logger.LogDebug($"chatsessiontimeout period {period}");
var warningPeriod = _botOptions.WarningInactivityPeriod;
_logger.LogDebug($"chatsessiontimeout warningPeriod {warningPeriod}");
var cacheOptions = new DistributedCacheEntryOptions
{
AbsoluteExpirationRelativeToNow = period.Add(period)
};
_logger.LogDebug($"cacheOptions {cacheOptions}");
AddConversationReference(turnContext.Activity as Activity);
await _distributedCache.SetStringAsync(key, JsonConvert.SerializeObject(DateTime.Now.Add(period)), cacheOptions);
await _distributedCache.SetStringAsync(warningKey, JsonConvert.SerializeObject(DateTime.Now.Add(warningPeriod)), cacheOptions);
var timerPeriod = period.Add(TimeSpan.FromSeconds(1));
var warningTimePeriod = warningPeriod.Add(TimeSpan.FromSeconds(1));
_warningTimer = null;
_warningTimerGCIssue = null;
_sessionTimers.TryGetValue(warningKey, out _warningTimer);
_logger.LogDebug($"warningTimer {_warningTimer}");
if (_warningTimer == null)
{
_logger.LogDebug("warningTimer is null");
_warningTimerGCIssue = new Timer(new TimerCallback(WarningCallback), (warningKey, turnContext), warningTimePeriod, warningTimePeriod);
_warningTimer = _sessionTimers.GetOrAdd(warningKey, _warningTimerGCIssue);
}
_timer = null;
_timerGCIssue = null;
_sessionTimers.TryGetValue(key, out _timer);
_logger.LogDebug($"timer {_timer}");
if (_timer == null)
{
_logger.LogDebug("timer is null");
_logger.LogDebug($"key {key}");
_timerGCIssue = new Timer(new TimerCallback(Callback), (key, turnContext), timerPeriod, timerPeriod);
_timer = _sessionTimers.GetOrAdd(key, _timerGCIssue);
}
_warningTimer.Change(warningTimePeriod, warningTimePeriod);
_logger.LogDebug("chatSessionTimeoutMiddleware timer change");
_timer.Change(timerPeriod, timerPeriod);
}
_logger.LogDebug("chatSessionTimeoutMiddleware nextturn");
await nextTurn(cancellationToken).ConfigureAwait(false);
}
private async void Callback(object target)
{
//telemetry.TrackEvent("ChatSessionMiddleware - InactivityCallback");
_logger.LogDebug("ChatSessionMiddleware InactivityCallback");
var tuple = ((string, ITurnContext))target;
ObjectKey = tuple.Item1;
var turnContext = tuple.Item2;
foreach (var conversationReference in _conversationReferences.Values)
{
_logger.LogDebug("ChatSessionMiddleware InactivityCallback for loop");
await ((BotAdapter)_adapter).ContinueConversationAsync(_appId, conversationReference, EndOfChatCallback, default(CancellationToken));
}
}
private async void WarningCallback(object target)
{
//telemetry.TrackEvent("ChatSessionMiddleware - WarningCallback");
_logger.LogDebug("ChatSessionMiddleware WarningCallback");
var tuple = ((string, ITurnContext))target;
ObjectKey = tuple.Item1;
var turnContext = tuple.Item2;
foreach (var conversationReference in _conversationReferences.Values)
{
_logger.LogDebug("ChatSessionMiddleware WarningCallback for loop");
await ((BotAdapter)_adapter).ContinueConversationAsync(_appId, conversationReference, WarningMessageCallback, default(CancellationToken));
}
}
private async Task WarningMessageCallback(ITurnContext turnContext, CancellationToken cancellationToken)
{
//telemetry.TrackEvent("ChatSessionMiddleware - WarningMessageCallback");
_logger.LogDebug("ChatSessionMiddleware WarningMessageCallback");
var customConversationState = await _customConversationStateAccessors.CustomConversationState.GetAsync(turnContext, () => new CustomConversationState());
void DisposeTimer()
{
bool found = _sessionTimers.TryRemove(ObjectKey, out var timer);
if (found)
{
timer.Dispose();
timer = null;
}
}
var json = await _distributedCache.GetStringAsync(ObjectKey);
var hasChatSessionRunning = await _sessionHandler.HasRunningSessionAsync(turnContext.Activity.Conversation.Id);
if (hasChatSessionRunning)
{
DisposeTimer();
return;
}
if (!string.IsNullOrEmpty(json))
{
var sessionEnd = JsonConvert.DeserializeObject<DateTime>(json);
if (DateTime.Now >= sessionEnd)
{
//telemetry.TrackEvent("ChatSessionMiddleware - SendingWarningMessage");
_logger.LogDebug("ChatSessionMiddleware SendingWarningMessage");
await _systemMessageSender.SendSystemMessage(turnContext, customConversationState, turnContext.Activity, ResourceIds.BotWarningEndOfChat);
}
}
DisposeTimer();
}
private async Task EndOfChatCallback(ITurnContext turnContext, CancellationToken cancellationToken)
{
//telemetry.TrackEvent("ChatSessionMiddleware - EndOfChatCallback");
_logger.LogDebug("ChatSessionMiddleware EndOfChatCallback");
var chatSdk = (IChatProvider)_serviceProvider.GetService(typeof(IChatProvider));
var customConversationState = await _customConversationStateAccessors.CustomConversationState.GetAsync(turnContext, () => new CustomConversationState());
void DisposeTimer()
{
bool found = _sessionTimers.TryRemove(ObjectKey, out var timer);
if (found)
{
timer.Dispose();
timer = null;
}
}
var json = await _distributedCache.GetStringAsync(ObjectKey);
var hasChatSessionRunning = await _sessionHandler.HasRunningSessionAsync(turnContext.Activity.Conversation.Id);
if (hasChatSessionRunning)
{
DisposeTimer();
return;
}
if (!string.IsNullOrEmpty(json))
{
var sessionEnd = JsonConvert.DeserializeObject<DateTime>(json);
if (DateTime.Now >= sessionEnd)
{
var parts = ObjectKey.Split(new char[] { ':' });
var dict = new Dictionary<string, string>
{
{"EndTime", json },
{"State", JsonConvert.SerializeObject(customConversationState) }
};
_telemetryManager.TrackEvent("AutomaticChatClosing", parts[parts.Length - 1], dict);
DisposeTimer();
//telemetry.TrackEvent("ChatSessionMiddleware - SendingEndOfChatMessage");
_logger.LogDebug("ChatSessionMiddleware SendingEndOfChatMessage");
await _systemMessageSender.SendSystemMessage(turnContext, customConversationState, turnContext.Activity, ResourceIds.BotAutomaticEndOfChat);
await Task.Delay(2000);
await chatSdk.EndChat(customConversationState.ChatContext, turnContext);
}
}
else
{
DisposeTimer();
}
}
}
}
```
|
non_process
|
timercallback function not getting hit in deployed environment c microsoft bot builder version i have a microsoft chatbot c code which has a timercallback function which runs after some inactive time by user this function gets hit when running code locally but the same function does not get hit when deployed on azure environment appologies for too much logging lines in the code it was just to verify that function gets hit on deployed environment or not here is the whole code csharp using system using system collections concurrent using system collections generic using system threading using system threading tasks using lebara crm bot core data using lebara crm bot core servicecontracts using microsoft bot builder using microsoft bot schema using microsoft extensions caching distributed using microsoft extensions configuration using microsoft extensions options using newtonsoft json using microsoft extensions dependencyinjection using microsoft bot builder integration aspnet core using microsoft aspnetcore mvc using microsoft applicationinsights using microsoft applicationinsights datacontracts using microsoft extensions logging namespace lebara crm bot services public class chatsessiontimeoutmiddleware activityhandler imiddleware private readonly idistributedcache distributedcache private readonly string cachekey private readonly botoptions botoptions private readonly isystemmessagesender systemmessagesender private readonly customconversationstateaccessors customconversationstateaccessors private readonly itelemetrymanager telemetrymanager private readonly isessionhandler sessionhandler private readonly iserviceprovider serviceprovider private static concurrentdictionary sessiontimers new concurrentdictionary private readonly ibotframeworkhttpadapter adapter private readonly string appid private readonly concurrentdictionary conversationreferences private string objectkey private readonly telemetryclient telemetry new telemetryclient private readonly ilogger logger timer related variables made global to avoid collected by garbage collector as they were created in memory previously private timer warningtimer null private timer warningtimergcissue null private timer timer null private timer timergcissue null public chatsessiontimeoutmiddleware idistributedcache distributedcache iconfiguration configuration ioptions options isystemmessagesender systemmessagesender customconversationstateaccessors customconversationstateaccessors itelemetrymanager telemetrymanager isessionhandler sessionhandler iserviceprovider serviceprovider ibotframeworkhttpadapter adapter concurrentdictionary conversationreferences ilogger logger cachekey configuration session timeout distributedcache distributedcache botoptions options value systemmessagesender systemmessagesender customconversationstateaccessors customconversationstateaccessors telemetrymanager telemetrymanager sessionhandler sessionhandler serviceprovider serviceprovider adapter adapter conversationreferences conversationreferences appid configuration string empty logger logger private void addconversationreference activity activity var conversationreference activity getconversationreference conversationreferences addorupdate conversationreference user id conversationreference key newvalue conversationreference public async task onturnasync iturncontext turncontext nextdelegate nextturn cancellationtoken cancellationtoken logger logdebug chat session timeout middleware telemetry trackevent chatsessionmiddleware onturnasync telemetry tracktrace chatsessionmiddleware onturnasync severitylevel warning null var customconversationstate await customconversationstateaccessors customconversationstate getasync turncontext new customconversationstate var haschatsessionrunning await sessionhandler hasrunningsessionasync turncontext activity conversation id if turncontext activity type activitytypes message string isnullorempty turncontext activity text haschatsessionrunning logger logdebug chatsessiontimeout onturnasync if statement var key cachekey turncontext activity conversation id logger logdebug key key var warningkey warning key logger logdebug warningkey warningkey var period botoptions inactivityperiod logger logdebug chatsessiontimeout period period var warningperiod botoptions warninginactivityperiod logger logdebug chatsessiontimeout warningperiod warningperiod var cacheoptions new distributedcacheentryoptions absoluteexpirationrelativetonow period add period logger logdebug cacheoptions cacheoptions addconversationreference turncontext activity as activity await distributedcache setstringasync key jsonconvert serializeobject datetime now add period cacheoptions await distributedcache setstringasync warningkey jsonconvert serializeobject datetime now add warningperiod cacheoptions var timerperiod period add timespan fromseconds var warningtimeperiod warningperiod add timespan fromseconds warningtimer null warningtimergcissue null sessiontimers trygetvalue warningkey out warningtimer logger logdebug warningtimer warningtimer if warningtimer null logger logdebug warningtimer is null warningtimergcissue new timer new timercallback warningcallback warningkey turncontext warningtimeperiod warningtimeperiod warningtimer sessiontimers getoradd warningkey warningtimergcissue timer null timergcissue null sessiontimers trygetvalue key out timer logger logdebug timer timer if timer null logger logdebug timer is null logger logdebug key key timergcissue new timer new timercallback callback key turncontext timerperiod timerperiod timer sessiontimers getoradd key timergcissue warningtimer change warningtimeperiod warningtimeperiod logger logdebug chatsessiontimeoutmiddleware timer change timer change timerperiod timerperiod logger logdebug chatsessiontimeoutmiddleware nextturn await nextturn cancellationtoken configureawait false private async void callback object target telemetry trackevent chatsessionmiddleware inactivitycallback logger logdebug chatsessionmiddleware inactivitycallback var tuple string iturncontext target objectkey tuple var turncontext tuple foreach var conversationreference in conversationreferences values logger logdebug chatsessionmiddleware inactivitycallback for loop await botadapter adapter continueconversationasync appid conversationreference endofchatcallback default cancellationtoken private async void warningcallback object target telemetry trackevent chatsessionmiddleware warningcallback logger logdebug chatsessionmiddleware warningcallback var tuple string iturncontext target objectkey tuple var turncontext tuple foreach var conversationreference in conversationreferences values logger logdebug chatsessionmiddleware warningcallback for loop await botadapter adapter continueconversationasync appid conversationreference warningmessagecallback default cancellationtoken private async task warningmessagecallback iturncontext turncontext cancellationtoken cancellationtoken telemetry trackevent chatsessionmiddleware warningmessagecallback logger logdebug chatsessionmiddleware warningmessagecallback var customconversationstate await customconversationstateaccessors customconversationstate getasync turncontext new customconversationstate void disposetimer bool found sessiontimers tryremove objectkey out var timer if found timer dispose timer null var json await distributedcache getstringasync objectkey var haschatsessionrunning await sessionhandler hasrunningsessionasync turncontext activity conversation id if haschatsessionrunning disposetimer return if string isnullorempty json var sessionend jsonconvert deserializeobject json if datetime now sessionend telemetry trackevent chatsessionmiddleware sendingwarningmessage logger logdebug chatsessionmiddleware sendingwarningmessage await systemmessagesender sendsystemmessage turncontext customconversationstate turncontext activity resourceids botwarningendofchat disposetimer private async task endofchatcallback iturncontext turncontext cancellationtoken cancellationtoken telemetry trackevent chatsessionmiddleware endofchatcallback logger logdebug chatsessionmiddleware endofchatcallback var chatsdk ichatprovider serviceprovider getservice typeof ichatprovider var customconversationstate await customconversationstateaccessors customconversationstate getasync turncontext new customconversationstate void disposetimer bool found sessiontimers tryremove objectkey out var timer if found timer dispose timer null var json await distributedcache getstringasync objectkey var haschatsessionrunning await sessionhandler hasrunningsessionasync turncontext activity conversation id if haschatsessionrunning disposetimer return if string isnullorempty json var sessionend jsonconvert deserializeobject json if datetime now sessionend var parts objectkey split new char var dict new dictionary endtime json state jsonconvert serializeobject customconversationstate telemetrymanager trackevent automaticchatclosing parts dict disposetimer telemetry trackevent chatsessionmiddleware sendingendofchatmessage logger logdebug chatsessionmiddleware sendingendofchatmessage await systemmessagesender sendsystemmessage turncontext customconversationstate turncontext activity resourceids botautomaticendofchat await task delay await chatsdk endchat customconversationstate chatcontext turncontext else disposetimer
| 0
|
282,575
| 30,889,370,997
|
IssuesEvent
|
2023-08-04 02:37:26
|
maddyCode23/linux-4.1.15
|
https://api.github.com/repos/maddyCode23/linux-4.1.15
|
reopened
|
CVE-2017-1000410 (High) detected in linux-stable-rtv4.1.33
|
Mend: dependency security vulnerability
|
## CVE-2017-1000410 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel version 3.3-rc1 and later is affected by a vulnerability lies in the processing of incoming L2CAP commands - ConfigRequest, and ConfigResponse messages. This info leak is a result of uninitialized stack variables that may be returned to an attacker in their uninitialized state. By manipulating the code flows that precede the handling of these configuration messages, an attacker can also gain some control over which data will be held in the uninitialized stack variables. This can allow him to bypass KASLR, and stack canaries protection - as both pointers and stack canaries may be leaked in this manner. Combining this vulnerability (for example) with the previously disclosed RCE vulnerability in L2CAP configuration parsing (CVE-2017-1000251) may allow an attacker to exploit the RCE against kernels which were built with the above mitigations. These are the specifics of this vulnerability: In the function l2cap_parse_conf_rsp and in the function l2cap_parse_conf_req the following variable is declared without initialization: struct l2cap_conf_efs efs; In addition, when parsing input configuration parameters in both of these functions, the switch case for handling EFS elements may skip the memcpy call that will write to the efs variable: ... case L2CAP_CONF_EFS: if (olen == sizeof(efs)) memcpy(&efs, (void *)val, olen); ... The olen in the above if is attacker controlled, and regardless of that if, in both of these functions the efs variable would eventually be added to the outgoing configuration request that is being built: l2cap_add_conf_opt(&ptr, L2CAP_CONF_EFS, sizeof(efs), (unsigned long) &efs); So by sending a configuration request, or response, that contains an L2CAP_CONF_EFS element, but with an element length that is not sizeof(efs) - the memcpy to the uninitialized efs variable can be avoided, and the uninitialized variable would be returned to the attacker (16 bytes).
<p>Publish Date: 2017-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-1000410>CVE-2017-1000410</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000410">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000410</a></p>
<p>Release Date: 2017-12-07</p>
<p>Fix Resolution: v5.1-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-1000410 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-1000410 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel version 3.3-rc1 and later is affected by a vulnerability lies in the processing of incoming L2CAP commands - ConfigRequest, and ConfigResponse messages. This info leak is a result of uninitialized stack variables that may be returned to an attacker in their uninitialized state. By manipulating the code flows that precede the handling of these configuration messages, an attacker can also gain some control over which data will be held in the uninitialized stack variables. This can allow him to bypass KASLR, and stack canaries protection - as both pointers and stack canaries may be leaked in this manner. Combining this vulnerability (for example) with the previously disclosed RCE vulnerability in L2CAP configuration parsing (CVE-2017-1000251) may allow an attacker to exploit the RCE against kernels which were built with the above mitigations. These are the specifics of this vulnerability: In the function l2cap_parse_conf_rsp and in the function l2cap_parse_conf_req the following variable is declared without initialization: struct l2cap_conf_efs efs; In addition, when parsing input configuration parameters in both of these functions, the switch case for handling EFS elements may skip the memcpy call that will write to the efs variable: ... case L2CAP_CONF_EFS: if (olen == sizeof(efs)) memcpy(&efs, (void *)val, olen); ... The olen in the above if is attacker controlled, and regardless of that if, in both of these functions the efs variable would eventually be added to the outgoing configuration request that is being built: l2cap_add_conf_opt(&ptr, L2CAP_CONF_EFS, sizeof(efs), (unsigned long) &efs); So by sending a configuration request, or response, that contains an L2CAP_CONF_EFS element, but with an element length that is not sizeof(efs) - the memcpy to the uninitialized efs variable can be avoided, and the uninitialized variable would be returned to the attacker (16 bytes).
<p>Publish Date: 2017-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-1000410>CVE-2017-1000410</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000410">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000410</a></p>
<p>Release Date: 2017-12-07</p>
<p>Fix Resolution: v5.1-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net bluetooth core c net bluetooth core c vulnerability details the linux kernel version and later is affected by a vulnerability lies in the processing of incoming commands configrequest and configresponse messages this info leak is a result of uninitialized stack variables that may be returned to an attacker in their uninitialized state by manipulating the code flows that precede the handling of these configuration messages an attacker can also gain some control over which data will be held in the uninitialized stack variables this can allow him to bypass kaslr and stack canaries protection as both pointers and stack canaries may be leaked in this manner combining this vulnerability for example with the previously disclosed rce vulnerability in configuration parsing cve may allow an attacker to exploit the rce against kernels which were built with the above mitigations these are the specifics of this vulnerability in the function parse conf rsp and in the function parse conf req the following variable is declared without initialization struct conf efs efs in addition when parsing input configuration parameters in both of these functions the switch case for handling efs elements may skip the memcpy call that will write to the efs variable case conf efs if olen sizeof efs memcpy efs void val olen the olen in the above if is attacker controlled and regardless of that if in both of these functions the efs variable would eventually be added to the outgoing configuration request that is being built add conf opt ptr conf efs sizeof efs unsigned long efs so by sending a configuration request or response that contains an conf efs element but with an element length that is not sizeof efs the memcpy to the uninitialized efs variable can be avoided and the uninitialized variable would be returned to the attacker bytes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.