Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
759,916 | 26,618,489,710 | IssuesEvent | 2023-01-24 09:24:10 | rpm-software-management/dnf5 | https://api.github.com/repos/rpm-software-management/dnf5 | closed | Add highlevel API for usage of "install_specs" | Priority: HIGH | dnf4 (`install_specs()` method of the `Base` class) has a feature on API that enables usage of so-called "install_specs" for transaction operations. Install spec is a string that describes any artifact the user wants to include into transaction:
- rpm from repository
- local rpm file
- URL to rpm file
- group, environmental group
- module
New API method should
- accept install spec on the input
- be able to parse install spec and decide whether the artifact is a package, a group, or a module
- download required package from given URL (for downloading the https://github.com/rpm-software-management/dnf5/issues/203 functionality is required)
- de-duplicate command line packages, inserting the same rpm file to libsolv pool should be avoided (would behave like multiple independent packages and possibly cause troubles)
- all downloads for given transaction are performed once during goal resolution
Implementation of this feature will enable us to move parsing of packages specified on the command line from client apps (dnf5, dnf5daemon-client) to the common place in the libdnf5 library. | 1.0 | Add highlevel API for usage of "install_specs" - dnf4 (`install_specs()` method of the `Base` class) has a feature on API that enables usage of so-called "install_specs" for transaction operations. Install spec is a string that describes any artifact the user wants to include into transaction:
- rpm from repository
- local rpm file
- URL to rpm file
- group, environmental group
- module
New API method should
- accept install spec on the input
- be able to parse install spec and decide whether the artifact is a package, a group, or a module
- download required package from given URL (for downloading the https://github.com/rpm-software-management/dnf5/issues/203 functionality is required)
- de-duplicate command line packages, inserting the same rpm file to libsolv pool should be avoided (would behave like multiple independent packages and possibly cause troubles)
- all downloads for given transaction are performed once during goal resolution
Implementation of this feature will enable us to move parsing of packages specified on the command line from client apps (dnf5, dnf5daemon-client) to the common place in the libdnf5 library. | priority | add highlevel api for usage of install specs install specs method of the base class has a feature on api that enables usage of so called install specs for transaction operations install spec is a string that describes any artifact the user wants to include into transaction rpm from repository local rpm file url to rpm file group environmental group module new api method should accept install spec on the input be able to parse install spec and decide whether the artifact is a package a group or a module download required package from given url for downloading the functionality is required de duplicate command line packages inserting the same rpm file to libsolv pool should be avoided would behave like multiple independent packages and possibly cause troubles all downloads for given transaction are performed once during goal resolution implementation of this feature will enable us to move parsing of packages specified on the command line from client apps client to the common place in the library | 1 |
143,965 | 11,589,311,876 | IssuesEvent | 2020-02-24 01:30:57 | urapadmin/kiosk | https://api.github.com/repos/urapadmin/kiosk | closed | tag analysis has a stupid default value | C: before or during next season bug filemaker needs testing |
tag analysis - Find a way to reset the standard input in tag analysis (that's why Wheeler always pops up)
| 1.0 | tag analysis has a stupid default value -
tag analysis - Find a way to reset the standard input in tag analysis (that's why Wheeler always pops up)
| non_priority | tag analysis has a stupid default value tag analysis find a way to reset the standard input in tag analysis that s why wheeler always pops up | 0 |
608,634 | 18,844,814,113 | IssuesEvent | 2021-11-11 13:51:38 | weaveio/woll-forum | https://api.github.com/repos/weaveio/woll-forum | closed | Weekly average dose rate per Morning/Noon/ Afternoon (週別の服用率 - test) | Fixed Priority1 | Please create a following graph.
User "週別の服用率 - TEST" for this ticket.
<br>
<br>
Mockup:
<img width="799" alt="スクリーンショット 2021-10-25 12 25 17" src="https://user-images.githubusercontent.com/61481039/138630770-f199fcd8-a741-4622-ab79-784095fab2f4.png">
It will look similar to what we have now, But Monday/Tuesday.....instead of month
<img width="1178" alt="スクリーンショット 2021-10-25 12 35 21" src="https://user-images.githubusercontent.com/61481039/138631127-9ab7dece-cb19-4512-a873-323dd5017bb7.png">
| 1.0 | Weekly average dose rate per Morning/Noon/ Afternoon (週別の服用率 - test) - Please create a following graph.
User "週別の服用率 - TEST" for this ticket.
<br>
<br>
Mockup:
<img width="799" alt="スクリーンショット 2021-10-25 12 25 17" src="https://user-images.githubusercontent.com/61481039/138630770-f199fcd8-a741-4622-ab79-784095fab2f4.png">
It will look similar to what we have now, But Monday/Tuesday.....instead of month
<img width="1178" alt="スクリーンショット 2021-10-25 12 35 21" src="https://user-images.githubusercontent.com/61481039/138631127-9ab7dece-cb19-4512-a873-323dd5017bb7.png">
| priority | weekly average dose rate per morning noon afternoon 週別の服用率 test please create a following graph user 週別の服用率 test for this ticket mockup img width alt スクリーンショット src it will look similar to what we have now but monday tuesday instead of month img width alt スクリーンショット src | 1 |
39,118 | 2,851,134,727 | IssuesEvent | 2015-06-01 02:48:48 | afollestad/cabinet-issue-tracker | https://api.github.com/repos/afollestad/cabinet-issue-tracker | closed | Use new support design library's widgets | enhancement high priority in progress | https://developer.android.com/tools/support-library/index.html
`NavigationView`: navigation drawer
`FloatingActionButton`: Unfortunately doesn't have menu right now (I don't think), but we should keep an eye out for that.
`Snackbar`: Perhaps we can use this for error messages or just to replace the `Toast`s
`AppBarLayout`: Responds to scrolling (like how we already have implemented) | 1.0 | Use new support design library's widgets - https://developer.android.com/tools/support-library/index.html
`NavigationView`: navigation drawer
`FloatingActionButton`: Unfortunately doesn't have menu right now (I don't think), but we should keep an eye out for that.
`Snackbar`: Perhaps we can use this for error messages or just to replace the `Toast`s
`AppBarLayout`: Responds to scrolling (like how we already have implemented) | priority | use new support design library s widgets navigationview navigation drawer floatingactionbutton unfortunately doesn t have menu right now i don t think but we should keep an eye out for that snackbar perhaps we can use this for error messages or just to replace the toast s appbarlayout responds to scrolling like how we already have implemented | 1 |
4,897 | 4,707,272,746 | IssuesEvent | 2016-10-13 19:38:37 | zulu-zsh/zulu | https://api.github.com/repos/zulu-zsh/zulu | opened | Improve performance of upgrade command | enhancement performance | Currently the `upgrade` command is very slow, as it has to perform a fetch on each package repository before it can check for changes.
One possibility might be asynchronously fetching from each of the repositories on init, and keeping a log of the last fetch timestamp for each of them, then only performing the fetch again if the fetch is older than x amount of time (probably an hour should be enough).
Any ideas welcome. | True | Improve performance of upgrade command - Currently the `upgrade` command is very slow, as it has to perform a fetch on each package repository before it can check for changes.
One possibility might be asynchronously fetching from each of the repositories on init, and keeping a log of the last fetch timestamp for each of them, then only performing the fetch again if the fetch is older than x amount of time (probably an hour should be enough).
Any ideas welcome. | non_priority | improve performance of upgrade command currently the upgrade command is very slow as it has to perform a fetch on each package repository before it can check for changes one possibility might be asynchronously fetching from each of the repositories on init and keeping a log of the last fetch timestamp for each of them then only performing the fetch again if the fetch is older than x amount of time probably an hour should be enough any ideas welcome | 0 |
191,588 | 15,298,249,525 | IssuesEvent | 2021-02-24 09:29:17 | SAP/luigi | https://api.github.com/repos/SAP/luigi | closed | Blog post with Luigi updates | documentation | Write a Luigi blog post explaining web components, new Luigi features and releases | 1.0 | Blog post with Luigi updates - Write a Luigi blog post explaining web components, new Luigi features and releases | non_priority | blog post with luigi updates write a luigi blog post explaining web components new luigi features and releases | 0 |
810,648 | 30,252,527,329 | IssuesEvent | 2023-07-06 21:58:47 | apcountryman/picolibrary | https://api.github.com/repos/apcountryman/picolibrary | closed | Add socket flags to WIZnet W5500 IP network stack | priority-normal status-awaiting_review type-enhancement | Add socket flags to WIZnet W5500 IP network stack (`::picolibrary::WIZnet::W5500::IP::Network_Stack`).
- [x] The `Network_Stack` class should support the following new operations:
- [x] `void set_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Set a socket's flags
- [x] `void clear_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Clear a socket's flags
- [x] `auto socket_flags( Socket_ID socket_id ) const noexcept -> std::uint_fast8_t;`: Get a socket's flags
- [x] Documentation | 1.0 | Add socket flags to WIZnet W5500 IP network stack - Add socket flags to WIZnet W5500 IP network stack (`::picolibrary::WIZnet::W5500::IP::Network_Stack`).
- [x] The `Network_Stack` class should support the following new operations:
- [x] `void set_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Set a socket's flags
- [x] `void clear_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Clear a socket's flags
- [x] `auto socket_flags( Socket_ID socket_id ) const noexcept -> std::uint_fast8_t;`: Get a socket's flags
- [x] Documentation | priority | add socket flags to wiznet ip network stack add socket flags to wiznet ip network stack picolibrary wiznet ip network stack the network stack class should support the following new operations void set socket flags socket id socket id std uint t mask noexcept set a socket s flags void clear socket flags socket id socket id std uint t mask noexcept clear a socket s flags auto socket flags socket id socket id const noexcept std uint t get a socket s flags documentation | 1 |
381,295 | 26,445,547,194 | IssuesEvent | 2023-01-16 06:52:24 | Chillandchat/mobile | https://api.github.com/repos/Chillandchat/mobile | closed | Bugs list | bug documentation enhancement help wanted need review | ### Bug when creating a room
The error screen will be deployed when the user creates a new room. A error message will be also displayed in the console of the app, which read:
```
ERROR [TypeError: undefined is not an object (evaluating 'returnedMessages[returnedMessages.length - 1].user')]
```
#### Reproduction steps
1. Sign in
2. Tap plus icon
3. Tap create room
4. Enter details
5. Click create
### Styling bug
A bug with the styling in the ```Menu``` screen was found. Specifically, in the ```RoomList``` component where the latest message will overflow to the outside of the screen. This prevents the user from reading it and is not aesthetcally pleasing as a contained text.
#### Reproduction steps
1. Sign in
That's All! | 1.0 | Bugs list - ### Bug when creating a room
The error screen will be deployed when the user creates a new room. A error message will be also displayed in the console of the app, which read:
```
ERROR [TypeError: undefined is not an object (evaluating 'returnedMessages[returnedMessages.length - 1].user')]
```
#### Reproduction steps
1. Sign in
2. Tap plus icon
3. Tap create room
4. Enter details
5. Click create
### Styling bug
A bug with the styling in the ```Menu``` screen was found. Specifically, in the ```RoomList``` component where the latest message will overflow to the outside of the screen. This prevents the user from reading it and is not aesthetcally pleasing as a contained text.
#### Reproduction steps
1. Sign in
That's All! | non_priority | bugs list bug when creating a room the error screen will be deployed when the user creates a new room a error message will be also displayed in the console of the app which read error user reproduction steps sign in tap plus icon tap create room enter details click create styling bug a bug with the styling in the menu screen was found specifically in the roomlist component where the latest message will overflow to the outside of the screen this prevents the user from reading it and is not aesthetcally pleasing as a contained text reproduction steps sign in that s all | 0 |
373,689 | 11,047,375,230 | IssuesEvent | 2019-12-09 18:50:34 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | Cannot disable default storage class | area/storage kind/bug priority/important-longterm | **Steps to reproduce**:
1. Start minikube (from scratch)
2. Disable addon :
```
> minikube addons disable default-storageclass
✅ "default-storageclass" was successfully disabled
```
3. List storage classes:
```
> kubectl get storageclasses.storage.k8s.io
NAME PROVISIONER AGE
standard (default) k8s.io/minikube-hostpath 4m14s
```
**Bug**: `standard` storage class is still marked as _default_.
Actually, the storage class is then annotated with both `storageclass.kubernetes.io/is-default-class: "true"` (from the addon) and `storageclass.beta.kubernetes.io/is-default-class: "false"` (added with `minikube addons disable default-storageclass`):
```
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
annotations:
storageclass.beta.kubernetes.io/is-default-class: "false"
storageclass.kubernetes.io/is-default-class: "true"
creationTimestamp: "2019-11-20T12:40:29Z"
labels:
addonmanager.kubernetes.io/mode: EnsureExists
name: standard
resourceVersion: "600"
selfLink: /apis/storage.k8s.io/v1/storageclasses/standard
uid: cdef4691-5dc5-42c5-8e02-db45e4b4265c
provisioner: k8s.io/minikube-hostpath
reclaimPolicy: Delete
volumeBindingMode: Immediate
```
**Minikube version**: v1.5.2
**Kubernetes version**: v1.16.2 | 1.0 | Cannot disable default storage class - **Steps to reproduce**:
1. Start minikube (from scratch)
2. Disable addon :
```
> minikube addons disable default-storageclass
✅ "default-storageclass" was successfully disabled
```
3. List storage classes:
```
> kubectl get storageclasses.storage.k8s.io
NAME PROVISIONER AGE
standard (default) k8s.io/minikube-hostpath 4m14s
```
**Bug**: `standard` storage class is still marked as _default_.
Actually, the storage class is then annotated with both `storageclass.kubernetes.io/is-default-class: "true"` (from the addon) and `storageclass.beta.kubernetes.io/is-default-class: "false"` (added with `minikube addons disable default-storageclass`):
```
apiVersion: storage.k8s.io/v1
kind: StorageClass
metadata:
annotations:
storageclass.beta.kubernetes.io/is-default-class: "false"
storageclass.kubernetes.io/is-default-class: "true"
creationTimestamp: "2019-11-20T12:40:29Z"
labels:
addonmanager.kubernetes.io/mode: EnsureExists
name: standard
resourceVersion: "600"
selfLink: /apis/storage.k8s.io/v1/storageclasses/standard
uid: cdef4691-5dc5-42c5-8e02-db45e4b4265c
provisioner: k8s.io/minikube-hostpath
reclaimPolicy: Delete
volumeBindingMode: Immediate
```
**Minikube version**: v1.5.2
**Kubernetes version**: v1.16.2 | priority | cannot disable default storage class steps to reproduce start minikube from scratch disable addon minikube addons disable default storageclass ✅ default storageclass was successfully disabled list storage classes kubectl get storageclasses storage io name provisioner age standard default io minikube hostpath bug standard storage class is still marked as default actually the storage class is then annotated with both storageclass kubernetes io is default class true from the addon and storageclass beta kubernetes io is default class false added with minikube addons disable default storageclass apiversion storage io kind storageclass metadata annotations storageclass beta kubernetes io is default class false storageclass kubernetes io is default class true creationtimestamp labels addonmanager kubernetes io mode ensureexists name standard resourceversion selflink apis storage io storageclasses standard uid provisioner io minikube hostpath reclaimpolicy delete volumebindingmode immediate minikube version kubernetes version | 1 |
509,994 | 14,767,114,421 | IssuesEvent | 2021-01-10 04:39:28 | fossasia/open-event-frontend | https://api.github.com/repos/fossasia/open-event-frontend | opened | Not able to create a speaker due to Error "Field may not be null" | Priority: High Priority: Urgent bug | Trying to create a speaker on next months Codeheat event shows an error message "Field may not be null" even though I filled in all fields.
Compare
https://eventyay.com/events/f4fc71b8/speakers/create

| 2.0 | Not able to create a speaker due to Error "Field may not be null" - Trying to create a speaker on next months Codeheat event shows an error message "Field may not be null" even though I filled in all fields.
Compare
https://eventyay.com/events/f4fc71b8/speakers/create

| priority | not able to create a speaker due to error field may not be null trying to create a speaker on next months codeheat event shows an error message field may not be null even though i filled in all fields compare | 1 |
79,701 | 3,541,590,382 | IssuesEvent | 2016-01-19 02:13:18 | OctopusDeploy/Issues | https://api.github.com/repos/OctopusDeploy/Issues | closed | Provide ability to ignore free space check on deployment | enhancement forreview priority | Some users have reported issues when trying to deploy to SSH endpoints where either [mono is reporting incorrect space] (https://github.com/OctopusDeploy/Calamari/blob/94e883c84e8605a7b032492e23e813e369811733/source/Calamari/Integration/FileSystem/NixPhysicalFileSystem.cs#L14) or, being on a linux machine, they just don't have or need 500Mb free on their drive. Alternatively they may be deploying the package ultimately to a seperate drive with different free space anyway and so the check is really not serving its real purpose.
For this reason it has been suggested that we keep the current default, but provide a mechanism by way of special variables to ignore the space check and/or provide a different threshold to the currently hard-coded 500Mb [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L458) and [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L469).
Since this calculation takes place within calamari this is a platform independent change and therefore will be available to Windows users too. | 1.0 | Provide ability to ignore free space check on deployment - Some users have reported issues when trying to deploy to SSH endpoints where either [mono is reporting incorrect space] (https://github.com/OctopusDeploy/Calamari/blob/94e883c84e8605a7b032492e23e813e369811733/source/Calamari/Integration/FileSystem/NixPhysicalFileSystem.cs#L14) or, being on a linux machine, they just don't have or need 500Mb free on their drive. Alternatively they may be deploying the package ultimately to a seperate drive with different free space anyway and so the check is really not serving its real purpose.
For this reason it has been suggested that we keep the current default, but provide a mechanism by way of special variables to ignore the space check and/or provide a different threshold to the currently hard-coded 500Mb [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L458) and [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L469).
Since this calculation takes place within calamari this is a platform independent change and therefore will be available to Windows users too. | priority | provide ability to ignore free space check on deployment some users have reported issues when trying to deploy to ssh endpoints where either or being on a linux machine they just don t have or need free on their drive alternatively they may be deploying the package ultimately to a seperate drive with different free space anyway and so the check is really not serving its real purpose for this reason it has been suggested that we keep the current default but provide a mechanism by way of special variables to ignore the space check and or provide a different threshold to the currently hard coded and since this calculation takes place within calamari this is a platform independent change and therefore will be available to windows users too | 1 |
290,836 | 8,908,281,880 | IssuesEvent | 2019-01-18 00:06:33 | nluedtke/linux_kernel_cves | https://api.github.com/repos/nluedtke/linux_kernel_cves | closed | Upgrade Webpack | Web Frontend high priority in progress | Upgrade webpack-dev-server to version 3.1.11 or later to fix vulnerable dependency. | 1.0 | Upgrade Webpack - Upgrade webpack-dev-server to version 3.1.11 or later to fix vulnerable dependency. | priority | upgrade webpack upgrade webpack dev server to version or later to fix vulnerable dependency | 1 |
22,064 | 10,722,644,831 | IssuesEvent | 2019-10-27 13:32:32 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Vulnerability roundup 51: libtiff-4.0.9 | 1.severity: security | ## libtiff-4.0.9: 3 advisories
[search](https://search.nix.gsc.io/?q=libtiff&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libtiff+in%3Apath&type=Code)
- [ ] [CVE-2018-17000](https://nvd.nist.gov/vuln/detail/CVE-2018-17000) (unstable, 18.09, 18.03)
- [ ] [CVE-2018-17100](https://nvd.nist.gov/vuln/detail/CVE-2018-17100) (unstable, 18.09, 18.03)
- [ ] [CVE-2018-17101](https://nvd.nist.gov/vuln/detail/CVE-2018-17101) (unstable, 18.09, 18.03)
| True | Vulnerability roundup 51: libtiff-4.0.9 - ## libtiff-4.0.9: 3 advisories
[search](https://search.nix.gsc.io/?q=libtiff&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libtiff+in%3Apath&type=Code)
- [ ] [CVE-2018-17000](https://nvd.nist.gov/vuln/detail/CVE-2018-17000) (unstable, 18.09, 18.03)
- [ ] [CVE-2018-17100](https://nvd.nist.gov/vuln/detail/CVE-2018-17100) (unstable, 18.09, 18.03)
- [ ] [CVE-2018-17101](https://nvd.nist.gov/vuln/detail/CVE-2018-17101) (unstable, 18.09, 18.03)
| non_priority | vulnerability roundup libtiff libtiff advisories unstable unstable unstable | 0 |
544,463 | 15,893,814,660 | IssuesEvent | 2021-04-11 07:43:45 | googleapis/nodejs-proto-files | https://api.github.com/repos/googleapis/nodejs-proto-files | closed | Synthesis failed for nodejs-proto-files | autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate nodejs-proto-files. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the nodejs-proto-files API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
MPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/.gitattributes
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/lint.cfg
.kokoro/continuous/node12/samples-test.cfg
.kokoro/continuous/node12/system-test.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/populate-secrets.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/samples-test.cfg
.kokoro/presubmit/node12/system-test.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/publish.sh
.kokoro/release/docs-devsite.cfg
.kokoro/release/docs-devsite.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.kokoro/trampoline_v2.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
.trampolinerc
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
api-extractor.json
renovate.json
samples/README.md
2021-04-09 00:42:57,323 synthtool [DEBUG] > Installing dependencies...
DEBUG:synthtool:Installing dependencies...
npm WARN deprecated sinon@10.0.1: Breaking change found in this patch version
npm WARN deprecated ts-simple-ast@12.4.0: NOTICE: ts-simple-ast has been renamed to ts-morph and version reset to 1.0.0. Switch at your leisure...
npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated opn@6.0.0: The package has been renamed to `open`
npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated core-js@2.6.12: core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to the actual version of core-js@3.
npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated
> core-js@2.6.12 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/core-js
> node -e "try{require('./postinstall')}catch(e){}"
[96mThank you for using core-js ([94m https://github.com/zloirock/core-js [96m) for polyfilling JavaScript standard library![0m
[96mThe project needs your help! Please consider supporting of core-js on Open Collective or Patreon: [0m
[96m>[94m https://opencollective.com/core-js [0m
[96m>[94m https://www.patreon.com/zloirock [0m
[96mAlso, the author of core-js ([94m https://github.com/zloirock [96m) is looking for a good job -)[0m
> @compodoc/compodoc@1.1.11 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@compodoc/compodoc
> opencollective-postinstall || exit 0
[96m[1mThank you for using @compodoc/compodoc![96m[1m
[0m[96mIf you rely on this package, please consider supporting our open collective:[22m[39m
> [94mhttps://opencollective.com/compodoc/donate[0m
> protobufjs@6.10.2 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/protobufjs
> node scripts/postinstall
> google-proto-files@2.4.0 prepare /home/kbuilder/.cache/synthtool/nodejs-proto-files
> npm run compile && node ./build/tools/prepublish.js
> google-proto-files@2.4.0 precompile /home/kbuilder/.cache/synthtool/nodejs-proto-files
> gts clean
version: 14
Removing build ...
> google-proto-files@2.4.0 compile /home/kbuilder/.cache/synthtool/nodejs-proto-files
> tsc -p .
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m778[0m:[93m36[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'TimerId'.
[7m778[0m type SinonTimerId = FakeTimers.TimerId;
[7m [0m [91m ~~~~~~~[0m
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m780[0m:[93m39[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'InstalledMethods'.
[7m780[0m type SinonFakeTimers = FakeTimers.InstalledMethods &
[7m [0m [91m ~~~~~~~~~~~~~~~~[0m
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m781[0m:[93m20[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'NodeClock'.
[7m781[0m FakeTimers.NodeClock &
[7m [0m [91m ~~~~~~~~~[0m
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m782[0m:[93m20[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'BrowserClock'.
[7m782[0m FakeTimers.BrowserClock & {
[7m [0m [91m ~~~~~~~~~~~~[0m
Found 4 errors.
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! google-proto-files@2.4.0 compile: `tsc -p .`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the google-proto-files@2.4.0 compile script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_479Z-debug.log
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! google-proto-files@2.4.0 prepare: `npm run compile && node ./build/tools/prepublish.js`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the google-proto-files@2.4.0 prepare script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_547Z-debug.log
2021-04-09 00:43:26,576 synthtool [ERROR] > Failed executing npm install:
None
ERROR:synthtool:Failed executing npm install:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/nodejs-proto-files/synth.py", line 27, in <module>
node.install()
File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 171, in install
shell.run(["npm", "install"], hide_output=hide_output)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1.
2021-04-09 00:43:26,626 autosynth [ERROR] > Synthesis failed
2021-04-09 00:43:26,627 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 6ffe0d7 chore(deps): update dependency sinon to v10 (#337)
2021-04-09 00:43:26,634 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2021-04-09 00:43:26,639 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Removing node_modules/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/7ee53bfe-efd4-4c1a-ab7f-3a67ce2765e5/targets/github%2Fsynthtool;config=default/tests;query=nodejs-proto-files;failed=false).
| 1.0 | Synthesis failed for nodejs-proto-files - Hello! Autosynth couldn't regenerate nodejs-proto-files. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the nodejs-proto-files API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
MPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/.gitattributes
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/lint.cfg
.kokoro/continuous/node12/samples-test.cfg
.kokoro/continuous/node12/system-test.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/populate-secrets.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/samples-test.cfg
.kokoro/presubmit/node12/system-test.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/publish.sh
.kokoro/release/docs-devsite.cfg
.kokoro/release/docs-devsite.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.kokoro/trampoline_v2.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
.trampolinerc
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
api-extractor.json
renovate.json
samples/README.md
2021-04-09 00:42:57,323 synthtool [DEBUG] > Installing dependencies...
DEBUG:synthtool:Installing dependencies...
npm WARN deprecated sinon@10.0.1: Breaking change found in this patch version
npm WARN deprecated ts-simple-ast@12.4.0: NOTICE: ts-simple-ast has been renamed to ts-morph and version reset to 1.0.0. Switch at your leisure...
npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies.
npm WARN deprecated opn@6.0.0: The package has been renamed to `open`
npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2.
npm WARN deprecated core-js@2.6.12: core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to the actual version of core-js@3.
npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated
npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated
> core-js@2.6.12 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/core-js
> node -e "try{require('./postinstall')}catch(e){}"
[96mThank you for using core-js ([94m https://github.com/zloirock/core-js [96m) for polyfilling JavaScript standard library![0m
[96mThe project needs your help! Please consider supporting of core-js on Open Collective or Patreon: [0m
[96m>[94m https://opencollective.com/core-js [0m
[96m>[94m https://www.patreon.com/zloirock [0m
[96mAlso, the author of core-js ([94m https://github.com/zloirock [96m) is looking for a good job -)[0m
> @compodoc/compodoc@1.1.11 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@compodoc/compodoc
> opencollective-postinstall || exit 0
[96m[1mThank you for using @compodoc/compodoc![96m[1m
[0m[96mIf you rely on this package, please consider supporting our open collective:[22m[39m
> [94mhttps://opencollective.com/compodoc/donate[0m
> protobufjs@6.10.2 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/protobufjs
> node scripts/postinstall
> google-proto-files@2.4.0 prepare /home/kbuilder/.cache/synthtool/nodejs-proto-files
> npm run compile && node ./build/tools/prepublish.js
> google-proto-files@2.4.0 precompile /home/kbuilder/.cache/synthtool/nodejs-proto-files
> gts clean
version: 14
Removing build ...
> google-proto-files@2.4.0 compile /home/kbuilder/.cache/synthtool/nodejs-proto-files
> tsc -p .
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m778[0m:[93m36[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'TimerId'.
[7m778[0m type SinonTimerId = FakeTimers.TimerId;
[7m [0m [91m ~~~~~~~[0m
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m780[0m:[93m39[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'InstalledMethods'.
[7m780[0m type SinonFakeTimers = FakeTimers.InstalledMethods &
[7m [0m [91m ~~~~~~~~~~~~~~~~[0m
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m781[0m:[93m20[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'NodeClock'.
[7m781[0m FakeTimers.NodeClock &
[7m [0m [91m ~~~~~~~~~[0m
[96mnode_modules/@types/sinon/index.d.ts[0m:[93m782[0m:[93m20[0m - [91merror[0m[90m TS2694: [0mNamespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'BrowserClock'.
[7m782[0m FakeTimers.BrowserClock & {
[7m [0m [91m ~~~~~~~~~~~~[0m
Found 4 errors.
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! google-proto-files@2.4.0 compile: `tsc -p .`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the google-proto-files@2.4.0 compile script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_479Z-debug.log
npm ERR! code ELIFECYCLE
npm ERR! errno 1
npm ERR! google-proto-files@2.4.0 prepare: `npm run compile && node ./build/tools/prepublish.js`
npm ERR! Exit status 1
npm ERR!
npm ERR! Failed at the google-proto-files@2.4.0 prepare script.
npm ERR! This is probably not a problem with npm. There is likely additional logging output above.
npm ERR! A complete log of this run can be found in:
npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_547Z-debug.log
2021-04-09 00:43:26,576 synthtool [ERROR] > Failed executing npm install:
None
ERROR:synthtool:Failed executing npm install:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/nodejs-proto-files/synth.py", line 27, in <module>
node.install()
File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 171, in install
shell.run(["npm", "install"], hide_output=hide_output)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1.
2021-04-09 00:43:26,626 autosynth [ERROR] > Synthesis failed
2021-04-09 00:43:26,627 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 6ffe0d7 chore(deps): update dependency sinon to v10 (#337)
2021-04-09 00:43:26,634 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2021-04-09 00:43:26,639 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Removing node_modules/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/7ee53bfe-efd4-4c1a-ab7f-3a67ce2765e5/targets/github%2Fsynthtool;config=default/tests;query=nodejs-proto-files;failed=false).
| priority | synthesis failed for nodejs proto files hello autosynth couldn t regenerate nodejs proto files broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the nodejs proto files api and the library grows stale see for trouble shooting tips here s the output from running synth py mplate support request md github pull request template md github release please yml github workflows ci yaml kokoro gitattributes kokoro common cfg kokoro continuous common cfg kokoro continuous docs cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous lint cfg kokoro continuous samples test cfg kokoro continuous system test cfg kokoro continuous test cfg kokoro docs sh kokoro lint sh kokoro populate secrets sh kokoro presubmit common cfg kokoro presubmit common cfg kokoro presubmit samples test cfg kokoro presubmit system test cfg kokoro presubmit test cfg kokoro publish sh kokoro release docs devsite cfg kokoro release docs devsite sh kokoro release docs cfg kokoro release docs sh kokoro release publish cfg kokoro samples test sh kokoro system test sh kokoro test bat kokoro test sh kokoro trampoline sh kokoro trampoline sh mocharc js nycrc prettierignore prettierrc js trampolinerc code of conduct md contributing md license readme md api extractor json renovate json samples readme md synthtool installing dependencies debug synthtool installing dependencies npm warn deprecated sinon breaking change found in this patch version npm warn deprecated ts simple ast notice ts simple ast has been renamed to ts morph and version reset to switch at your leisure npm warn deprecated chokidar chokidar will break on node upgrade to chokidar with less dependencies npm warn deprecated opn the package has been renamed to open npm warn deprecated fsevents fsevents will break on node and could be using insecure binaries upgrade to fsevents npm warn deprecated core js core js is no longer maintained and not recommended for usage due to the number of issues please upgrade your dependencies to the actual version of core js npm warn deprecated resolve url npm warn deprecated urix please see core js postinstall home kbuilder cache synthtool nodejs proto files node modules core js node e try require postinstall catch e you for using core js for polyfilling javascript standard library project needs your help please consider supporting of core js on open collective or patreon the author of core js is looking for a good job compodoc compodoc postinstall home kbuilder cache synthtool nodejs proto files node modules compodoc compodoc opencollective postinstall exit you for using compodoc compodoc you rely on this package please consider supporting our open collective protobufjs postinstall home kbuilder cache synthtool nodejs proto files node modules protobufjs node scripts postinstall google proto files prepare home kbuilder cache synthtool nodejs proto files npm run compile node build tools prepublish js google proto files precompile home kbuilder cache synthtool nodejs proto files gts clean version removing build google proto files compile home kbuilder cache synthtool nodejs proto files tsc p modules types sinon index d ts home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member timerid type sinontimerid faketimers timerid modules types sinon index d ts home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member installedmethods type sinonfaketimers faketimers installedmethods modules types sinon index d ts home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member nodeclock faketimers nodeclock modules types sinon index d ts home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member browserclock faketimers browserclock found errors npm err code elifecycle npm err errno npm err google proto files compile tsc p npm err exit status npm err npm err failed at the google proto files compile script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log npm err code elifecycle npm err errno npm err google proto files prepare npm run compile node build tools prepublish js npm err exit status npm err npm err failed at the google proto files prepare script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool failed executing npm install none error synthtool failed executing npm install none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool nodejs proto files synth py line in node install file tmpfs src github synthtool synthtool languages node py line in install shell run hide output hide output file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at chore deps update dependency sinon to autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache removing node modules traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth toolbox py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 1 |
8,817 | 3,220,225,124 | IssuesEvent | 2015-10-08 14:01:22 | jqPlot/jqPlot | https://api.github.com/repos/jqPlot/jqPlot | closed | bar-charts on the examples website | bug documentation | hi, i wondered if this is the same thing. why does it not load any of the graphs?
http://www.jqplot.com/tests/bar-charts.php | 1.0 | bar-charts on the examples website - hi, i wondered if this is the same thing. why does it not load any of the graphs?
http://www.jqplot.com/tests/bar-charts.php | non_priority | bar charts on the examples website hi i wondered if this is the same thing why does it not load any of the graphs | 0 |
3,975 | 2,610,085,101 | IssuesEvent | 2015-02-26 18:25:54 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 深圳彩光治痤疮 | auto-migrated Priority-Medium Type-Defect | ```
深圳彩光治痤疮【深圳韩方科颜全国热线400-869-1818,24小时QQ4
008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方��
�—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科�
��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康
祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治��
�粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘�
��
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:02 | 1.0 | 深圳彩光治痤疮 - ```
深圳彩光治痤疮【深圳韩方科颜全国热线400-869-1818,24小时QQ4
008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方��
�—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科�
��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康
祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治��
�粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘�
��
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:02 | non_priority | 深圳彩光治痤疮 深圳彩光治痤疮【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� original issue reported on code google com by szft com on may at | 0 |
493,911 | 14,240,929,200 | IssuesEvent | 2020-11-18 22:28:53 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | `nim js --gc:arc` gives bad error: undeclared identifier: '+!' | ARC/ORC Error messages JS Low Priority | Function `echo` outputs the wrong string.
### Example
```nim
# can be an empty file
```
### Current Output
nim js --gc:arc --skipparentcfg --skipusercfg main.nim
lib/system/arc.nim(84, 42) Error: undeclared identifier: '+!'
### Expected Output
* either make `--gc:arc` a noop (and code should work)
* or give error `--gc:arc` not supported with js backend
### Additional Information
devel 1.5.1 87d3e5331a7838aa255deca8a840c2080b5424ce
this (or fixing it by making it not work with js backend) would prevent things like this in testament:
```nim
discard """
targets: "c cpp js"
matrix: ";--gc:arc"
"""
```
but testament could also be made smart about this and skip `js --gc:arc` combination, although this isn't trivial to do it cleanly
### note
curiously, this works:
```nim
# main2.nim:
echo compileOption("gc", "arc")
# cmd:
nim r -b:js --gc:arc main2.nim
false
```
likewise with `nim js -r -b:js --gc:arc main2.nim`
but this doesn't:
```
nim js -r --gc:arc main2.nim
Error: undeclared identifier: '+!'
```
| 1.0 | `nim js --gc:arc` gives bad error: undeclared identifier: '+!' - Function `echo` outputs the wrong string.
### Example
```nim
# can be an empty file
```
### Current Output
nim js --gc:arc --skipparentcfg --skipusercfg main.nim
lib/system/arc.nim(84, 42) Error: undeclared identifier: '+!'
### Expected Output
* either make `--gc:arc` a noop (and code should work)
* or give error `--gc:arc` not supported with js backend
### Additional Information
devel 1.5.1 87d3e5331a7838aa255deca8a840c2080b5424ce
this (or fixing it by making it not work with js backend) would prevent things like this in testament:
```nim
discard """
targets: "c cpp js"
matrix: ";--gc:arc"
"""
```
but testament could also be made smart about this and skip `js --gc:arc` combination, although this isn't trivial to do it cleanly
### note
curiously, this works:
```nim
# main2.nim:
echo compileOption("gc", "arc")
# cmd:
nim r -b:js --gc:arc main2.nim
false
```
likewise with `nim js -r -b:js --gc:arc main2.nim`
but this doesn't:
```
nim js -r --gc:arc main2.nim
Error: undeclared identifier: '+!'
```
| priority | nim js gc arc gives bad error undeclared identifier function echo outputs the wrong string example nim can be an empty file current output nim js gc arc skipparentcfg skipusercfg main nim lib system arc nim error undeclared identifier expected output either make gc arc a noop and code should work or give error gc arc not supported with js backend additional information devel this or fixing it by making it not work with js backend would prevent things like this in testament nim discard targets c cpp js matrix gc arc but testament could also be made smart about this and skip js gc arc combination although this isn t trivial to do it cleanly note curiously this works nim nim echo compileoption gc arc cmd nim r b js gc arc nim false likewise with nim js r b js gc arc nim but this doesn t nim js r gc arc nim error undeclared identifier | 1 |
30,443 | 4,209,838,484 | IssuesEvent | 2016-06-29 07:37:36 | gandm/language-babel | https://api.github.com/repos/gandm/language-babel | closed | 'react' incompatibility | as-designed | I'm getting an error saying that the package 'react' is not compatible with 'language-babel'. This happens after migrating to language-babel@2.29.0. | 1.0 | 'react' incompatibility - I'm getting an error saying that the package 'react' is not compatible with 'language-babel'. This happens after migrating to language-babel@2.29.0. | non_priority | react incompatibility i m getting an error saying that the package react is not compatible with language babel this happens after migrating to language babel | 0 |
3,047 | 5,207,647,206 | IssuesEvent | 2017-01-25 00:21:22 | saenzramiro/rambox | https://api.github.com/repos/saenzramiro/rambox | closed | Amium shows browser unsupported message | service-bug | <!-- Please complete the follow issue template. If is not an issue, feel free to remove the steps and behaviors -->
### Steps to reproduce
1. Login to Amium
### Expected behavior
Should detect the user agent as Chrome or Firefox instead of electron.
<!-- Write here -->
### Actual behavior
Shows this message "Oops, this browser isn't supported yet!"
<!-- Write here -->
### ENVIRONMENT
<!-- Complete your environment -->
> Rambox: v0.4.4
> OS: Ubuntu 16.04
> Arch: x64
| 1.0 | Amium shows browser unsupported message - <!-- Please complete the follow issue template. If is not an issue, feel free to remove the steps and behaviors -->
### Steps to reproduce
1. Login to Amium
### Expected behavior
Should detect the user agent as Chrome or Firefox instead of electron.
<!-- Write here -->
### Actual behavior
Shows this message "Oops, this browser isn't supported yet!"
<!-- Write here -->
### ENVIRONMENT
<!-- Complete your environment -->
> Rambox: v0.4.4
> OS: Ubuntu 16.04
> Arch: x64
| non_priority | amium shows browser unsupported message steps to reproduce login to amium expected behavior should detect the user agent as chrome or firefox instead of electron actual behavior shows this message oops this browser isn t supported yet environment rambox os ubuntu arch | 0 |
343,555 | 10,332,507,093 | IssuesEvent | 2019-09-03 00:32:07 | HackGT/bolt | https://api.github.com/repos/HackGT/bolt | closed | Logout is broken | component / auth priority / high type / bug | When you click the logout button, you get the following error
```js
TypeError: "listener" argument must be a function
at ClientRequest.once (events.js:340:11)
at new ClientRequest (_http_client.js:164:10)
at Object.request (http.js:38:10)
at Object.request (https.js:239:15)
at exports.authRoutes.all (C:\Users\Evan\Documents\GitHub\bolt\server\build\auth\auth.js:97:19)
at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5)
at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:137:13)
at Route.dispatch (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:112:3)
at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5)
at C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:281:22
at Function.process_params (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:335:12)
at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:275:10)
at Function.handle (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:174:3)
at router (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:47:12)
at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5)
at trim_prefix (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:317:13)``` | 1.0 | Logout is broken - When you click the logout button, you get the following error
```js
TypeError: "listener" argument must be a function
at ClientRequest.once (events.js:340:11)
at new ClientRequest (_http_client.js:164:10)
at Object.request (http.js:38:10)
at Object.request (https.js:239:15)
at exports.authRoutes.all (C:\Users\Evan\Documents\GitHub\bolt\server\build\auth\auth.js:97:19)
at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5)
at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:137:13)
at Route.dispatch (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:112:3)
at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5)
at C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:281:22
at Function.process_params (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:335:12)
at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:275:10)
at Function.handle (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:174:3)
at router (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:47:12)
at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5)
at trim_prefix (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:317:13)``` | priority | logout is broken when you click the logout button you get the following error js typeerror listener argument must be a function at clientrequest once events js at new clientrequest http client js at object request http js at object request https js at exports authroutes all c users evan documents github bolt server build auth auth js at layer handle c users evan documents github bolt server node modules express lib router layer js at next c users evan documents github bolt server node modules express lib router route js at route dispatch c users evan documents github bolt server node modules express lib router route js at layer handle c users evan documents github bolt server node modules express lib router layer js at c users evan documents github bolt server node modules express lib router index js at function process params c users evan documents github bolt server node modules express lib router index js at next c users evan documents github bolt server node modules express lib router index js at function handle c users evan documents github bolt server node modules express lib router index js at router c users evan documents github bolt server node modules express lib router index js at layer handle c users evan documents github bolt server node modules express lib router layer js at trim prefix c users evan documents github bolt server node modules express lib router index js | 1 |
191,816 | 14,596,491,050 | IssuesEvent | 2020-12-20 16:04:07 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | hello-mr-code/terraform-oci: oci/identity_network_source_test.go; 16 LoC | fresh small test |
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/identity_network_source_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to networkSourceId is reassigned at line 281
[Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, networkSourceId := range networkSourceIds {
if ok := SweeperDefaultResourceId[networkSourceId]; !ok {
deleteNetworkSourceRequest := oci_identity.DeleteNetworkSourceRequest{}
deleteNetworkSourceRequest.NetworkSourceId = &networkSourceId
deleteNetworkSourceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "identity")
_, error := identityClient.DeleteNetworkSource(context.Background(), deleteNetworkSourceRequest)
if error != nil {
fmt.Printf("Error deleting NetworkSource %s %s, It is possible that the resource is already deleted. Please verify manually \n", networkSourceId, error)
continue
}
waitTillCondition(testAccProvider, &networkSourceId, networkSourceSweepWaitCondition, time.Duration(3*time.Minute),
networkSourceSweepResponseFetchOperation, "identity", true)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
| 1.0 | hello-mr-code/terraform-oci: oci/identity_network_source_test.go; 16 LoC -
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/identity_network_source_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> reference to networkSourceId is reassigned at line 281
[Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292)
<details>
<summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary>
```go
for _, networkSourceId := range networkSourceIds {
if ok := SweeperDefaultResourceId[networkSourceId]; !ok {
deleteNetworkSourceRequest := oci_identity.DeleteNetworkSourceRequest{}
deleteNetworkSourceRequest.NetworkSourceId = &networkSourceId
deleteNetworkSourceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "identity")
_, error := identityClient.DeleteNetworkSource(context.Background(), deleteNetworkSourceRequest)
if error != nil {
fmt.Printf("Error deleting NetworkSource %s %s, It is possible that the resource is already deleted. Please verify manually \n", networkSourceId, error)
continue
}
waitTillCondition(testAccProvider, &networkSourceId, networkSourceSweepWaitCondition, time.Duration(3*time.Minute),
networkSourceSweepResponseFetchOperation, "identity", true)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
| non_priority | hello mr code terraform oci oci identity network source test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to networksourceid is reassigned at line click here to show the line s of go which triggered the analyzer go for networksourceid range networksourceids if ok sweeperdefaultresourceid ok deletenetworksourcerequest oci identity deletenetworksourcerequest deletenetworksourcerequest networksourceid networksourceid deletenetworksourcerequest requestmetadata retrypolicy getretrypolicy true identity error identityclient deletenetworksource context background deletenetworksourcerequest if error nil fmt printf error deleting networksource s s it is possible that the resource is already deleted please verify manually n networksourceid error continue waittillcondition testaccprovider networksourceid networksourcesweepwaitcondition time duration time minute networksourcesweepresponsefetchoperation identity true leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
432,586 | 30,288,544,900 | IssuesEvent | 2023-07-09 01:28:12 | aigic8/warmlight | https://api.github.com/repos/aigic8/warmlight | opened | document adding bot to Bot Father | documentation | It should be specified in the docs how to exactly add the bot to the Bot Father, especially bot commands, so they can be copied and sent to Bot Father | 1.0 | document adding bot to Bot Father - It should be specified in the docs how to exactly add the bot to the Bot Father, especially bot commands, so they can be copied and sent to Bot Father | non_priority | document adding bot to bot father it should be specified in the docs how to exactly add the bot to the bot father especially bot commands so they can be copied and sent to bot father | 0 |
42,704 | 17,267,417,850 | IssuesEvent | 2021-07-22 15:17:00 | terraform-providers/terraform-provider-azurerm | https://api.github.com/repos/terraform-providers/terraform-provider-azurerm | closed | Support for EnableAdvancedFilteringOnArrays on Event Grid Subscription | enhancement service/event-grid | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
The current [event grid subscription](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/eventgrid_event_subscription#advanced_filter) terraform does not support the property `EnableAdvancedFilteringOnArrays`
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_eventgrid_event_subscription
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_eventgrid_event_subscription" "default" {
name = "defaultEventSubscription"
scope = azurerm_resource_group.default.id
storage_queue_endpoint {
storage_account_id = azurerm_storage_account.default.id
queue_name = azurerm_storage_queue.default.name
}
advanced_filtering_on_arrays_enabled = true
}
```
### References
https://stackoverflow.com/questions/68345240/event-grid-subscription-terraform-missing-enable-advanced-filtering-on-arrays/
| 1.0 | Support for EnableAdvancedFilteringOnArrays on Event Grid Subscription - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
The current [event grid subscription](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/eventgrid_event_subscription#advanced_filter) terraform does not support the property `EnableAdvancedFilteringOnArrays`
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_eventgrid_event_subscription
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_eventgrid_event_subscription" "default" {
name = "defaultEventSubscription"
scope = azurerm_resource_group.default.id
storage_queue_endpoint {
storage_account_id = azurerm_storage_account.default.id
queue_name = azurerm_storage_queue.default.name
}
advanced_filtering_on_arrays_enabled = true
}
```
### References
https://stackoverflow.com/questions/68345240/event-grid-subscription-terraform-missing-enable-advanced-filtering-on-arrays/
| non_priority | support for enableadvancedfilteringonarrays on event grid subscription community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the current terraform does not support the property enableadvancedfilteringonarrays new or affected resource s azurerm eventgrid event subscription potential terraform configuration hcl resource azurerm eventgrid event subscription default name defaulteventsubscription scope azurerm resource group default id storage queue endpoint storage account id azurerm storage account default id queue name azurerm storage queue default name advanced filtering on arrays enabled true references | 0 |
314,176 | 26,980,012,445 | IssuesEvent | 2023-02-09 12:25:43 | astronomer/astronomer-providers | https://api.github.com/repos/astronomer/astronomer-providers | closed | Install astronomer-provider with Airflow constraint file | feature pri/high testing | For example astronomer-providers-1.0.0 with airflow 2.4.0 and Python3.8
```
pip install "astronomer-providers[all]==1.0.0" -c https://raw.githubusercontent.com/apache/airflow/constraints-2.4.0/constraints-3.8.txt
``` | 1.0 | Install astronomer-provider with Airflow constraint file - For example astronomer-providers-1.0.0 with airflow 2.4.0 and Python3.8
```
pip install "astronomer-providers[all]==1.0.0" -c https://raw.githubusercontent.com/apache/airflow/constraints-2.4.0/constraints-3.8.txt
``` | non_priority | install astronomer provider with airflow constraint file for example astronomer providers with airflow and pip install astronomer providers c | 0 |
97,771 | 28,497,475,101 | IssuesEvent | 2023-04-18 15:05:57 | deephaven/deephaven-core | https://api.github.com/repos/deephaven/deephaven-core | opened | Add build information to jars | feature request build | For benchmarking and other provenance use cases, it would be useful to have more build information as part of the jars.
The most useful pieces of information would probably be:
* git commit hash (-dirty)
* Is this a release build?
* Build date / timestamp (this may come into conflict with development workflows which may favor cacheability, so ostensibly may only be desired on release builds)
* Other build environment details
The most canonical place to add this information would be in the jar manifest `META-INF/MANIFEST.MF`: https://docs.oracle.com/en/java/javase/11/docs/specs/jar/jar.html#jar-manifest.
Here is an example of our current manifest:
```
Manifest-Version: 1.0
Implementation-Title: Deephaven
Implementation-Version: 0.23.0
Provider: gradle
```
Note: regardless of adding more build information, we may want to change `Implementation-Title` to actually be the archive base name instead of "Deephaven".
We _could_ add build information to a separate file in the jar, but unless there is a strong compelling reason why we _shouldn't_ add it to the manifest, the manifest is likely the best choice.
As an added bonus, this would give us an opportunity to add information that automating tooling likes to look for (license, PURL info, ...). We recently did this for https://github.com/deephaven/deephaven-csv/pull/93:
```
Manifest-Version: 1.0
Specification-Title: deephaven-csv
Specification-Version: 0.10.0
Specification-Vendor: Deephaven Data Labs
Implementation-Title: deephaven-csv
Implementation-Version: 0.10.0
Implementation-Vendor: Deephaven Data Labs
Implementation-Vendor-Id: io.deephaven
Implementation-URL: https://github.com/deephaven/deephaven-csv
Bundle-License: https://www.apache.org/licenses/LICENSE-2.0.txt
```
| 1.0 | Add build information to jars - For benchmarking and other provenance use cases, it would be useful to have more build information as part of the jars.
The most useful pieces of information would probably be:
* git commit hash (-dirty)
* Is this a release build?
* Build date / timestamp (this may come into conflict with development workflows which may favor cacheability, so ostensibly may only be desired on release builds)
* Other build environment details
The most canonical place to add this information would be in the jar manifest `META-INF/MANIFEST.MF`: https://docs.oracle.com/en/java/javase/11/docs/specs/jar/jar.html#jar-manifest.
Here is an example of our current manifest:
```
Manifest-Version: 1.0
Implementation-Title: Deephaven
Implementation-Version: 0.23.0
Provider: gradle
```
Note: regardless of adding more build information, we may want to change `Implementation-Title` to actually be the archive base name instead of "Deephaven".
We _could_ add build information to a separate file in the jar, but unless there is a strong compelling reason why we _shouldn't_ add it to the manifest, the manifest is likely the best choice.
As an added bonus, this would give us an opportunity to add information that automating tooling likes to look for (license, PURL info, ...). We recently did this for https://github.com/deephaven/deephaven-csv/pull/93:
```
Manifest-Version: 1.0
Specification-Title: deephaven-csv
Specification-Version: 0.10.0
Specification-Vendor: Deephaven Data Labs
Implementation-Title: deephaven-csv
Implementation-Version: 0.10.0
Implementation-Vendor: Deephaven Data Labs
Implementation-Vendor-Id: io.deephaven
Implementation-URL: https://github.com/deephaven/deephaven-csv
Bundle-License: https://www.apache.org/licenses/LICENSE-2.0.txt
```
| non_priority | add build information to jars for benchmarking and other provenance use cases it would be useful to have more build information as part of the jars the most useful pieces of information would probably be git commit hash dirty is this a release build build date timestamp this may come into conflict with development workflows which may favor cacheability so ostensibly may only be desired on release builds other build environment details the most canonical place to add this information would be in the jar manifest meta inf manifest mf here is an example of our current manifest manifest version implementation title deephaven implementation version provider gradle note regardless of adding more build information we may want to change implementation title to actually be the archive base name instead of deephaven we could add build information to a separate file in the jar but unless there is a strong compelling reason why we shouldn t add it to the manifest the manifest is likely the best choice as an added bonus this would give us an opportunity to add information that automating tooling likes to look for license purl info we recently did this for manifest version specification title deephaven csv specification version specification vendor deephaven data labs implementation title deephaven csv implementation version implementation vendor deephaven data labs implementation vendor id io deephaven implementation url bundle license | 0 |
312,013 | 26,831,659,970 | IssuesEvent | 2023-02-02 16:24:23 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] GeoIpDownloaderCliIT testGeoIpDatabasesDownloadNoGeoipProcessors failing | >test-failure :Data Management/Ingest Node Team:Data Management | Failed because of a NPE in the test itself.
**Build scan:**
https://gradle-enterprise.elastic.co/s/zztkk4y3bucwu/tests/:modules:ingest-geoip:internalClusterTest/org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT/testGeoIpDatabasesDownloadNoGeoipProcessors
**Reproduction line:**
```
./gradlew ':modules:ingest-geoip:internalClusterTest' --tests "org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT.testGeoIpDatabasesDownloadNoGeoipProcessors" -Dtests.seed=59DC63CBE0AC5A4C -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Grand_Turk -Druntime.java=17
```
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT&tests.test=testGeoIpDatabasesDownloadNoGeoipProcessors
**Failure excerpt:**
```
java.lang.NullPointerException: Cannot invoke "org.elasticsearch.persistent.PersistentTasksCustomMetadata$PersistentTask.getState()" because the return value of "org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.getTask()" is null
at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.lambda$testGeoIpDatabasesDownloadNoGeoipProcessors$14(GeoIpDownloaderIT.java:315)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1109)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1094)
at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.testGeoIpDatabasesDownloadNoGeoipProcessors(GeoIpDownloaderIT.java:313)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:48)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | 1.0 | [CI] GeoIpDownloaderCliIT testGeoIpDatabasesDownloadNoGeoipProcessors failing - Failed because of a NPE in the test itself.
**Build scan:**
https://gradle-enterprise.elastic.co/s/zztkk4y3bucwu/tests/:modules:ingest-geoip:internalClusterTest/org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT/testGeoIpDatabasesDownloadNoGeoipProcessors
**Reproduction line:**
```
./gradlew ':modules:ingest-geoip:internalClusterTest' --tests "org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT.testGeoIpDatabasesDownloadNoGeoipProcessors" -Dtests.seed=59DC63CBE0AC5A4C -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Grand_Turk -Druntime.java=17
```
**Applicable branches:**
main
**Reproduces locally?:**
No
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT&tests.test=testGeoIpDatabasesDownloadNoGeoipProcessors
**Failure excerpt:**
```
java.lang.NullPointerException: Cannot invoke "org.elasticsearch.persistent.PersistentTasksCustomMetadata$PersistentTask.getState()" because the return value of "org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.getTask()" is null
at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.lambda$testGeoIpDatabasesDownloadNoGeoipProcessors$14(GeoIpDownloaderIT.java:315)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1109)
at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1094)
at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.testGeoIpDatabasesDownloadNoGeoipProcessors(GeoIpDownloaderIT.java:313)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:568)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946)
at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:48)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850)
at java.lang.Thread.run(Thread.java:833)
``` | non_priority | geoipdownloadercliit testgeoipdatabasesdownloadnogeoipprocessors failing failed because of a npe in the test itself build scan reproduction line gradlew modules ingest geoip internalclustertest tests org elasticsearch ingest geoip geoipdownloadercliit testgeoipdatabasesdownloadnogeoipprocessors dtests seed dtests locale sr latn ba dtests timezone america grand turk druntime java applicable branches main reproduces locally no failure history failure excerpt java lang nullpointerexception cannot invoke org elasticsearch persistent persistenttaskscustommetadata persistenttask getstate because the return value of org elasticsearch ingest geoip geoipdownloaderit gettask is null at org elasticsearch ingest geoip geoipdownloaderit lambda testgeoipdatabasesdownloadnogeoipprocessors geoipdownloaderit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch ingest geoip geoipdownloaderit testgeoipdatabasesdownloadnogeoipprocessors geoipdownloaderit java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java | 0 |
53,624 | 13,261,989,026 | IssuesEvent | 2020-08-20 20:54:10 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | millipede / taupede: error message "Assertion `alpha >= 0' failed" leads to abortion (Trac #1773) | Migrated from Trac combo reconstruction defect | When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better?
A test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2
I am using a monopod-based seed, from "CscdL4_Monopod_noBD_4" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1773">https://code.icecube.wisc.edu/projects/icecube/ticket/1773</a>, reported by jstachurskaand owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-07-12T16:09:01",
"_ts": "1468339741343204",
"description": "When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better?\n\nA test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2 \nI am using a monopod-based seed, from \"CscdL4_Monopod_noBD_4\" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed.",
"reporter": "jstachurska",
"cc": "",
"resolution": "fixed",
"time": "2016-07-06T19:12:52",
"component": "combo reconstruction",
"summary": "millipede / taupede: error message \"Assertion `alpha >= 0' failed\" leads to abortion",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| 1.0 | millipede / taupede: error message "Assertion `alpha >= 0' failed" leads to abortion (Trac #1773) - When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better?
A test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2
I am using a monopod-based seed, from "CscdL4_Monopod_noBD_4" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1773">https://code.icecube.wisc.edu/projects/icecube/ticket/1773</a>, reported by jstachurskaand owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-07-12T16:09:01",
"_ts": "1468339741343204",
"description": "When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better?\n\nA test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2 \nI am using a monopod-based seed, from \"CscdL4_Monopod_noBD_4\" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed.",
"reporter": "jstachurska",
"cc": "",
"resolution": "fixed",
"time": "2016-07-06T19:12:52",
"component": "combo reconstruction",
"summary": "millipede / taupede: error message \"Assertion `alpha >= 0' failed\" leads to abortion",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| non_priority | millipede taupede error message assertion alpha failed leads to abortion trac when running taupede icerec trunc version reconstruction the above problem causes the whole job to abort maybe returning a warning and continuing with the next frame would be better a test file with unproblematic events and events causing abortion can be found here data user jstachurska nutauana trunk icerec i am using a monopod based seed from monopod nobd and setting the seed length max m p energy pev m for a positive length seed and the negative of that for a negative length seed migrated from json status closed changetime ts description when running taupede icerec trunc version reconstruction the above problem causes the whole job to abort maybe returning a warning and continuing with the next frame would be better n na test file with unproblematic events and events causing abortion can be found here data user jstachurska nutauana trunk icerec ni am using a monopod based seed from monopod nobd and setting the seed length max m p energy pev m for a positive length seed and the negative of that for a negative length seed reporter jstachurska cc resolution fixed time component combo reconstruction summary millipede taupede error message assertion alpha failed leads to abortion priority normal keywords milestone owner jbraun type defect | 0 |
60,008 | 3,117,669,176 | IssuesEvent | 2015-09-04 03:58:19 | framingeinstein/issues-test | https://api.github.com/repos/framingeinstein/issues-test | closed | SPK-41: Blog Category: Remove More Categories from page. | priority:low resolution:will-not-fix | Hi Andrew,
Let's just remove the "More Categories" functionality from the blog category to save time. See screenshot.
[~Spencer] - We can move this to 1.1 if needbe. | 1.0 | SPK-41: Blog Category: Remove More Categories from page. - Hi Andrew,
Let's just remove the "More Categories" functionality from the blog category to save time. See screenshot.
[~Spencer] - We can move this to 1.1 if needbe. | priority | spk blog category remove more categories from page hi andrew let s just remove the more categories functionality from the blog category to save time see screenshot we can move this to if needbe | 1 |
1,129 | 3,381,145,652 | IssuesEvent | 2015-11-26 00:11:05 | BCDevExchange/BCDevExchange-app | https://api.github.com/repos/BCDevExchange/BCDevExchange-app | opened | As a user I would like to test against the API methods so that I can determine if the API service delivers the information I need. | API Services Bulls Eye Center | Users need to be able to test the listed APIs to see if the API delivers information that is salient to the business need.
Created by Paul R
DOD
1. Users can test the API according to the spec.
2. Users can test against modern programming languages (JAVA, NODE, PHP etc)
Connect with
Peter, Todd, Fred, Leo | 1.0 | As a user I would like to test against the API methods so that I can determine if the API service delivers the information I need. - Users need to be able to test the listed APIs to see if the API delivers information that is salient to the business need.
Created by Paul R
DOD
1. Users can test the API according to the spec.
2. Users can test against modern programming languages (JAVA, NODE, PHP etc)
Connect with
Peter, Todd, Fred, Leo | non_priority | as a user i would like to test against the api methods so that i can determine if the api service delivers the information i need users need to be able to test the listed apis to see if the api delivers information that is salient to the business need created by paul r dod users can test the api according to the spec users can test against modern programming languages java node php etc connect with peter todd fred leo | 0 |
197,923 | 22,617,748,194 | IssuesEvent | 2022-06-30 01:04:38 | turkdevops/sanity | https://api.github.com/repos/turkdevops/sanity | opened | CVE-2022-2216 (High) detected in parse-url-5.0.2.tgz | security vulnerability | ## CVE-2022-2216 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary>
<p>An advanced url parser supporting git urls too.</p>
<p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/parse-url/package.json</p>
<p>
Dependency Hierarchy:
- lerna-3.22.1.tgz (Root Library)
- version-3.22.1.tgz
- github-client-3.22.0.tgz
- git-url-parse-11.4.4.tgz
- git-up-4.0.2.tgz
- :x: **parse-url-5.0.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>next</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 7.0.0.
<p>Publish Date: 2022-06-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2216>CVE-2022-2216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/">https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/</a></p>
<p>Release Date: 2022-06-27</p>
<p>Fix Resolution: parse-url - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-2216 (High) detected in parse-url-5.0.2.tgz - ## CVE-2022-2216 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary>
<p>An advanced url parser supporting git urls too.</p>
<p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/parse-url/package.json</p>
<p>
Dependency Hierarchy:
- lerna-3.22.1.tgz (Root Library)
- version-3.22.1.tgz
- github-client-3.22.0.tgz
- git-url-parse-11.4.4.tgz
- git-up-4.0.2.tgz
- :x: **parse-url-5.0.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>next</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 7.0.0.
<p>Publish Date: 2022-06-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2216>CVE-2022-2216</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/">https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/</a></p>
<p>Release Date: 2022-06-27</p>
<p>Fix Resolution: parse-url - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in parse url tgz cve high severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href path to dependency file package json path to vulnerable library node modules parse url package json dependency hierarchy lerna tgz root library version tgz github client tgz git url parse tgz git up tgz x parse url tgz vulnerable library found in base branch next vulnerability details server side request forgery ssrf in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution parse url step up your open source security game with mend | 0 |
84,407 | 15,720,901,073 | IssuesEvent | 2021-03-29 01:33:55 | benchmarkdebricked/generator-jhipster | https://api.github.com/repos/benchmarkdebricked/generator-jhipster | closed | CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed | security vulnerability | ## CVE-2020-5398 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library)
- :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input.
<p>Publish Date: 2020-01-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p>
<p>Release Date: 2020-01-17</p>
<p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed - ## CVE-2020-5398 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library)
- :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input.
<p>Publish Date: 2020-01-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p>
<p>Release Date: 2020-01-17</p>
<p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in spring web release jar autoclosed cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file tmp ws scm generator jhipster test templates ci cd maven ngx yarn pom xml path to vulnerable library root repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in head commit a href vulnerability details in spring framework versions x prior to versions x prior to and versions x prior to an application is vulnerable to a reflected file download rfd attack when it sets a content disposition header in the response where the filename attribute is derived from user supplied input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web release org springframework spring web release org springframework spring web release step up your open source security game with whitesource | 0 |
64,349 | 6,900,338,494 | IssuesEvent | 2017-11-24 17:59:21 | emfoundation/ce100-app | https://api.github.com/repos/emfoundation/ce100-app | closed | filter organistions by tags (from the API) | bug DAM-API please-test T2h | Organisations can be filtered by tags and locations:
- [x] update the list of tags and location
- [x] update the queries to search for tag or location | 1.0 | filter organistions by tags (from the API) - Organisations can be filtered by tags and locations:
- [x] update the list of tags and location
- [x] update the queries to search for tag or location | non_priority | filter organistions by tags from the api organisations can be filtered by tags and locations update the list of tags and location update the queries to search for tag or location | 0 |
89,157 | 3,790,282,569 | IssuesEvent | 2016-03-21 20:54:27 | phetsims/fraction-matcher | https://api.github.com/repos/phetsims/fraction-matcher | closed | Shapes are not correctly positioned in the match collection | priority:2-high type:bug type:performance | The position of representations (circles, squares, etc...) in their gray box will be incorrect if they are included in the last match of a perfectly scored level. This was seen to happen on Witten (iPad2 9.2.1), diSessa (iPad3 9.2.1), and Tycho (iPad2 0.3 Beta 6).
I'm labeling this as a performance issue since the iPad Air 2 (Leibniz) does not show this problem. The Air blows the other iPads out of the water with its performance. There is a noticeable lag between pressing "OK" on the last match and when the reward cascade begins. During this lag is when the match animates towards the gray box and comes to rest incorrectly.

Zoomed:

Assigning to @ariel-phet to figure out who will deal with this. For phetsims/tasks/issues/541. | 1.0 | Shapes are not correctly positioned in the match collection - The position of representations (circles, squares, etc...) in their gray box will be incorrect if they are included in the last match of a perfectly scored level. This was seen to happen on Witten (iPad2 9.2.1), diSessa (iPad3 9.2.1), and Tycho (iPad2 0.3 Beta 6).
I'm labeling this as a performance issue since the iPad Air 2 (Leibniz) does not show this problem. The Air blows the other iPads out of the water with its performance. There is a noticeable lag between pressing "OK" on the last match and when the reward cascade begins. During this lag is when the match animates towards the gray box and comes to rest incorrectly.

Zoomed:

Assigning to @ariel-phet to figure out who will deal with this. For phetsims/tasks/issues/541. | priority | shapes are not correctly positioned in the match collection the position of representations circles squares etc in their gray box will be incorrect if they are included in the last match of a perfectly scored level this was seen to happen on witten disessa and tycho beta i m labeling this as a performance issue since the ipad air leibniz does not show this problem the air blows the other ipads out of the water with its performance there is a noticeable lag between pressing ok on the last match and when the reward cascade begins during this lag is when the match animates towards the gray box and comes to rest incorrectly zoomed assigning to ariel phet to figure out who will deal with this for phetsims tasks issues | 1 |
216,184 | 7,301,965,700 | IssuesEvent | 2018-02-27 08:00:00 | Motoxpro/WorldCupStatsSite | https://api.github.com/repos/Motoxpro/WorldCupStatsSite | closed | Add Timed Training Overall | Medium Priority Data Issue Medium Priority Feature MySQL | Create a timed training overall by adding up all of the finishes and taking the lowest score. If a rider hasn't done a timed training round, their finish for that round is last place or just the number of riders in the session.
Don't bother breaking ties | 2.0 | Add Timed Training Overall - Create a timed training overall by adding up all of the finishes and taking the lowest score. If a rider hasn't done a timed training round, their finish for that round is last place or just the number of riders in the session.
Don't bother breaking ties | priority | add timed training overall create a timed training overall by adding up all of the finishes and taking the lowest score if a rider hasn t done a timed training round their finish for that round is last place or just the number of riders in the session don t bother breaking ties | 1 |
301,748 | 9,223,542,809 | IssuesEvent | 2019-03-12 03:58:30 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | net: icmpv4: Zephyr drops valid echo request | area: Conformance area: Networking bug priority: medium | Zephyr drops echo request with valid checksum. Zephyr must respond to valid echo request.
[icmpv4-valid-chksum2.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610165/icmpv4-valid-chksum2.pcap.gz)
[icmpv4-valid-chksum1.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610166/icmpv4-valid-chksum1.pcap.gz)
| 1.0 | net: icmpv4: Zephyr drops valid echo request - Zephyr drops echo request with valid checksum. Zephyr must respond to valid echo request.
[icmpv4-valid-chksum2.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610165/icmpv4-valid-chksum2.pcap.gz)
[icmpv4-valid-chksum1.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610166/icmpv4-valid-chksum1.pcap.gz)
| priority | net zephyr drops valid echo request zephyr drops echo request with valid checksum zephyr must respond to valid echo request | 1 |
36,340 | 14,979,486,378 | IssuesEvent | 2021-01-28 12:20:01 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [Drilldowns] Enable Drilldowns in edit mode for faster testing | Feature:Dashboard Feature:Drilldowns Team:AppServices enhancement | Currently, it is very tedious to test changes in a drilldown as you have to:
1. save a drilldown
2. save the dashboard
3. try the drilldown
We intentionally disabled drilldowns in edit mode, because we were afraid the users would lose unsaved changes when navigating away.
But after [this awesome pr](https://github.com/elastic/kibana/pull/86939) that stores drafts in session storage this isn't a problem anymore.
So it seems like we can safely enable drilldowns in the edit mode and remove one step in between editing and trying-out the drilldown.
We should also remove text from drilldown toasts that mention "save you dashboard before testing"
cc @streamich @elastic-jb | 1.0 | [Drilldowns] Enable Drilldowns in edit mode for faster testing - Currently, it is very tedious to test changes in a drilldown as you have to:
1. save a drilldown
2. save the dashboard
3. try the drilldown
We intentionally disabled drilldowns in edit mode, because we were afraid the users would lose unsaved changes when navigating away.
But after [this awesome pr](https://github.com/elastic/kibana/pull/86939) that stores drafts in session storage this isn't a problem anymore.
So it seems like we can safely enable drilldowns in the edit mode and remove one step in between editing and trying-out the drilldown.
We should also remove text from drilldown toasts that mention "save you dashboard before testing"
cc @streamich @elastic-jb | non_priority | enable drilldowns in edit mode for faster testing currently it is very tedious to test changes in a drilldown as you have to save a drilldown save the dashboard try the drilldown we intentionally disabled drilldowns in edit mode because we were afraid the users would lose unsaved changes when navigating away but after that stores drafts in session storage this isn t a problem anymore so it seems like we can safely enable drilldowns in the edit mode and remove one step in between editing and trying out the drilldown we should also remove text from drilldown toasts that mention save you dashboard before testing cc streamich elastic jb | 0 |
443,384 | 12,793,665,996 | IssuesEvent | 2020-07-02 04:46:39 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | Cannot delete conversation when Notification component is deactivated | Has-PR bug priority: medium status: needs review | **Describe the bug**
Cannot delete conversation when Notification component is deactivated
**To Reproduce**
Steps to reproduce the behavior:
1. Deactivate notification component
2. Try to delete a conversation, notice that it will not be deleted.
**Screencast**
https://drive.google.com/file/d/1KuPuBwPM7Av406m2KrErIHH36c3PMC0R/view
**Support ticket links**
https://secure.helpscout.net/conversation/1205558198/79889
| 1.0 | Cannot delete conversation when Notification component is deactivated - **Describe the bug**
Cannot delete conversation when Notification component is deactivated
**To Reproduce**
Steps to reproduce the behavior:
1. Deactivate notification component
2. Try to delete a conversation, notice that it will not be deleted.
**Screencast**
https://drive.google.com/file/d/1KuPuBwPM7Av406m2KrErIHH36c3PMC0R/view
**Support ticket links**
https://secure.helpscout.net/conversation/1205558198/79889
| priority | cannot delete conversation when notification component is deactivated describe the bug cannot delete conversation when notification component is deactivated to reproduce steps to reproduce the behavior deactivate notification component try to delete a conversation notice that it will not be deleted screencast support ticket links | 1 |
338,493 | 10,230,066,965 | IssuesEvent | 2019-08-17 18:03:43 | simplitech/simpli-web-sdk | https://api.github.com/repos/simplitech/simpli-web-sdk | closed | input components should validate and validation and on schema should be a bind property instead of separate avj info | bug highpriority | selects are not currently validable | 1.0 | input components should validate and validation and on schema should be a bind property instead of separate avj info - selects are not currently validable | priority | input components should validate and validation and on schema should be a bind property instead of separate avj info selects are not currently validable | 1 |
384,251 | 11,386,054,434 | IssuesEvent | 2020-01-29 12:28:26 | NukkitX/Nukkit | https://api.github.com/repos/NukkitX/Nukkit | closed | Speedbug | [Priority] Medium [Status] Unconfirmed [Type] Bug |
### Actual Behavior
<!--- What actually happened -->
I had a problem on my server using Nukkit X last version. Players spawn 9/10 times with a speed effect (not visible on effect menu) very high. Please fix this bug !
### Steps to Reproduce
<!--- Reliable steps which someone can use to reproduce the issue. Please do not create issues for non reproducible bug! -->
I dont know.... sorry :C
| 1.0 | Speedbug -
### Actual Behavior
<!--- What actually happened -->
I had a problem on my server using Nukkit X last version. Players spawn 9/10 times with a speed effect (not visible on effect menu) very high. Please fix this bug !
### Steps to Reproduce
<!--- Reliable steps which someone can use to reproduce the issue. Please do not create issues for non reproducible bug! -->
I dont know.... sorry :C
| priority | speedbug actual behavior i had a problem on my server using nukkit x last version players spawn times with a speed effect not visible on effect menu very high please fix this bug steps to reproduce i dont know sorry c | 1 |
30,367 | 8,526,463,699 | IssuesEvent | 2018-11-02 16:19:30 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | build with default cuda options fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable" | type:build/install | **System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 16.04.5 x86_64
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): master branch from 10/31/18 -
- Python version:
- Bazel version (if compiling from source): 0.18.0
- GCC/Compiler version (if compiling from source): 5.4.0 20160609
- CUDA/cuDNN version: 9.0/7
- GPU model and memory: 2x Tesla P100 16 GB
**Describe the current behavior**
Configuring TF to build with CUDA support and choosing all the default option for all the other questions in ./configure including the default for NCCL "Please specify the locally installed NCCL version you want to use. [Default is to use https://github.com/nvidia/nccl]:", the build fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable"
This has been broken since this commit:
https://github.com/tensorflow/tensorflow/commit/fc6cd33c334f88759ce637e29e1586733076e094
Running in the docker container: nvidia/cuda:9.0-cudnn7-devel (for ease of setup)
export TF_NEED_CUDA=1
"" | ./configure
bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package
failed with:
ERROR: /root/.cache/bazel/_bazel_root/68a62076e91007a7908bc42a32e4cff9/external/nccl_archive/BUILD.bazel:139:1: Executing genrule @nccl_archive//:device_code_fatbin_h failed (Exit 1)
/bin/bash: CUDA_TOOLKIT_PATH: unbound variable
Target //tensorflow/tools/pip_package:build_pip_package failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 267.057s, Critical Path: 93.66s
INFO: 3606 processes: 3606 local.
FAILED: Build did NOT complete successfully
**Describe the expected behavior**
Build should succeed.
**Code to reproduce the issue**
export TF_NEED_CUDA=1
"" | ./configure
bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package
**Other info / logs**
Fails on both x86 and ppc64le, tried with both bazel 0.15.0 and 0.18.0
| 1.0 | build with default cuda options fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable" - **System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 16.04.5 x86_64
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A
- TensorFlow installed from (source or binary): source
- TensorFlow version (use command below): master branch from 10/31/18 -
- Python version:
- Bazel version (if compiling from source): 0.18.0
- GCC/Compiler version (if compiling from source): 5.4.0 20160609
- CUDA/cuDNN version: 9.0/7
- GPU model and memory: 2x Tesla P100 16 GB
**Describe the current behavior**
Configuring TF to build with CUDA support and choosing all the default option for all the other questions in ./configure including the default for NCCL "Please specify the locally installed NCCL version you want to use. [Default is to use https://github.com/nvidia/nccl]:", the build fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable"
This has been broken since this commit:
https://github.com/tensorflow/tensorflow/commit/fc6cd33c334f88759ce637e29e1586733076e094
Running in the docker container: nvidia/cuda:9.0-cudnn7-devel (for ease of setup)
export TF_NEED_CUDA=1
"" | ./configure
bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package
failed with:
ERROR: /root/.cache/bazel/_bazel_root/68a62076e91007a7908bc42a32e4cff9/external/nccl_archive/BUILD.bazel:139:1: Executing genrule @nccl_archive//:device_code_fatbin_h failed (Exit 1)
/bin/bash: CUDA_TOOLKIT_PATH: unbound variable
Target //tensorflow/tools/pip_package:build_pip_package failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 267.057s, Critical Path: 93.66s
INFO: 3606 processes: 3606 local.
FAILED: Build did NOT complete successfully
**Describe the expected behavior**
Build should succeed.
**Code to reproduce the issue**
export TF_NEED_CUDA=1
"" | ./configure
bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package
**Other info / logs**
Fails on both x86 and ppc64le, tried with both bazel 0.15.0 and 0.18.0
| non_priority | build with default cuda options fails with bin bash cuda toolkit path unbound variable system information have i written custom code as opposed to using a stock example script provided in tensorflow no os platform and distribution e g linux ubuntu linux ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device n a tensorflow installed from source or binary source tensorflow version use command below master branch from python version bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory tesla gb describe the current behavior configuring tf to build with cuda support and choosing all the default option for all the other questions in configure including the default for nccl please specify the locally installed nccl version you want to use the build fails with bin bash cuda toolkit path unbound variable this has been broken since this commit running in the docker container nvidia cuda devel for ease of setup export tf need cuda configure bazel build config opt config cuda tensorflow tools pip package build pip package failed with error root cache bazel bazel root external nccl archive build bazel executing genrule nccl archive device code fatbin h failed exit bin bash cuda toolkit path unbound variable target tensorflow tools pip package build pip package failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes local failed build did not complete successfully describe the expected behavior build should succeed code to reproduce the issue export tf need cuda configure bazel build config opt config cuda tensorflow tools pip package build pip package other info logs fails on both and tried with both bazel and | 0 |
827,353 | 31,767,098,978 | IssuesEvent | 2023-09-12 09:26:00 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | [Playlist] Brave crashes when `Remove from Playlist` is clicked in the address bar playlist + button dropdown menu. | crash OS/Windows priority/P3 QA/No release-notes/exclude features/playlist OS/Desktop | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
@sangwoo108
Clicking on `Remove from Playlist` is in the address bar is crashing the browser, it works fine when removing from the sidebar Playlist panel.
I also noticed the 'Open in playlist' doesn't work, but doesn't crash the browser.
1.60.7 Chromium: 117.0.5938.48 (Official Build) nightly (64-bit) Windows 11 22H2
https://github.com/brave/brave-browser/assets/122518587/20a51737-9390-433b-bc0a-7be85c7979b8
Have a good day!
BTW, something unrelated to this issue but about Playlist.
In the past, you could manually add links to playlist, will that come back at some point?
But more important, will Brave Playlist support grabbing many links/videos at once and using something like Youtube Playlists/Mixes?
There is the `More media on this site` button, for other type of sites, but I thought I would ask, since it would be nice to have like a way to just paste Playlists links or get them from a playlist link, just to get videos added more easily, instead of manually go one by one; also, that would be useful to to avoid going to Youtube to get videos in the playlist, like if someone got the the link from the search engine or something like that.
| 1.0 | [Playlist] Brave crashes when `Remove from Playlist` is clicked in the address bar playlist + button dropdown menu. - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
@sangwoo108
Clicking on `Remove from Playlist` is in the address bar is crashing the browser, it works fine when removing from the sidebar Playlist panel.
I also noticed the 'Open in playlist' doesn't work, but doesn't crash the browser.
1.60.7 Chromium: 117.0.5938.48 (Official Build) nightly (64-bit) Windows 11 22H2
https://github.com/brave/brave-browser/assets/122518587/20a51737-9390-433b-bc0a-7be85c7979b8
Have a good day!
BTW, something unrelated to this issue but about Playlist.
In the past, you could manually add links to playlist, will that come back at some point?
But more important, will Brave Playlist support grabbing many links/videos at once and using something like Youtube Playlists/Mixes?
There is the `More media on this site` button, for other type of sites, but I thought I would ask, since it would be nice to have like a way to just paste Playlists links or get them from a playlist link, just to get videos added more easily, instead of manually go one by one; also, that would be useful to to avoid going to Youtube to get videos in the playlist, like if someone got the the link from the search engine or something like that.
| priority | brave crashes when remove from playlist is clicked in the address bar playlist button dropdown menu have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description clicking on remove from playlist is in the address bar is crashing the browser it works fine when removing from the sidebar playlist panel i also noticed the open in playlist doesn t work but doesn t crash the browser chromium official build nightly bit windows have a good day btw something unrelated to this issue but about playlist in the past you could manually add links to playlist will that come back at some point but more important will brave playlist support grabbing many links videos at once and using something like youtube playlists mixes there is the more media on this site button for other type of sites but i thought i would ask since it would be nice to have like a way to just paste playlists links or get them from a playlist link just to get videos added more easily instead of manually go one by one also that would be useful to to avoid going to youtube to get videos in the playlist like if someone got the the link from the search engine or something like that | 1 |
618,379 | 19,433,775,646 | IssuesEvent | 2021-12-21 14:52:35 | BIDMCDigitalPsychiatry/LAMP-platform | https://api.github.com/repos/BIDMCDigitalPsychiatry/LAMP-platform | closed | User Hierarchy Update | bug 2day frontend priority 1 | Can you please add the permission for User Admin roles to have access to the manage credentials button in the top left corner when they click on their listed role (where it says User Administrator in the corner). Additionally, make it so they can only assign the roles of user administrator and practice lead, not system administrator | 1.0 | User Hierarchy Update - Can you please add the permission for User Admin roles to have access to the manage credentials button in the top left corner when they click on their listed role (where it says User Administrator in the corner). Additionally, make it so they can only assign the roles of user administrator and practice lead, not system administrator | priority | user hierarchy update can you please add the permission for user admin roles to have access to the manage credentials button in the top left corner when they click on their listed role where it says user administrator in the corner additionally make it so they can only assign the roles of user administrator and practice lead not system administrator | 1 |
297,400 | 25,727,678,935 | IssuesEvent | 2022-12-07 17:47:23 | liberland/liberland_substrate | https://api.github.com/repos/liberland/liberland_substrate | opened | LLM - multiple calls lack proper permission checks | tests-blocker | Following calls can be executed by any signed origin:
* `fake_send` - we should only have it enabled on dev. Having it on/off should be part of chain_spec, it's not acceptable to have to change code when releasing to mainnet.
* `freeze_llm` - this call allows freezing LLM for any account, as long as origin is signed. It should either require specific origin or only apply to account of origin
* `createllm` - is this even needed?
* `mint_llm` - I'm not sure here, as there are some cap checks inside, so maybe it's OK for anyone to call it anytime... but isn't it already handled by `try_mint` run on block initialize? Maybe this should be removed? | 1.0 | LLM - multiple calls lack proper permission checks - Following calls can be executed by any signed origin:
* `fake_send` - we should only have it enabled on dev. Having it on/off should be part of chain_spec, it's not acceptable to have to change code when releasing to mainnet.
* `freeze_llm` - this call allows freezing LLM for any account, as long as origin is signed. It should either require specific origin or only apply to account of origin
* `createllm` - is this even needed?
* `mint_llm` - I'm not sure here, as there are some cap checks inside, so maybe it's OK for anyone to call it anytime... but isn't it already handled by `try_mint` run on block initialize? Maybe this should be removed? | non_priority | llm multiple calls lack proper permission checks following calls can be executed by any signed origin fake send we should only have it enabled on dev having it on off should be part of chain spec it s not acceptable to have to change code when releasing to mainnet freeze llm this call allows freezing llm for any account as long as origin is signed it should either require specific origin or only apply to account of origin createllm is this even needed mint llm i m not sure here as there are some cap checks inside so maybe it s ok for anyone to call it anytime but isn t it already handled by try mint run on block initialize maybe this should be removed | 0 |
103,312 | 8,895,997,387 | IssuesEvent | 2019-01-16 10:13:57 | tracim/tracim | https://api.github.com/repos/tracim/tracim | opened | Cypress test: test not working good all the time tiny_mce__can_type_content_spec.js | backend test | **At present**
The test "app_html-document/tiny_mce__can_type_content_spec.js" is not working correctly each time is starting. Sometime you just need to restart this test and its working
**Expected**
- [ ] If this test not working good all the time, we need to disabled this test.
| 1.0 | Cypress test: test not working good all the time tiny_mce__can_type_content_spec.js - **At present**
The test "app_html-document/tiny_mce__can_type_content_spec.js" is not working correctly each time is starting. Sometime you just need to restart this test and its working
**Expected**
- [ ] If this test not working good all the time, we need to disabled this test.
| non_priority | cypress test test not working good all the time tiny mce can type content spec js at present the test app html document tiny mce can type content spec js is not working correctly each time is starting sometime you just need to restart this test and its working expected if this test not working good all the time we need to disabled this test | 0 |
414,550 | 27,991,906,811 | IssuesEvent | 2023-03-27 05:00:28 | CarsOk/castillo_del_rey | https://api.github.com/repos/CarsOk/castillo_del_rey | closed | Script HTML, CSS | documentation | # Historias de usuario
## Yo como
Aprendiz sena
# Quiero
llevar a cabo la realización de las interfaces de mi proyecto con su orden en el index
# Para
que de esta forma la avancemos con nuestro aplicativo.
# Criterios de aceptación:
- [] Debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente.
- [] la información debe ser verídica y completa. | 1.0 | Script HTML, CSS - # Historias de usuario
## Yo como
Aprendiz sena
# Quiero
llevar a cabo la realización de las interfaces de mi proyecto con su orden en el index
# Para
que de esta forma la avancemos con nuestro aplicativo.
# Criterios de aceptación:
- [] Debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente.
- [] la información debe ser verídica y completa. | non_priority | script html css historias de usuario yo como aprendiz sena quiero llevar a cabo la realización de las interfaces de mi proyecto con su orden en el index para que de esta forma la avancemos con nuestro aplicativo criterios de aceptación debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente la información debe ser verídica y completa | 0 |
7,737 | 5,176,455,958 | IssuesEvent | 2017-01-19 00:46:53 | geneontology/go-site | https://api.github.com/repos/geneontology/go-site | closed | Make sure that users.yaml is the sole source for user metadata | bug (B: affects usability) | Now that users.yaml (and its schema) have been merged (#106) added to the repo, no other user metadata should be edited or used.
Talking with @cmungall, some of this will be taken care of by documentation sooner (#105), but at some point we will have to do one of the following:
1. remove the files: GO.curators_dbxrefs, GO.user_data.json (TermGenie), termgenie-user-permissions.json (TermGenie)
2. make the listed files read-only versions that are produced by the users.yaml master (like dbxrefs.yaml)
This is a ticket to pick back up on this later when we have more bandwidth.
| True | Make sure that users.yaml is the sole source for user metadata - Now that users.yaml (and its schema) have been merged (#106) added to the repo, no other user metadata should be edited or used.
Talking with @cmungall, some of this will be taken care of by documentation sooner (#105), but at some point we will have to do one of the following:
1. remove the files: GO.curators_dbxrefs, GO.user_data.json (TermGenie), termgenie-user-permissions.json (TermGenie)
2. make the listed files read-only versions that are produced by the users.yaml master (like dbxrefs.yaml)
This is a ticket to pick back up on this later when we have more bandwidth.
| non_priority | make sure that users yaml is the sole source for user metadata now that users yaml and its schema have been merged added to the repo no other user metadata should be edited or used talking with cmungall some of this will be taken care of by documentation sooner but at some point we will have to do one of the following remove the files go curators dbxrefs go user data json termgenie termgenie user permissions json termgenie make the listed files read only versions that are produced by the users yaml master like dbxrefs yaml this is a ticket to pick back up on this later when we have more bandwidth | 0 |
151,968 | 13,440,196,295 | IssuesEvent | 2020-09-08 00:00:24 | InstituteforDiseaseModeling/covasim | https://api.github.com/repos/InstituteforDiseaseModeling/covasim | closed | WebApp: Tooltips for testing and tracing intervention should mention that people are isolated. | CovasimUI documentation enhancement | Covasim, WebApp, click to expand Interventions Bar, hover over the intervention tabs
Current descriptions for Testing and Tracing interventions are:
"Testing rates for people with symptoms" and "Contact Tracing of diagnosed cases (requires testing intervention to be set)"
Nothing in this indicates how that reduces the spread.
I propose to expand the descriptions to "Testing rates and, if found positive isolating, for people with symptoms" and "Contact tracing of diagnosed cases, those found positive are isolated (requires testing intervention to be set)" | 1.0 | WebApp: Tooltips for testing and tracing intervention should mention that people are isolated. - Covasim, WebApp, click to expand Interventions Bar, hover over the intervention tabs
Current descriptions for Testing and Tracing interventions are:
"Testing rates for people with symptoms" and "Contact Tracing of diagnosed cases (requires testing intervention to be set)"
Nothing in this indicates how that reduces the spread.
I propose to expand the descriptions to "Testing rates and, if found positive isolating, for people with symptoms" and "Contact tracing of diagnosed cases, those found positive are isolated (requires testing intervention to be set)" | non_priority | webapp tooltips for testing and tracing intervention should mention that people are isolated covasim webapp click to expand interventions bar hover over the intervention tabs current descriptions for testing and tracing interventions are testing rates for people with symptoms and contact tracing of diagnosed cases requires testing intervention to be set nothing in this indicates how that reduces the spread i propose to expand the descriptions to testing rates and if found positive isolating for people with symptoms and contact tracing of diagnosed cases those found positive are isolated requires testing intervention to be set | 0 |
95,070 | 10,865,387,877 | IssuesEvent | 2019-11-14 18:53:27 | AIR-FOI-HR/AIR1925 | https://api.github.com/repos/AIR-FOI-HR/AIR1925 | closed | Izrada korisničkih zahtjeva | documentation | Identificiranje te popisivanje korisničkih zahtjeva sa razrađenim opisom i scenarijima. | 1.0 | Izrada korisničkih zahtjeva - Identificiranje te popisivanje korisničkih zahtjeva sa razrađenim opisom i scenarijima. | non_priority | izrada korisničkih zahtjeva identificiranje te popisivanje korisničkih zahtjeva sa razrađenim opisom i scenarijima | 0 |
148,282 | 19,529,190,873 | IssuesEvent | 2021-12-30 13:41:27 | developerone12/WebGoat-WhiteSource-Bolt | https://api.github.com/repos/developerone12/WebGoat-WhiteSource-Bolt | opened | CVE-2019-10768 (High) detected in angular-1.2.16.js, angular-1.2.16.min.js | security vulnerability | ## CVE-2019-10768 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.2.16.js</b>, <b>angular-1.2.16.min.js</b></p></summary>
<p>
<details><summary><b>angular-1.2.16.js</b></p></summary>
<p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js</a></p>
<p>Path to vulnerable library: /src/main/webapp/js/angular/angular.js</p>
<p>
Dependency Hierarchy:
- :x: **angular-1.2.16.js** (Vulnerable Library)
</details>
<details><summary><b>angular-1.2.16.min.js</b></p></summary>
<p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js</a></p>
<p>Path to vulnerable library: /src/main/webapp/js/angular/angular.min.js</p>
<p>
Dependency Hierarchy:
- :x: **angular-1.2.16.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/developerone12/WebGoat-WhiteSource-Bolt/commit/c42e663814e4b88294ff90339ad577ca1afcf531">c42e663814e4b88294ff90339ad577ca1afcf531</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In AngularJS before 1.7.9 the function `merge()` could be tricked into adding or modifying properties of `Object.prototype` using a `__proto__` payload.
<p>Publish Date: 2019-11-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10768>CVE-2019-10768</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a">https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a</a></p>
<p>Release Date: 2019-11-19</p>
<p>Fix Resolution: angularjs - 1.7.8,1.7.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-10768 (High) detected in angular-1.2.16.js, angular-1.2.16.min.js - ## CVE-2019-10768 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.2.16.js</b>, <b>angular-1.2.16.min.js</b></p></summary>
<p>
<details><summary><b>angular-1.2.16.js</b></p></summary>
<p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js</a></p>
<p>Path to vulnerable library: /src/main/webapp/js/angular/angular.js</p>
<p>
Dependency Hierarchy:
- :x: **angular-1.2.16.js** (Vulnerable Library)
</details>
<details><summary><b>angular-1.2.16.min.js</b></p></summary>
<p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js</a></p>
<p>Path to vulnerable library: /src/main/webapp/js/angular/angular.min.js</p>
<p>
Dependency Hierarchy:
- :x: **angular-1.2.16.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/developerone12/WebGoat-WhiteSource-Bolt/commit/c42e663814e4b88294ff90339ad577ca1afcf531">c42e663814e4b88294ff90339ad577ca1afcf531</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In AngularJS before 1.7.9 the function `merge()` could be tricked into adding or modifying properties of `Object.prototype` using a `__proto__` payload.
<p>Publish Date: 2019-11-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10768>CVE-2019-10768</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a">https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a</a></p>
<p>Release Date: 2019-11-19</p>
<p>Fix Resolution: angularjs - 1.7.8,1.7.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in angular js angular min js cve high severity vulnerability vulnerable libraries angular js angular min js angular js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to vulnerable library src main webapp js angular angular js dependency hierarchy x angular js vulnerable library angular min js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to vulnerable library src main webapp js angular angular min js dependency hierarchy x angular min js vulnerable library found in head commit a href found in base branch master vulnerability details in angularjs before the function merge could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angularjs step up your open source security game with whitesource | 0 |
476,317 | 13,736,755,215 | IssuesEvent | 2020-10-05 12:15:00 | wso2/streaming-integrator-tooling | https://api.github.com/repos/wso2/streaming-integrator-tooling | opened | ETL Wizard - Multiple ETL wizards simultaneously do not work | Priority/High Severity/Critical Type/Bug | **Description:**
When multiple ETL wizards are opened, they are not working properly.
**Affected Product Version:**
SI-Tooling 1.1.0
| 1.0 | ETL Wizard - Multiple ETL wizards simultaneously do not work - **Description:**
When multiple ETL wizards are opened, they are not working properly.
**Affected Product Version:**
SI-Tooling 1.1.0
| priority | etl wizard multiple etl wizards simultaneously do not work description when multiple etl wizards are opened they are not working properly affected product version si tooling | 1 |
6,001 | 8,674,599,627 | IssuesEvent | 2018-11-30 08:16:12 | FundacionParaguaya/MentorApp | https://api.github.com/repos/FundacionParaguaya/MentorApp | reopened | Validation on day and year of birth date | UX Requirement bug | - [ ] Validation of Day and Year in the dropdown its delayed - An error message only appears when the user clicks on the next input field - When the user clicks on the next field it hides the invalid message - thus creating a usability issue
- [ ] Validation on year is incorrect - This allows any number even those such as 4444 | 1.0 | Validation on day and year of birth date - - [ ] Validation of Day and Year in the dropdown its delayed - An error message only appears when the user clicks on the next input field - When the user clicks on the next field it hides the invalid message - thus creating a usability issue
- [ ] Validation on year is incorrect - This allows any number even those such as 4444 | non_priority | validation on day and year of birth date validation of day and year in the dropdown its delayed an error message only appears when the user clicks on the next input field when the user clicks on the next field it hides the invalid message thus creating a usability issue validation on year is incorrect this allows any number even those such as | 0 |
44,558 | 11,461,490,745 | IssuesEvent | 2020-02-07 12:04:08 | Qiskit/qiskit-aer | https://api.github.com/repos/Qiskit/qiskit-aer | closed | Pass user specified directories to the build system | build system enhancement | <!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues -->
### Informations
- **Qiskit AER version**: 0.1
- **Operating system**: OSX
### What is the current behavior?
The make system does not find the OpenBLAS libraries in their default location in OSX. Instead, the Apple Accelerate libs are always used.
### Steps to reproduce the problem
### What is the expected behavior?
The installer should find the OpenBLAS libs if they are installed in the default dirs.
### Suggested solutions
| 1.0 | Pass user specified directories to the build system - <!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues -->
### Informations
- **Qiskit AER version**: 0.1
- **Operating system**: OSX
### What is the current behavior?
The make system does not find the OpenBLAS libraries in their default location in OSX. Instead, the Apple Accelerate libs are always used.
### Steps to reproduce the problem
### What is the expected behavior?
The installer should find the OpenBLAS libs if they are installed in the default dirs.
### Suggested solutions
| non_priority | pass user specified directories to the build system informations qiskit aer version operating system osx what is the current behavior the make system does not find the openblas libraries in their default location in osx instead the apple accelerate libs are always used steps to reproduce the problem what is the expected behavior the installer should find the openblas libs if they are installed in the default dirs suggested solutions | 0 |
241,428 | 7,811,899,804 | IssuesEvent | 2018-06-12 11:44:09 | our-city-app/oca-backend | https://api.github.com/repos/our-city-app/oca-backend | closed | Missing translation in sidebar? | priority_critical state_verification type_bug | The configuration looks the same as Lochristi, but in Niel it's called "Regional/Local News" instead of "Regionaal/Lokaal Nieuws".
Are there other apps which have the same problem?


| 1.0 | Missing translation in sidebar? - The configuration looks the same as Lochristi, but in Niel it's called "Regional/Local News" instead of "Regionaal/Lokaal Nieuws".
Are there other apps which have the same problem?


| priority | missing translation in sidebar the configuration looks the same as lochristi but in niel it s called regional local news instead of regionaal lokaal nieuws are there other apps which have the same problem | 1 |
19,467 | 5,888,929,298 | IssuesEvent | 2017-05-17 11:32:32 | mrr0088/Python_DataClassification | https://api.github.com/repos/mrr0088/Python_DataClassification | closed | Funcionalidad Clasificación (Parte 3) | code enhancement | Conseguir enlazar los botones de Jinja con el código Python para ejecutar en backend la clasificación de la noticia. | 1.0 | Funcionalidad Clasificación (Parte 3) - Conseguir enlazar los botones de Jinja con el código Python para ejecutar en backend la clasificación de la noticia. | non_priority | funcionalidad clasificación parte conseguir enlazar los botones de jinja con el código python para ejecutar en backend la clasificación de la noticia | 0 |
117,430 | 9,935,037,769 | IssuesEvent | 2019-07-02 15:39:02 | somaliaims/Somali-AIMSUI | https://api.github.com/repos/somaliaims/Somali-AIMSUI | closed | If set start and end year of report to 2019 - no projects | july testing | This setting should show all projects active at any point in 2019 | 1.0 | If set start and end year of report to 2019 - no projects - This setting should show all projects active at any point in 2019 | non_priority | if set start and end year of report to no projects this setting should show all projects active at any point in | 0 |
763,379 | 26,754,702,661 | IssuesEvent | 2023-01-30 22:49:53 | nv-morpheus/Morpheus | https://api.github.com/repos/nv-morpheus/Morpheus | reopened | [FEA] Remove Direct `cuda-python` Dependency | feature request Priority 2 | **Is your feature request related to a problem? Please describe.**
Found in issue #250, the update of `cuda-python` from `11.7.0` to `11.7.1` has broken both CI and the conda build. The main cause is the switch from the conda package `cudatoolkit` to `cuda-toolkit`. The new package doesnt seem to include the necessary driver stubs that are used during linking.
To fix this issue, we needed to pin `cuda-python <=11.7.0`. However, this isnt a direct dependency of ours and should be handled by `rmm` and `cudf`.
**Describe the solution you'd like**
Ideally, once a clear strategy for how to handle the different conda packages `cudatoolkit` and `cuda-toolkit` emerges, RMM and cuDF would handle setting the correct `cuda-python` requirements. When that happens, we should remove our pinned versions of `cuda-python`.
**Describe alternatives you've considered**
Alternatively, we could add the `cuda-driver-dev` package which does include the CUDA Driver stubs. However, this doesnt solve the issue of which conda package we should be using.
| 1.0 | [FEA] Remove Direct `cuda-python` Dependency - **Is your feature request related to a problem? Please describe.**
Found in issue #250, the update of `cuda-python` from `11.7.0` to `11.7.1` has broken both CI and the conda build. The main cause is the switch from the conda package `cudatoolkit` to `cuda-toolkit`. The new package doesnt seem to include the necessary driver stubs that are used during linking.
To fix this issue, we needed to pin `cuda-python <=11.7.0`. However, this isnt a direct dependency of ours and should be handled by `rmm` and `cudf`.
**Describe the solution you'd like**
Ideally, once a clear strategy for how to handle the different conda packages `cudatoolkit` and `cuda-toolkit` emerges, RMM and cuDF would handle setting the correct `cuda-python` requirements. When that happens, we should remove our pinned versions of `cuda-python`.
**Describe alternatives you've considered**
Alternatively, we could add the `cuda-driver-dev` package which does include the CUDA Driver stubs. However, this doesnt solve the issue of which conda package we should be using.
| priority | remove direct cuda python dependency is your feature request related to a problem please describe found in issue the update of cuda python from to has broken both ci and the conda build the main cause is the switch from the conda package cudatoolkit to cuda toolkit the new package doesnt seem to include the necessary driver stubs that are used during linking to fix this issue we needed to pin cuda python however this isnt a direct dependency of ours and should be handled by rmm and cudf describe the solution you d like ideally once a clear strategy for how to handle the different conda packages cudatoolkit and cuda toolkit emerges rmm and cudf would handle setting the correct cuda python requirements when that happens we should remove our pinned versions of cuda python describe alternatives you ve considered alternatively we could add the cuda driver dev package which does include the cuda driver stubs however this doesnt solve the issue of which conda package we should be using | 1 |
713,400 | 24,527,460,056 | IssuesEvent | 2022-10-11 14:05:34 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | When reading the properties of `RpcException`, an unrelated exception may be thrown | kind/bug lang/C# area/api priority/P2 | I found this in 2.36.4 but it appears to still be an issue in the latest. When reading the properties of `RcpException`, an unrelated exception can be thrown. Specifically, `RpcException.Trailers` is of type `Metadata` and `Metadata.Value` can throw an exception from its getter.
This is generally bad practice, and is recommended against by Microsoft. The fact that exceptions are often serialized for logging exacerbates the issue, as this would obfuscate the initial exception. Whether or not to call this a "bug" is subjective since this appears to be by design, but I think it's clearly a flaw that should be addressed either way.
From a more general design perspective, it's generally a bad idea to throw an exception inside a property getter or setter. If the behavior involved could explicitly throw an exception, it's better to use a method instead of a property. Microsoft's guidance on this [can be found here](https://docs.microsoft.com/en-us/dotnet/standard/design-guidelines/property?redirectedfrom=MSDN). Relevant excerpt (emphasis mine):
> ❌ AVOID throwing exceptions from property getters.
>
> Property getters should be simple operations and should not have any preconditions. If a getter can throw an exception, **it should probably be redesigned to be a method**. Notice that this rule does not apply to indexers, where we do expect exceptions as a result of validating the arguments.
`RpcException.Trailers` is of type `Metadata`. `Metadata.Value` is a property as follows:
```
public string Value
{
get
{
GrpcPreconditions.CheckState(!IsBinary, "Cannot access string value of a binary metadata entry");
return value;
}
}
```
When the metadata is recognized as binary, an exception is thrown. This can interrupt exception handling for the RpcException and make it difficult to diagnose the issue. In our case, when our logger attempts to parse one of these during serialization, the resulting exception obfuscates the real exception.
One option would be to change it to a getter method such as `bool TryGetValue(out string value)`
- Prevents it from being read automatically by serializers
- Method name makes it clear to the caller that it may fail
- Fails safely (return `false` instead of throwing an exception)
| 1.0 | When reading the properties of `RpcException`, an unrelated exception may be thrown - I found this in 2.36.4 but it appears to still be an issue in the latest. When reading the properties of `RcpException`, an unrelated exception can be thrown. Specifically, `RpcException.Trailers` is of type `Metadata` and `Metadata.Value` can throw an exception from its getter.
This is generally bad practice, and is recommended against by Microsoft. The fact that exceptions are often serialized for logging exacerbates the issue, as this would obfuscate the initial exception. Whether or not to call this a "bug" is subjective since this appears to be by design, but I think it's clearly a flaw that should be addressed either way.
From a more general design perspective, it's generally a bad idea to throw an exception inside a property getter or setter. If the behavior involved could explicitly throw an exception, it's better to use a method instead of a property. Microsoft's guidance on this [can be found here](https://docs.microsoft.com/en-us/dotnet/standard/design-guidelines/property?redirectedfrom=MSDN). Relevant excerpt (emphasis mine):
> ❌ AVOID throwing exceptions from property getters.
>
> Property getters should be simple operations and should not have any preconditions. If a getter can throw an exception, **it should probably be redesigned to be a method**. Notice that this rule does not apply to indexers, where we do expect exceptions as a result of validating the arguments.
`RpcException.Trailers` is of type `Metadata`. `Metadata.Value` is a property as follows:
```
public string Value
{
get
{
GrpcPreconditions.CheckState(!IsBinary, "Cannot access string value of a binary metadata entry");
return value;
}
}
```
When the metadata is recognized as binary, an exception is thrown. This can interrupt exception handling for the RpcException and make it difficult to diagnose the issue. In our case, when our logger attempts to parse one of these during serialization, the resulting exception obfuscates the real exception.
One option would be to change it to a getter method such as `bool TryGetValue(out string value)`
- Prevents it from being read automatically by serializers
- Method name makes it clear to the caller that it may fail
- Fails safely (return `false` instead of throwing an exception)
| priority | when reading the properties of rpcexception an unrelated exception may be thrown i found this in but it appears to still be an issue in the latest when reading the properties of rcpexception an unrelated exception can be thrown specifically rpcexception trailers is of type metadata and metadata value can throw an exception from its getter this is generally bad practice and is recommended against by microsoft the fact that exceptions are often serialized for logging exacerbates the issue as this would obfuscate the initial exception whether or not to call this a bug is subjective since this appears to be by design but i think it s clearly a flaw that should be addressed either way from a more general design perspective it s generally a bad idea to throw an exception inside a property getter or setter if the behavior involved could explicitly throw an exception it s better to use a method instead of a property microsoft s guidance on this relevant excerpt emphasis mine ❌ avoid throwing exceptions from property getters property getters should be simple operations and should not have any preconditions if a getter can throw an exception it should probably be redesigned to be a method notice that this rule does not apply to indexers where we do expect exceptions as a result of validating the arguments rpcexception trailers is of type metadata metadata value is a property as follows public string value get grpcpreconditions checkstate isbinary cannot access string value of a binary metadata entry return value when the metadata is recognized as binary an exception is thrown this can interrupt exception handling for the rpcexception and make it difficult to diagnose the issue in our case when our logger attempts to parse one of these during serialization the resulting exception obfuscates the real exception one option would be to change it to a getter method such as bool trygetvalue out string value prevents it from being read automatically by serializers method name makes it clear to the caller that it may fail fails safely return false instead of throwing an exception | 1 |
410,506 | 11,992,452,507 | IssuesEvent | 2020-04-08 10:08:15 | PostHog/posthog | https://api.github.com/repos/PostHog/posthog | opened | Add SMTP credentials in UX | enhancement priority-6 | **Is your feature request related to a problem? Please describe.**
For users to have a password reset, they currently have to set env variables. This isn't user friendly.
**Describe the solution you'd like**
In the settings page of the app, be able to set:
```
EMAIL_HOST
EMAIL_PORT
EMAIL_HOST_USER
EMAIL_HOST_PASSWORD
EMAIL_USE_TLS
EMAIL_USE_SSL
DEFAULT_FROM_EMAIL
```
**Describe alternatives you've considered**
Set environment variables.
| 1.0 | Add SMTP credentials in UX - **Is your feature request related to a problem? Please describe.**
For users to have a password reset, they currently have to set env variables. This isn't user friendly.
**Describe the solution you'd like**
In the settings page of the app, be able to set:
```
EMAIL_HOST
EMAIL_PORT
EMAIL_HOST_USER
EMAIL_HOST_PASSWORD
EMAIL_USE_TLS
EMAIL_USE_SSL
DEFAULT_FROM_EMAIL
```
**Describe alternatives you've considered**
Set environment variables.
| priority | add smtp credentials in ux is your feature request related to a problem please describe for users to have a password reset they currently have to set env variables this isn t user friendly describe the solution you d like in the settings page of the app be able to set email host email port email host user email host password email use tls email use ssl default from email describe alternatives you ve considered set environment variables | 1 |
290,522 | 25,073,326,738 | IssuesEvent | 2022-11-07 13:49:56 | jackmcgrath20088629/notes-app-2 | https://api.github.com/repos/jackmcgrath20088629/notes-app-2 | closed | New Functionality - Update a Note in the Notes Collection | enhancement test driven development | This new functionality should allow the user to update a note in the notes collection. The flow should be:
- All notes [along with the index number in the ArrayList] should be printed to the console.
- The user should be prompted to enter the index number of the note they wish to update.
- The user is then asked to enter new data for the title, priority, and category. Note: Archive status will be managed in later versions.
- The note is then updated - the user is informed of whether the update was successful or not (e.g if the user entered an index number that was not valid, this would result in a note not being updated.)
The JUnit tests associated with this functionality should be completed as part of this issue. | 1.0 | New Functionality - Update a Note in the Notes Collection - This new functionality should allow the user to update a note in the notes collection. The flow should be:
- All notes [along with the index number in the ArrayList] should be printed to the console.
- The user should be prompted to enter the index number of the note they wish to update.
- The user is then asked to enter new data for the title, priority, and category. Note: Archive status will be managed in later versions.
- The note is then updated - the user is informed of whether the update was successful or not (e.g if the user entered an index number that was not valid, this would result in a note not being updated.)
The JUnit tests associated with this functionality should be completed as part of this issue. | non_priority | new functionality update a note in the notes collection this new functionality should allow the user to update a note in the notes collection the flow should be all notes should be printed to the console the user should be prompted to enter the index number of the note they wish to update the user is then asked to enter new data for the title priority and category note archive status will be managed in later versions the note is then updated the user is informed of whether the update was successful or not e g if the user entered an index number that was not valid this would result in a note not being updated the junit tests associated with this functionality should be completed as part of this issue | 0 |
733,306 | 25,299,828,766 | IssuesEvent | 2022-11-17 09:52:40 | software-mansion/starknet.py | https://api.github.com/repos/software-mansion/starknet.py | closed | Investigate if methods using token_address must be updated | question priority: high | It seems like something related to token address was changed in general config: perhaps methods like get_balance need updating.

| 1.0 | Investigate if methods using token_address must be updated - It seems like something related to token address was changed in general config: perhaps methods like get_balance need updating.

| priority | investigate if methods using token address must be updated it seems like something related to token address was changed in general config perhaps methods like get balance need updating | 1 |
6,870 | 15,686,961,734 | IssuesEvent | 2021-03-25 13:09:24 | nikolagsiderov/APPartment | https://api.github.com/repos/nikolagsiderov/APPartment | closed | Introduce Areas in routing | General Architecture General UI | During development of #271 from @fr0wBtW, I figured it would be best if we separated the whole controllers' structure in areas... | 1.0 | Introduce Areas in routing - During development of #271 from @fr0wBtW, I figured it would be best if we separated the whole controllers' structure in areas... | non_priority | introduce areas in routing during development of from i figured it would be best if we separated the whole controllers structure in areas | 0 |
33,982 | 2,774,133,775 | IssuesEvent | 2015-05-04 05:21:38 | galenp/PROPERTYCOMPASS.ISSUES | https://api.github.com/repos/galenp/PROPERTYCOMPASS.ISSUES | closed | Payment details page - Credit Card number and Total field | 1.3 Type-Suggestion 2.2 Priority-Low 3.5 Status-Resolved | Please refer attached image
thanks,

| 1.0 | Payment details page - Credit Card number and Total field - Please refer attached image
thanks,

| priority | payment details page credit card number and total field please refer attached image thanks | 1 |
672,293 | 22,820,549,178 | IssuesEvent | 2022-07-12 01:32:27 | LuanRT/YouTube.js | https://api.github.com/repos/LuanRT/YouTube.js | opened | Properly organize parser classes | enhancement good first issue priority: low | ### Describe your suggestion
The parser is mostly complete at this point, but I reckon the `classes` folder is a it messy. With f924a394091feda6f38168eb84ccb38ab1c701ce we should be able to put Renderers/DA classes in specific folders — that would allow us to find problematic code more easily. I already started doing that but only with files related to recently added features such as `LiveChat` and `Comments`
### Other details
_No response_
### Checklist
- [X] I am running the latest version.
- [X] I checked the documentation and found no answer.
- [X] I have searched the existing issues and made sure this is not a duplicate.
- [X] I have provided sufficient information. | 1.0 | Properly organize parser classes - ### Describe your suggestion
The parser is mostly complete at this point, but I reckon the `classes` folder is a it messy. With f924a394091feda6f38168eb84ccb38ab1c701ce we should be able to put Renderers/DA classes in specific folders — that would allow us to find problematic code more easily. I already started doing that but only with files related to recently added features such as `LiveChat` and `Comments`
### Other details
_No response_
### Checklist
- [X] I am running the latest version.
- [X] I checked the documentation and found no answer.
- [X] I have searched the existing issues and made sure this is not a duplicate.
- [X] I have provided sufficient information. | priority | properly organize parser classes describe your suggestion the parser is mostly complete at this point but i reckon the classes folder is a it messy with we should be able to put renderers da classes in specific folders — that would allow us to find problematic code more easily i already started doing that but only with files related to recently added features such as livechat and comments other details no response checklist i am running the latest version i checked the documentation and found no answer i have searched the existing issues and made sure this is not a duplicate i have provided sufficient information | 1 |
4,491 | 6,996,111,643 | IssuesEvent | 2017-12-15 22:26:44 | amcnulty/Whoop-It-Up | https://api.github.com/repos/amcnulty/Whoop-It-Up | closed | Mobile User page | Requirement | Basic mobile layout for the landing page
- bootstrap 4
- modify grunt file (if needed) to compile scss
- basic JS, no animation needed yet
- basic CSS/JS for avatar selection
- "Update Profile" button require backend support | 1.0 | Mobile User page - Basic mobile layout for the landing page
- bootstrap 4
- modify grunt file (if needed) to compile scss
- basic JS, no animation needed yet
- basic CSS/JS for avatar selection
- "Update Profile" button require backend support | non_priority | mobile user page basic mobile layout for the landing page bootstrap modify grunt file if needed to compile scss basic js no animation needed yet basic css js for avatar selection update profile button require backend support | 0 |
53,389 | 13,161,489,515 | IssuesEvent | 2020-08-10 19:38:32 | mozilla-mobile/fenix | https://api.github.com/repos/mozilla-mobile/fenix | closed | FNX-4726 ⁃ Dependency substitution for android-components | E5 P2 eng:build | This will allow us to easily develop against a local checkout of `android-components`.
Currently blocked on a-c moving to AndroidX (https://github.com/mozilla-mobile/android-components/issues/842). Also, see similar r-b issue: https://github.com/mozilla-mobile/reference-browser/issues/365
| 1.0 | FNX-4726 ⁃ Dependency substitution for android-components - This will allow us to easily develop against a local checkout of `android-components`.
Currently blocked on a-c moving to AndroidX (https://github.com/mozilla-mobile/android-components/issues/842). Also, see similar r-b issue: https://github.com/mozilla-mobile/reference-browser/issues/365
| non_priority | fnx ⁃ dependency substitution for android components this will allow us to easily develop against a local checkout of android components currently blocked on a c moving to androidx also see similar r b issue | 0 |
215,825 | 16,707,166,899 | IssuesEvent | 2021-06-09 11:24:55 | ita-social-projects/EventsExpress | https://api.github.com/repos/ita-social-projects/EventsExpress | opened | Verify that the event owner can create an online event and add a link to that event | test | **Date Time**
06-09-2021 15:15
**Priority**
Medium
**Description**
This test verifies if the event owner can provide a link to the online event.
**Pre-requisites**
1. Move to https://eventsexpress-test.azurewebsites.net/home/events?page=1&status=active
2. Authorize as a user (E-Mail: eventsexpress.testuser3@gmail.com Password: 1qaz2wsx1qaz2wsx)
3. 'Home' page is opened
**Test Cases**
| S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated |
|:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:|
| 1<br> | Click on the 'Add event' button<br> | <br> | The ‘Edit event draft’ page is open<br> | <br> | Pass<br> | <br> |
| 2<br> | Fill in the 'Title', 'Max Count Of Participants', 'Date', 'Description' and ' Hashtags' fields<br> | e.g. 'Нова онлайн подія', 35, 'Подія для провірки встановлення посилання на онлайн подію', '#travel'<br> | <br> | <br> | Pass<br> | <br> |
| 3<br> | Select the 'Online' radio button<br> | <br> | The input field appeared under the 'Online' radio button<br> | <br> | Pass<br> | <br> |
| 4<br> | Insert the event link in the input field<br> | e.g. https://www.google.com.ua<br> | <br> | <br> | Pass<br> | <br> |
| 5<br> | Сlick 'Save' button<br> | <br> | In the lower left corner of the page a message about the successful saving of the event appears<br> | <br> | Pass<br> | <br> |
**Environment:**
- OS: [Windows 10 (64 bit)]
- Browser [Chrome Version 91.0.4472.77]
- Version [20210609.3]
| 1.0 | Verify that the event owner can create an online event and add a link to that event - **Date Time**
06-09-2021 15:15
**Priority**
Medium
**Description**
This test verifies if the event owner can provide a link to the online event.
**Pre-requisites**
1. Move to https://eventsexpress-test.azurewebsites.net/home/events?page=1&status=active
2. Authorize as a user (E-Mail: eventsexpress.testuser3@gmail.com Password: 1qaz2wsx1qaz2wsx)
3. 'Home' page is opened
**Test Cases**
| S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated |
|:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:|
| 1<br> | Click on the 'Add event' button<br> | <br> | The ‘Edit event draft’ page is open<br> | <br> | Pass<br> | <br> |
| 2<br> | Fill in the 'Title', 'Max Count Of Participants', 'Date', 'Description' and ' Hashtags' fields<br> | e.g. 'Нова онлайн подія', 35, 'Подія для провірки встановлення посилання на онлайн подію', '#travel'<br> | <br> | <br> | Pass<br> | <br> |
| 3<br> | Select the 'Online' radio button<br> | <br> | The input field appeared under the 'Online' radio button<br> | <br> | Pass<br> | <br> |
| 4<br> | Insert the event link in the input field<br> | e.g. https://www.google.com.ua<br> | <br> | <br> | Pass<br> | <br> |
| 5<br> | Сlick 'Save' button<br> | <br> | In the lower left corner of the page a message about the successful saving of the event appears<br> | <br> | Pass<br> | <br> |
**Environment:**
- OS: [Windows 10 (64 bit)]
- Browser [Chrome Version 91.0.4472.77]
- Version [20210609.3]
| non_priority | verify that the event owner can create an online event and add a link to that event date time priority medium description this test verifies if the event owner can provide a link to the online event pre requisites move to authorize as a user e mail eventsexpress gmail com password home page is opened test cases s action test data expected result actual result p f automated click on the add event button the ‘edit event draft’ page is open pass fill in the title max count of participants date description and hashtags fields e g нова онлайн подія подія для провірки встановлення посилання на онлайн подію travel pass select the online radio button the input field appeared under the online radio button pass insert the event link in the input field e g pass сlick save button in the lower left corner of the page a message about the successful saving of the event appears pass environment os browser version | 0 |
329,621 | 10,022,309,203 | IssuesEvent | 2019-07-16 16:23:10 | strapi/strapi | https://api.github.com/repos/strapi/strapi | closed | Authenticating a user with Google is throwing 500 on the 2nd login | priority: medium status: have to reproduce type: bug 🐛 | <!-- ⚠️ If you do not respect this template your issue will be closed. -->
<!-- =============================================================================== -->
<!-- ⚠️ If you are not using the current Strapi release, you will be asked to update. -->
<!-- Please see the wiki for guides on upgrading to the latest release. -->
<!-- =============================================================================== -->
<!-- ⚠️ Make sure to browse the opened and closed issues before submitting your issue. -->
<!-- ⚠️ Before writing your issue make sure you are using:-->
<!-- Node 10.x.x -->
<!-- npm 6.x.x -->
<!-- The latest version of Strapi. -->
**Informations**
- **Node.js version**: 10.13.0
- **npm version**: 6.4.1
- **Strapi version**: v3.0.0-alpha.14.5
- **Database**: mongodb 3.6.5
- **Operating system**: Win10
**What is the current behavior?**
After the first successful authentication with google, additional attempts to authenticate, at least within the next minutes, fail at redirecting to GET /auth/google/callback as long as the user remains signed in with google in the browser.
Response is 500 at http://localhost:4200/auth/callback/google?error%5Berror%5D=invalid_grant&error%5Berror_description%5D=Malformed%20auth%20code.
Signing off from Google fixes this issue and returns jwt and user response after redirecting to GET /auth/google/callback
**Steps to reproduce the problem**
Authenticate with google
Stay signed in with google
Authenticate again with google
**What is the expected behavior?**
After the first authentication subsequent authentication attempts should also issue a new token.
**Suggested solutions**
I'm not quite sure why this happens since the initial redirect from accounts.google.com/o/oauth2/auth to /connect/google/callback looks pretty much the same on the first and subsequent calls, only difference i could recognize was that the first redirect is encoded while the next ones are not
First attempt, working:
http://localhost:1337/connect/google/callback?code=4%2FogD_bJ10kjNs7l8gqTF25hLsYQPUU-rkatS6jK5shBXcw-lLN0wlJpbSDbyx8zFP2yuDyVLDA1ScgSxLaZxxxxx&scope=openid+email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fuserinfo.email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fplus.me
Next attempt, failing:
http://localhost:1337/connect/google/callback?code=4/ogCWJEaBayzfL2PoY4NJ3rL-11Wu4FNjrPOxRlFG003qPuGCy7nV6qCnlX1vvh_Dt1cUmJZrFn5ESxxxxx&scope=openid+email+https://www.googleapis.com/auth/plus.me+https://www.googleapis.com/auth/userinfo.email | 1.0 | Authenticating a user with Google is throwing 500 on the 2nd login - <!-- ⚠️ If you do not respect this template your issue will be closed. -->
<!-- =============================================================================== -->
<!-- ⚠️ If you are not using the current Strapi release, you will be asked to update. -->
<!-- Please see the wiki for guides on upgrading to the latest release. -->
<!-- =============================================================================== -->
<!-- ⚠️ Make sure to browse the opened and closed issues before submitting your issue. -->
<!-- ⚠️ Before writing your issue make sure you are using:-->
<!-- Node 10.x.x -->
<!-- npm 6.x.x -->
<!-- The latest version of Strapi. -->
**Informations**
- **Node.js version**: 10.13.0
- **npm version**: 6.4.1
- **Strapi version**: v3.0.0-alpha.14.5
- **Database**: mongodb 3.6.5
- **Operating system**: Win10
**What is the current behavior?**
After the first successful authentication with google, additional attempts to authenticate, at least within the next minutes, fail at redirecting to GET /auth/google/callback as long as the user remains signed in with google in the browser.
Response is 500 at http://localhost:4200/auth/callback/google?error%5Berror%5D=invalid_grant&error%5Berror_description%5D=Malformed%20auth%20code.
Signing off from Google fixes this issue and returns jwt and user response after redirecting to GET /auth/google/callback
**Steps to reproduce the problem**
Authenticate with google
Stay signed in with google
Authenticate again with google
**What is the expected behavior?**
After the first authentication subsequent authentication attempts should also issue a new token.
**Suggested solutions**
I'm not quite sure why this happens since the initial redirect from accounts.google.com/o/oauth2/auth to /connect/google/callback looks pretty much the same on the first and subsequent calls, only difference i could recognize was that the first redirect is encoded while the next ones are not
First attempt, working:
http://localhost:1337/connect/google/callback?code=4%2FogD_bJ10kjNs7l8gqTF25hLsYQPUU-rkatS6jK5shBXcw-lLN0wlJpbSDbyx8zFP2yuDyVLDA1ScgSxLaZxxxxx&scope=openid+email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fuserinfo.email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fplus.me
Next attempt, failing:
http://localhost:1337/connect/google/callback?code=4/ogCWJEaBayzfL2PoY4NJ3rL-11Wu4FNjrPOxRlFG003qPuGCy7nV6qCnlX1vvh_Dt1cUmJZrFn5ESxxxxx&scope=openid+email+https://www.googleapis.com/auth/plus.me+https://www.googleapis.com/auth/userinfo.email | priority | authenticating a user with google is throwing on the login informations node js version npm version strapi version alpha database mongodb operating system what is the current behavior after the first successful authentication with google additional attempts to authenticate at least within the next minutes fail at redirecting to get auth google callback as long as the user remains signed in with google in the browser response is at signing off from google fixes this issue and returns jwt and user response after redirecting to get auth google callback steps to reproduce the problem authenticate with google stay signed in with google authenticate again with google what is the expected behavior after the first authentication subsequent authentication attempts should also issue a new token suggested solutions i m not quite sure why this happens since the initial redirect from accounts google com o auth to connect google callback looks pretty much the same on the first and subsequent calls only difference i could recognize was that the first redirect is encoded while the next ones are not first attempt working next attempt failing | 1 |
260,463 | 22,622,957,629 | IssuesEvent | 2022-06-30 08:12:42 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Release 4.3.5 | release test/4.3.5 | Main issue gathering all the tasks regarding the preparation and validation of Wazuh version 4.3.5 release.
Close the issue only after the GA release.
## Release info
| | |
|---------------------------------------|------------------------------------------------------|
| Release name| Wazuh v4.3.5 |
| Tag | https://github.com/wazuh/wazuh/tree/v4.3.5 |
| Branch | https://github.com/wazuh/wazuh/tree/4.3 |
| Changelog | https://github.com/wazuh/wazuh/blob/4.3/CHANGELOG.md |
| Latest release candidate (GA version) | N/A | |
## Release candidates
As usual, in order to publish, one or more release candidate versions are generated and fully tested, in the links below you can check the status and evolution of each version.
| Version | Issue | Tag | Status |
|---------|-------|-------------------------------------------------|-----------|
| RC1 | #13966 | [v4.3.5-rc1](https://github.com/wazuh/wazuh/tree/v4.3.5-rc1) | Completed |
- _In progress: Fixes are being implemented._
- _In review: All the fixes have been implemented. The RC is now tagged and the QA process kicks off._
- _Closed: Fixes and QA process has finished._
## Tests
All the tests results for this release can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1)
## Fixes/Known issues
All the fixes and the issues detected in the testing phase that won't be solved in this version can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1)
| 1.0 | Release 4.3.5 - Main issue gathering all the tasks regarding the preparation and validation of Wazuh version 4.3.5 release.
Close the issue only after the GA release.
## Release info
| | |
|---------------------------------------|------------------------------------------------------|
| Release name| Wazuh v4.3.5 |
| Tag | https://github.com/wazuh/wazuh/tree/v4.3.5 |
| Branch | https://github.com/wazuh/wazuh/tree/4.3 |
| Changelog | https://github.com/wazuh/wazuh/blob/4.3/CHANGELOG.md |
| Latest release candidate (GA version) | N/A | |
## Release candidates
As usual, in order to publish, one or more release candidate versions are generated and fully tested, in the links below you can check the status and evolution of each version.
| Version | Issue | Tag | Status |
|---------|-------|-------------------------------------------------|-----------|
| RC1 | #13966 | [v4.3.5-rc1](https://github.com/wazuh/wazuh/tree/v4.3.5-rc1) | Completed |
- _In progress: Fixes are being implemented._
- _In review: All the fixes have been implemented. The RC is now tagged and the QA process kicks off._
- _Closed: Fixes and QA process has finished._
## Tests
All the tests results for this release can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1)
## Fixes/Known issues
All the fixes and the issues detected in the testing phase that won't be solved in this version can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1)
| non_priority | release main issue gathering all the tasks regarding the preparation and validation of wazuh version release close the issue only after the ga release release info release name wazuh tag branch changelog latest release candidate ga version n a release candidates as usual in order to publish one or more release candidate versions are generated and fully tested in the links below you can check the status and evolution of each version version issue tag status completed in progress fixes are being implemented in review all the fixes have been implemented the rc is now tagged and the qa process kicks off closed fixes and qa process has finished tests all the tests results for this release can be found in the fixes known issues all the fixes and the issues detected in the testing phase that won t be solved in this version can be found in the | 0 |
21,378 | 29,202,228,594 | IssuesEvent | 2023-05-21 00:36:56 | devssa/onde-codar-em-salvador | https://api.github.com/repos/devssa/onde-codar-em-salvador | closed | [Remoto] Product Manager na Coodesh | SALVADOR GESTÃO DE PROJETOS JIRA REQUISITOS REMOTO PROCESSOS GITHUB UMA POWER BI APIs NEGÓCIOS PRODUCT MANAGER Stale | ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>Techsocial</strong> está em busca de <strong><ins>Product Manager</ins></strong> para compor seu time!</p>
<p></p>
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas. A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. </p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Entendimento dos módulos do nossa Plataforma; </li>
<li>Conhecimento dos nossos negócios e suas peculiaridades;</li>
<li>Identificação de gaps e oportunidades de melhoria; </li>
<li>Auxílio no mapeamento do processo e levantamento de requisitos de mudanças;</li>
<li>Confecção de tickets de melhoria alinhado ao time de processos; </li>
<li>Apresentação e revisão de tickets junto ao time de desenvolvimento; </li>
<li>Definição de prioridades de implementação; </li>
<li>Identificação de interfaces do(s) módulo(s) que representa para que requisitos e regras não sejam alterados sem alinhamento e definição com outros módulos, etc.</li>
</ul>
## Techsocial:
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas.</p>
<p>A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. Nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas.</p><a href='https://coodesh.com/empresas/techsocial-tecnologia-e-consultoria-ltda'>Veja mais no site</a>
## Habilidades:
- JIRA
- Análise de requisitos
- Gestão e Negociação com Cliente
## Local:
100% Remoto
## Requisitos:
- Conhecimento em Levantamento de requisitos;
- Mapeamento de Processos;
- Experiência em Gestão de Projetos;
- Experiência em Gestão de Produtos.
## Diferenciais:
- Power BI;
- Experíência na Plataforma Jira.
## Benefícios:
- Convênio Médico;
- Trabalho Remoto.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Product Manager na Techsocial](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Gestão em TI | 1.0 | [Remoto] Product Manager na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>Techsocial</strong> está em busca de <strong><ins>Product Manager</ins></strong> para compor seu time!</p>
<p></p>
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas. A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. </p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Entendimento dos módulos do nossa Plataforma; </li>
<li>Conhecimento dos nossos negócios e suas peculiaridades;</li>
<li>Identificação de gaps e oportunidades de melhoria; </li>
<li>Auxílio no mapeamento do processo e levantamento de requisitos de mudanças;</li>
<li>Confecção de tickets de melhoria alinhado ao time de processos; </li>
<li>Apresentação e revisão de tickets junto ao time de desenvolvimento; </li>
<li>Definição de prioridades de implementação; </li>
<li>Identificação de interfaces do(s) módulo(s) que representa para que requisitos e regras não sejam alterados sem alinhamento e definição com outros módulos, etc.</li>
</ul>
## Techsocial:
<p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas.</p>
<p>A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. Nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas.</p><a href='https://coodesh.com/empresas/techsocial-tecnologia-e-consultoria-ltda'>Veja mais no site</a>
## Habilidades:
- JIRA
- Análise de requisitos
- Gestão e Negociação com Cliente
## Local:
100% Remoto
## Requisitos:
- Conhecimento em Levantamento de requisitos;
- Mapeamento de Processos;
- Experiência em Gestão de Projetos;
- Experiência em Gestão de Produtos.
## Diferenciais:
- Power BI;
- Experíência na Plataforma Jira.
## Benefícios:
- Convênio Médico;
- Trabalho Remoto.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Product Manager na Techsocial](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Remoto
#### Categoria
Gestão em TI | non_priority | product manager na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a techsocial está em busca de product manager para compor seu time somos uma empresa de soluções tecnológicas que busca transformar os dados e informações de nossos clientes em resultados evoluímos a partir de consultoria em gestão empresarial somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas a tech é uma empresa inovadora desenvolvemos e aportamos inteligência em softwares aplicativos rpas apis entre outras soluções digitais nbsp responsabilidades entendimento dos módulos do nossa plataforma nbsp conhecimento dos nossos negócios e suas peculiaridades identificação de gaps e oportunidades de melhoria nbsp auxílio no mapeamento do processo e levantamento de requisitos de mudanças confecção de tickets de melhoria alinhado ao time de processos nbsp apresentação e revisão de tickets junto ao time de desenvolvimento nbsp definição de prioridades de implementação nbsp identificação de interfaces do s módulo s que representa para que requisitos e regras não sejam alterados sem alinhamento e definição com outros módulos etc techsocial somos uma empresa de soluções tecnológicas que busca transformar os dados e informações de nossos clientes em resultados evoluímos a partir de consultoria em gestão empresarial somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas a tech é uma empresa inovadora desenvolvemos e aportamos inteligência em softwares aplicativos rpas apis entre outras soluções digitais nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas habilidades jira análise de requisitos gestão e negociação com cliente local remoto requisitos conhecimento em levantamento de requisitos mapeamento de processos experiência em gestão de projetos experiência em gestão de produtos diferenciais power bi experíência na plataforma jira benefícios convênio médico trabalho remoto como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria gestão em ti | 0 |
517,461 | 15,013,967,458 | IssuesEvent | 2021-02-01 05:39:45 | TerriaJS/nationalmap | https://api.github.com/repos/TerriaJS/nationalmap | closed | NationalMap - Jan 2021 release (v8) | High priority | **HIGH PRIORITY:**
- [ ] AREMI catalogue transfer to NatMap https://github.com/TerriaJS/aremi-natmap/issues/437 - Data transfer mapping is [HERE ](https://csiroau.sharepoint.com/:w:/r/sites/WebDataPlatformsgroup/Shared%20Documents/General/Terria/AREMI/Data/Decommissioning%20-%20data%20transfer%20to%20NatMap/AREMI%20data%20transfer%20to%20NatMap%20-%20mapping%20to%20catalogue.docx?d=w807583b6b8684cfa85cbc4fefa064729&csf=1&web=1&e=DqTrGg)
V8 porting related issues for AREMI data:
- Token implementation for WA services: https://github.com/TerriaJS/aremi-natmap/issues/437
- ISF data - investigate CSS custom sheets and vector data handling: https://github.com/TerriaJS/aremi-natmap/issues/439
- Traffic volumes - NSW transport API - https://github.com/TerriaJS/aremi-natmap/issues/440
- AEMO data - csv polling - feature info disappears after 1 cycle / polling - insert ticket https://github.com/terriajs/terriajs/issues/4768
**MEDIUM PRIORITY:**
- [ ] Help page (under About) needs to be deleted
- [ ] update some GA layers: https://github.com/TerriaJS/nationalmap/issues/986
- [ ] NationalMap catalogue https://github.com/TerriaJS/nationalmap/issues/985 - needs SDMX fixes and https://github.com/TerriaJS/terriajs/pull/5085 merged
- [ ] YourDataYourRegions https://github.com/TerriaJS/terriajs/issues/4861
- [ ] CKAN support for data.gov.au - https://github.com/terriajs/nationalmap/issues/987 (show data with no filter)
- [ ] MAGDA API support for data.gov.au - search?
- [ ] MAGDA integration _ https://github.com/TerriaJS/nationalmap/issues/981 and https://github.com/TerriaJS/TerriaMap/issues/489
- [ ] GeoJSON
- [ ] KML/KMZ
- [ ] Boundaries: e.g. http://geoserver.nationalmap.nicta.com.au/admin_bnds_abs/ows
- [ ] region mapping
- [ ] vector tiles
- [ ] csw https://github.com/TerriaJS/terriajs/issues/4693
**LOW PRIORITY**
- [ ] NSW decimal issue: https://github.com/TerriaJS/nationalmap/issues/980
- [ ] Update Mobile black spot data position in catalogue - https://github.com/TerriaJS/nationalmap/issues/979
| 1.0 | NationalMap - Jan 2021 release (v8) - **HIGH PRIORITY:**
- [ ] AREMI catalogue transfer to NatMap https://github.com/TerriaJS/aremi-natmap/issues/437 - Data transfer mapping is [HERE ](https://csiroau.sharepoint.com/:w:/r/sites/WebDataPlatformsgroup/Shared%20Documents/General/Terria/AREMI/Data/Decommissioning%20-%20data%20transfer%20to%20NatMap/AREMI%20data%20transfer%20to%20NatMap%20-%20mapping%20to%20catalogue.docx?d=w807583b6b8684cfa85cbc4fefa064729&csf=1&web=1&e=DqTrGg)
V8 porting related issues for AREMI data:
- Token implementation for WA services: https://github.com/TerriaJS/aremi-natmap/issues/437
- ISF data - investigate CSS custom sheets and vector data handling: https://github.com/TerriaJS/aremi-natmap/issues/439
- Traffic volumes - NSW transport API - https://github.com/TerriaJS/aremi-natmap/issues/440
- AEMO data - csv polling - feature info disappears after 1 cycle / polling - insert ticket https://github.com/terriajs/terriajs/issues/4768
**MEDIUM PRIORITY:**
- [ ] Help page (under About) needs to be deleted
- [ ] update some GA layers: https://github.com/TerriaJS/nationalmap/issues/986
- [ ] NationalMap catalogue https://github.com/TerriaJS/nationalmap/issues/985 - needs SDMX fixes and https://github.com/TerriaJS/terriajs/pull/5085 merged
- [ ] YourDataYourRegions https://github.com/TerriaJS/terriajs/issues/4861
- [ ] CKAN support for data.gov.au - https://github.com/terriajs/nationalmap/issues/987 (show data with no filter)
- [ ] MAGDA API support for data.gov.au - search?
- [ ] MAGDA integration _ https://github.com/TerriaJS/nationalmap/issues/981 and https://github.com/TerriaJS/TerriaMap/issues/489
- [ ] GeoJSON
- [ ] KML/KMZ
- [ ] Boundaries: e.g. http://geoserver.nationalmap.nicta.com.au/admin_bnds_abs/ows
- [ ] region mapping
- [ ] vector tiles
- [ ] csw https://github.com/TerriaJS/terriajs/issues/4693
**LOW PRIORITY**
- [ ] NSW decimal issue: https://github.com/TerriaJS/nationalmap/issues/980
- [ ] Update Mobile black spot data position in catalogue - https://github.com/TerriaJS/nationalmap/issues/979
| priority | nationalmap jan release high priority aremi catalogue transfer to natmap data transfer mapping is porting related issues for aremi data token implementation for wa services isf data investigate css custom sheets and vector data handling traffic volumes nsw transport api aemo data csv polling feature info disappears after cycle polling insert ticket medium priority help page under about needs to be deleted update some ga layers nationalmap catalogue needs sdmx fixes and merged yourdatayourregions ckan support for data gov au show data with no filter magda api support for data gov au search magda integration and geojson kml kmz boundaries e g region mapping vector tiles csw low priority nsw decimal issue update mobile black spot data position in catalogue | 1 |
268,400 | 23,366,318,859 | IssuesEvent | 2022-08-10 15:38:56 | cosmos/ibc-go | https://api.github.com/repos/cosmos/ibc-go | opened | E2E: Timeout IBC transfer | testing | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Word of caution: poorly thought-out proposals may be rejected
v without deliberation
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
Add a test which times out IBC transfers
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged/assigned
| 1.0 | E2E: Timeout IBC transfer - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Word of caution: poorly thought-out proposals may be rejected
v without deliberation
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
Add a test which times out IBC transfers
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate contributors tagged/assigned
| non_priority | timeout ibc transfer ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v word of caution poorly thought out proposals may be rejected v without deliberation ☺ summary add a test which times out ibc transfers for admin use not duplicate issue appropriate labels applied appropriate contributors tagged assigned | 0 |
40,304 | 5,199,405,735 | IssuesEvent | 2017-01-23 20:47:51 | azavea/cac-tripplanner | https://api.github.com/repos/azavea/cac-tripplanner | opened | Explore nav button has no effect on Learn page | bug redesign | On staging, go to the Learn page. Click the Explore link in the top nav. Nothing happens. Should go to Explore mode. | 1.0 | Explore nav button has no effect on Learn page - On staging, go to the Learn page. Click the Explore link in the top nav. Nothing happens. Should go to Explore mode. | non_priority | explore nav button has no effect on learn page on staging go to the learn page click the explore link in the top nav nothing happens should go to explore mode | 0 |
3,925 | 6,845,689,064 | IssuesEvent | 2017-11-13 09:17:46 | openvstorage/volumedriver | https://api.github.com/repos/openvstorage/volumedriver | closed | Redirected volume removal times out and leaves stale metadata | priority_normal process_wontfix type_bug | Scenario:
- config:
- redirect timeout 5 secs
- backend timeout 15 secs
- volume V is running on node N
- SCO write W to the backend is happening in the background
- volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish
- W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires
=> unlink call times out on N
| 1.0 | Redirected volume removal times out and leaves stale metadata - Scenario:
- config:
- redirect timeout 5 secs
- backend timeout 15 secs
- volume V is running on node N
- SCO write W to the backend is happening in the background
- volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish
- W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires
=> unlink call times out on N
| non_priority | redirected volume removal times out and leaves stale metadata scenario config redirect timeout secs backend timeout secs volume v is running on node n sco write w to the backend is happening in the background volume removal is started on node m redirect to node n removal needs to wait for w to finish w encounters backend issues finishes unsuccessfully when the backend timeout fires unlink call times out on n | 0 |
23,626 | 12,042,422,297 | IssuesEvent | 2020-04-14 10:32:04 | flutter/gallery | https://api.github.com/repos/flutter/gallery | closed | [Performance] Use deferred loading to improve performance | performance | ## Acceptance Criteria
* The studies load asynchronously after the initial page load
* Load non necessary locales asynchronously
https://dart.dev/guides/language/language-tour#deferred-loading
Migrated from https://github.com/material-components/material-components-flutter-gallery/issues/571 | True | [Performance] Use deferred loading to improve performance - ## Acceptance Criteria
* The studies load asynchronously after the initial page load
* Load non necessary locales asynchronously
https://dart.dev/guides/language/language-tour#deferred-loading
Migrated from https://github.com/material-components/material-components-flutter-gallery/issues/571 | non_priority | use deferred loading to improve performance acceptance criteria the studies load asynchronously after the initial page load load non necessary locales asynchronously migrated from | 0 |
35,025 | 9,531,211,295 | IssuesEvent | 2019-04-29 15:29:58 | Guestlogix/traveler-ios | https://api.github.com/repos/Guestlogix/traveler-ios | closed | Minor | Text in "Airline Name" and "Flight Number" field disappear when dismissing keyboard on those fields | :bug: bug Future Build iOS pm-review | **Steps:**
1. Launch App
2. Navigate to booking form
3. Click "Airline Name" or "Flight Number"
4. Dismiss keyboard by dragging down on the screen
**Actual result:**
Keyboard is dismissed but text inside that field also disappears
**Expected result:**
Keyboard dismisses without text disappearing
**Screenshots:**

Build 22 0.3(40)
| 1.0 | Minor | Text in "Airline Name" and "Flight Number" field disappear when dismissing keyboard on those fields - **Steps:**
1. Launch App
2. Navigate to booking form
3. Click "Airline Name" or "Flight Number"
4. Dismiss keyboard by dragging down on the screen
**Actual result:**
Keyboard is dismissed but text inside that field also disappears
**Expected result:**
Keyboard dismisses without text disappearing
**Screenshots:**

Build 22 0.3(40)
| non_priority | minor text in airline name and flight number field disappear when dismissing keyboard on those fields steps launch app navigate to booking form click airline name or flight number dismiss keyboard by dragging down on the screen actual result keyboard is dismissed but text inside that field also disappears expected result keyboard dismisses without text disappearing screenshots build | 0 |
271,769 | 8,489,111,978 | IssuesEvent | 2018-10-26 18:48:00 | wevote/WebApp | https://api.github.com/repos/wevote/WebApp | closed | Race Level & Decided State: Save in Store | Difficulty: Medium Priority: 1 | ### Please describe the issue (What happens? What do you expect?)
When a voter dives in to look at a candidate or measure from the Ballot page, and then comes back to the ballot, the "Race Level" (Federal, State, Measure, Local), and the "Decided State" (Remaining Decisions, All Items, Items Decided) both get unset. Store these "states" in a Store so they remain the same when the voter returns to the Ballot.
| 1.0 | Race Level & Decided State: Save in Store - ### Please describe the issue (What happens? What do you expect?)
When a voter dives in to look at a candidate or measure from the Ballot page, and then comes back to the ballot, the "Race Level" (Federal, State, Measure, Local), and the "Decided State" (Remaining Decisions, All Items, Items Decided) both get unset. Store these "states" in a Store so they remain the same when the voter returns to the Ballot.
| priority | race level decided state save in store please describe the issue what happens what do you expect when a voter dives in to look at a candidate or measure from the ballot page and then comes back to the ballot the race level federal state measure local and the decided state remaining decisions all items items decided both get unset store these states in a store so they remain the same when the voter returns to the ballot | 1 |
26,516 | 26,903,976,433 | IssuesEvent | 2023-02-06 17:33:21 | DCS-LCSR/SignStream3 | https://api.github.com/repos/DCS-LCSR/SignStream3 | opened | Changing Utterance name UI not working as expected | bug severity LOW usability concern | Changing the name of an utterance appears to have some bugs. From the User's Guide...
> 4. To edit the label (i.e., assign or change the name of Utterance), click on the label
> in the Utterance Bar of the Temporal Partition and hold the mouse down until
> the text becomes editable (just as you would edit a filename in the Finder in Mac
> OS). A visible text box for the label will appear. Edit the text of the label, and hit
> Return. The label will also change for all associated Segment Bars, as illustrated
> in Figure 29.
But clicking on the name and holding in SS does not work as it does in Mac Finder.
You can get it to pop up the name change box other ways, but it's a bit buggy. | True | Changing Utterance name UI not working as expected - Changing the name of an utterance appears to have some bugs. From the User's Guide...
> 4. To edit the label (i.e., assign or change the name of Utterance), click on the label
> in the Utterance Bar of the Temporal Partition and hold the mouse down until
> the text becomes editable (just as you would edit a filename in the Finder in Mac
> OS). A visible text box for the label will appear. Edit the text of the label, and hit
> Return. The label will also change for all associated Segment Bars, as illustrated
> in Figure 29.
But clicking on the name and holding in SS does not work as it does in Mac Finder.
You can get it to pop up the name change box other ways, but it's a bit buggy. | non_priority | changing utterance name ui not working as expected changing the name of an utterance appears to have some bugs from the user s guide to edit the label i e assign or change the name of utterance click on the label in the utterance bar of the temporal partition and hold the mouse down until the text becomes editable just as you would edit a filename in the finder in mac os a visible text box for the label will appear edit the text of the label and hit return the label will also change for all associated segment bars as illustrated in figure but clicking on the name and holding in ss does not work as it does in mac finder you can get it to pop up the name change box other ways but it s a bit buggy | 0 |
631,357 | 20,151,151,704 | IssuesEvent | 2022-02-09 12:31:48 | ita-social-projects/horondi_admin | https://api.github.com/repos/ita-social-projects/horondi_admin | closed | (SP:1)Material initial values don`t show on form | bug priority: high Admin | Steps to reproduce:
1) Go to https://horondi-admin-staging.azurewebsites.net/materials
2) Open material
Actual result:

Expected result:
Initial values should be shown in material form for each material
| 1.0 | (SP:1)Material initial values don`t show on form - Steps to reproduce:
1) Go to https://horondi-admin-staging.azurewebsites.net/materials
2) Open material
Actual result:

Expected result:
Initial values should be shown in material form for each material
| priority | sp material initial values don t show on form steps to reproduce go to open material actual result expected result initial values should be shown in material form for each material | 1 |
108,028 | 4,325,558,759 | IssuesEvent | 2016-07-26 00:28:34 | syb0rg/Khronos | https://api.github.com/repos/syb0rg/Khronos | closed | CMake Rebuild Problems | Priority: Medium Status: Completed Status: Pending Type: Bug | Upon rebuilding, CMake sometimes fails to download external dependencies. Deleting the project's folder in the `libs` solves the issue. | 1.0 | CMake Rebuild Problems - Upon rebuilding, CMake sometimes fails to download external dependencies. Deleting the project's folder in the `libs` solves the issue. | priority | cmake rebuild problems upon rebuilding cmake sometimes fails to download external dependencies deleting the project s folder in the libs solves the issue | 1 |
120,540 | 15,776,445,503 | IssuesEvent | 2021-04-01 04:42:44 | microsoft/react-native-windows | https://api.github.com/repos/microsoft/react-native-windows | closed | How to deal with NM breaking changes | Agenda Area: Native Modules enhancement needs dev design | Meta-point: we have breaking changes for NM across releases (C# for 0.63, e.g.). How do we deal with keeping community modules working despite breaking changes in the platform?
Breaking changes in Native Modules should have a higher bar than other breaking changes.
Do we need to make the breaking change now at all?
_Originally posted by @asklar in https://github.com/microsoft/react-native-windows/issues/6444#issuecomment-724225605_ | 1.0 | How to deal with NM breaking changes - Meta-point: we have breaking changes for NM across releases (C# for 0.63, e.g.). How do we deal with keeping community modules working despite breaking changes in the platform?
Breaking changes in Native Modules should have a higher bar than other breaking changes.
Do we need to make the breaking change now at all?
_Originally posted by @asklar in https://github.com/microsoft/react-native-windows/issues/6444#issuecomment-724225605_ | non_priority | how to deal with nm breaking changes meta point we have breaking changes for nm across releases c for e g how do we deal with keeping community modules working despite breaking changes in the platform breaking changes in native modules should have a higher bar than other breaking changes do we need to make the breaking change now at all originally posted by asklar in | 0 |
54,636 | 7,904,502,742 | IssuesEvent | 2018-07-02 05:05:01 | SalatielBairros/HomeLibrary | https://api.github.com/repos/SalatielBairros/HomeLibrary | opened | Terminar documentação inicial | documentation help wanted | Terminar documentação inicial do projeto para entrega do Marco Avaliativo 2. | 1.0 | Terminar documentação inicial - Terminar documentação inicial do projeto para entrega do Marco Avaliativo 2. | non_priority | terminar documentação inicial terminar documentação inicial do projeto para entrega do marco avaliativo | 0 |
34,006 | 16,373,389,657 | IssuesEvent | 2021-05-15 15:59:27 | Dart-Code/Dart-Code | https://api.github.com/repos/Dart-Code/Dart-Code | closed | Unusable in Large Monorepo | awaiting info in editor in lsp/analysis server is performance | We operate a large (~230k lines) monorepo consisting of several Flutter teams, several apps and several dozens of packages all using various Dart versions (null-safe and non-null-safe, though the entire monorepo uses Flutter 2.0.5). As the monorepo has continued to grow, intellisense, auto-complete, auto-import etc performance in VSCode has continued to degrade.
We're very heavy users of Git sparse-checkout to remove apps and packages that aren't relevant for people working on specific teams. For awhile this actually solved all of our performance problems in VSCode. Now however, even the smallest team's sparse-checkout is ~18% of the monorepo and performance is as bad as ever.
**To Reproduce**
Have a big monorepo.
**Expected behavior**
Intellisense, auto-complete, auto-import should work.
**Versions (please complete the following information):**
- VS Code version: 1.56.0
- Dart extension version: 3.22.0
- Dart/Flutter SDK version: 2.0.6
Is there a way to capture a performance trace that can be submitted somewhere?
| True | Unusable in Large Monorepo - We operate a large (~230k lines) monorepo consisting of several Flutter teams, several apps and several dozens of packages all using various Dart versions (null-safe and non-null-safe, though the entire monorepo uses Flutter 2.0.5). As the monorepo has continued to grow, intellisense, auto-complete, auto-import etc performance in VSCode has continued to degrade.
We're very heavy users of Git sparse-checkout to remove apps and packages that aren't relevant for people working on specific teams. For awhile this actually solved all of our performance problems in VSCode. Now however, even the smallest team's sparse-checkout is ~18% of the monorepo and performance is as bad as ever.
**To Reproduce**
Have a big monorepo.
**Expected behavior**
Intellisense, auto-complete, auto-import should work.
**Versions (please complete the following information):**
- VS Code version: 1.56.0
- Dart extension version: 3.22.0
- Dart/Flutter SDK version: 2.0.6
Is there a way to capture a performance trace that can be submitted somewhere?
| non_priority | unusable in large monorepo we operate a large lines monorepo consisting of several flutter teams several apps and several dozens of packages all using various dart versions null safe and non null safe though the entire monorepo uses flutter as the monorepo has continued to grow intellisense auto complete auto import etc performance in vscode has continued to degrade we re very heavy users of git sparse checkout to remove apps and packages that aren t relevant for people working on specific teams for awhile this actually solved all of our performance problems in vscode now however even the smallest team s sparse checkout is of the monorepo and performance is as bad as ever to reproduce have a big monorepo expected behavior intellisense auto complete auto import should work versions please complete the following information vs code version dart extension version dart flutter sdk version is there a way to capture a performance trace that can be submitted somewhere | 0 |
105,160 | 9,037,227,639 | IssuesEvent | 2019-02-09 08:14:08 | chartjs/chartjs-plugin-datalabels | https://api.github.com/repos/chartjs/chartjs-plugin-datalabels | closed | [ask] get error from script source | needs test case | hello, i try this plugin on my html, but when i put the script then reload the page, i get 2 error like this..


do you guys know why this happen? Thanks! | 1.0 | [ask] get error from script source - hello, i try this plugin on my html, but when i put the script then reload the page, i get 2 error like this..


do you guys know why this happen? Thanks! | non_priority | get error from script source hello i try this plugin on my html but when i put the script then reload the page i get error like this do you guys know why this happen thanks | 0 |
139,372 | 12,855,294,750 | IssuesEvent | 2020-07-09 04:46:13 | rnpgp/rnp | https://api.github.com/repos/rnpgp/rnp | opened | Update Fuzzer section of `develop.adoc` due to #1195 | documentation | From @ni4 :
> LGTM, except minor copyright year mismatch.
> And it would be also good to update Fuzzer section of develop.adoc once this is merged.
| 1.0 | Update Fuzzer section of `develop.adoc` due to #1195 - From @ni4 :
> LGTM, except minor copyright year mismatch.
> And it would be also good to update Fuzzer section of develop.adoc once this is merged.
| non_priority | update fuzzer section of develop adoc due to from lgtm except minor copyright year mismatch and it would be also good to update fuzzer section of develop adoc once this is merged | 0 |
67,393 | 12,953,373,853 | IssuesEvent | 2020-07-20 00:20:58 | Abbassihraf/P-curiosity-LAB | https://api.github.com/repos/Abbassihraf/P-curiosity-LAB | closed | Programm details | Code In progress | ### **All versions**
- [x] Include nav
- [x] Include hero header
- [ ] Programms details section
- [x] Include footer | 1.0 | Programm details - ### **All versions**
- [x] Include nav
- [x] Include hero header
- [ ] Programms details section
- [x] Include footer | non_priority | programm details all versions include nav include hero header programms details section include footer | 0 |
171,995 | 6,497,755,605 | IssuesEvent | 2017-08-22 15:00:02 | fossasia/susi_skill_cms | https://api.github.com/repos/fossasia/susi_skill_cms | closed | Follow web chat and use same icons in dropdown | bug Priority: High | 
Please implement and use the same icons.
| 1.0 | Follow web chat and use same icons in dropdown - 
Please implement and use the same icons.
| priority | follow web chat and use same icons in dropdown please implement and use the same icons | 1 |
102,088 | 11,274,343,838 | IssuesEvent | 2020-01-14 18:21:57 | lxc/lxd | https://api.github.com/repos/lxc/lxd | closed | Multi-architecture clustering | API Documentation Easy Feature | LXD clustering can be used to turn multiple LXD servers into one large instance.
Right now, this assumes that all servers in the cluster are of the same architecture.
While that's certainly the common case, there are times where it would be useful to have a single LXD cluster which supports multiple architectures, usually a mix of Intel and Arm hardware.
To make this possible, we'd need to:
- Register the native architecture of each server in the `nodes` database table
- Update the generated /1.0 output to advertise all architectures supported by the cluster
- Update the default placement rule such that if asked to deploy an image specific to a given architecture, we find the least busy cluster member that supports that architecture | 1.0 | Multi-architecture clustering - LXD clustering can be used to turn multiple LXD servers into one large instance.
Right now, this assumes that all servers in the cluster are of the same architecture.
While that's certainly the common case, there are times where it would be useful to have a single LXD cluster which supports multiple architectures, usually a mix of Intel and Arm hardware.
To make this possible, we'd need to:
- Register the native architecture of each server in the `nodes` database table
- Update the generated /1.0 output to advertise all architectures supported by the cluster
- Update the default placement rule such that if asked to deploy an image specific to a given architecture, we find the least busy cluster member that supports that architecture | non_priority | multi architecture clustering lxd clustering can be used to turn multiple lxd servers into one large instance right now this assumes that all servers in the cluster are of the same architecture while that s certainly the common case there are times where it would be useful to have a single lxd cluster which supports multiple architectures usually a mix of intel and arm hardware to make this possible we d need to register the native architecture of each server in the nodes database table update the generated output to advertise all architectures supported by the cluster update the default placement rule such that if asked to deploy an image specific to a given architecture we find the least busy cluster member that supports that architecture | 0 |
57,435 | 11,756,093,988 | IssuesEvent | 2020-03-13 10:50:20 | fac19/week2-hklo | https://api.github.com/repos/fac19/week2-hklo | closed | G R E A T J O B 😍 | code review compliment | I love the concept of the game - you've done an amazing job fetching and displaying the data on the page! Well done you, this is great!
And... a super mega well done for working from home. I really admire what you've managed to create together working remotely. Wow!
PS: I'm a big fan of your game :)
 | 1.0 | G R E A T J O B 😍 - I love the concept of the game - you've done an amazing job fetching and displaying the data on the page! Well done you, this is great!
And... a super mega well done for working from home. I really admire what you've managed to create together working remotely. Wow!
PS: I'm a big fan of your game :)
 | non_priority | g r e a t j o b 😍 i love the concept of the game you ve done an amazing job fetching and displaying the data on the page well done you this is great and a super mega well done for working from home i really admire what you ve managed to create together working remotely wow ps i m a big fan of your game | 0 |
35,726 | 7,800,101,936 | IssuesEvent | 2018-06-09 04:49:46 | StrikeNP/trac_test | https://api.github.com/repos/StrikeNP/trac_test | closed | Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012 (Trac #491) | Migrated from Trac clubb_src defect roehl@uwm.edu | The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear.
Steven, could you please investigate?
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/491
```json
{
"status": "closed",
"changetime": "2012-02-07T15:32:28",
"description": "The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear.\n\nSteven, could you please investigate?",
"reporter": "vlarson@uwm.edu",
"cc": "vlarson@uwm.edu, dschanen@uwm.edu, connork@uwm.edu",
"resolution": "fixed",
"_ts": "1328628748000000",
"component": "clubb_src",
"summary": "Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012",
"priority": "critical",
"keywords": "",
"time": "2012-01-30T12:21:11",
"milestone": "",
"owner": "roehl@uwm.edu",
"type": "defect"
}
```
| 1.0 | Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012 (Trac #491) - The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear.
Steven, could you please investigate?
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/491
```json
{
"status": "closed",
"changetime": "2012-02-07T15:32:28",
"description": "The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear.\n\nSteven, could you please investigate?",
"reporter": "vlarson@uwm.edu",
"cc": "vlarson@uwm.edu, dschanen@uwm.edu, connork@uwm.edu",
"resolution": "fixed",
"_ts": "1328628748000000",
"component": "clubb_src",
"summary": "Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012",
"priority": "critical",
"keywords": "",
"time": "2012-01-30T12:21:11",
"milestone": "",
"owner": "roehl@uwm.edu",
"type": "defect"
}
```
| non_priority | nightly plots do not appear for cam gfdl and sam on jan trac the jpg images are missing when i try to display the nightly plots for cam gfdl and sam however clubb standalone plots do appear steven could you please investigate attachments migrated from json status closed changetime description the jpg images are missing when i try to display the nightly plots for cam gfdl and sam however clubb standalone plots do appear n nsteven could you please investigate reporter vlarson uwm edu cc vlarson uwm edu dschanen uwm edu connork uwm edu resolution fixed ts component clubb src summary nightly plots do not appear for cam gfdl and sam on jan priority critical keywords time milestone owner roehl uwm edu type defect | 0 |
366,310 | 10,819,566,544 | IssuesEvent | 2019-11-08 14:39:30 | raz0red/wii-mednafen | https://api.github.com/repos/raz0red/wii-mednafen | closed | NTFS Support | Priority-Medium Type-Enhancement auto-migrated | ```
Please add Support to NTFS Hard Disk!
```
Original issue reported on code.google.com by `cardelli...@gmail.com` on 3 Mar 2012 at 9:41
| 1.0 | NTFS Support - ```
Please add Support to NTFS Hard Disk!
```
Original issue reported on code.google.com by `cardelli...@gmail.com` on 3 Mar 2012 at 9:41
| priority | ntfs support please add support to ntfs hard disk original issue reported on code google com by cardelli gmail com on mar at | 1 |
251,860 | 21,526,307,611 | IssuesEvent | 2022-04-28 18:49:29 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | server: TestClusterVersionUpgrade failed | C-test-failure O-robot GA-blocker T-server-and-security branch-release-22.1.0 | server.TestClusterVersionUpgrade [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=artifacts#/) on release-22.1.0 @ [dc225064c5f50ae59bd2216283b5a1ee13adbd1d](https://github.com/cockroachdb/cockroach/commits/dc225064c5f50ae59bd2216283b5a1ee13adbd1d):
```
=== RUN TestClusterVersionUpgrade
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/ab5462d2cb3989e2a33018db970f23de/logTestClusterVersionUpgrade2444000157
test_log_scope.go:80: use -show-logs to present logs inline
version_cluster_test.go:271: condition failed to evaluate within 45s: 0: v21.2-112 active=true (wanted false)
goroutine 520664 [running]:
runtime/debug.Stack()
GOROOT/src/runtime/debug/stack.go:24 +0x65
github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithin({0x54dd058, 0xc000ce11e0}, 0x0, 0x406aa60)
github.com/cockroachdb/cockroach/pkg/testutils/soon.go:60 +0x5f
github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon({0x54dd058, 0xc000ce11e0}, 0xc006a11710)
github.com/cockroachdb/cockroach/pkg/testutils/soon.go:41 +0x4a
github.com/cockroachdb/cockroach/pkg/server_test.TestClusterVersionUpgrade(0xc000ce11e0)
github.com/cockroachdb/cockroach/pkg/server_test/pkg/server/version_cluster_test.go:271 +0x896
testing.tRunner(0xc000ce11e0, 0x4494a48)
GOROOT/src/testing/testing.go:1259 +0x102
created by testing.(*T).Run
GOROOT/src/testing/testing.go:1306 +0x35a
panic.go:642: -- test log scope end --
--- FAIL: TestClusterVersionUpgrade (46.93s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #79162 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-21.2]
- #78126 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-22.1]
</p>
</details>
/cc @cockroachdb/server
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestClusterVersionUpgrade.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 1.0 | server: TestClusterVersionUpgrade failed - server.TestClusterVersionUpgrade [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=artifacts#/) on release-22.1.0 @ [dc225064c5f50ae59bd2216283b5a1ee13adbd1d](https://github.com/cockroachdb/cockroach/commits/dc225064c5f50ae59bd2216283b5a1ee13adbd1d):
```
=== RUN TestClusterVersionUpgrade
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/ab5462d2cb3989e2a33018db970f23de/logTestClusterVersionUpgrade2444000157
test_log_scope.go:80: use -show-logs to present logs inline
version_cluster_test.go:271: condition failed to evaluate within 45s: 0: v21.2-112 active=true (wanted false)
goroutine 520664 [running]:
runtime/debug.Stack()
GOROOT/src/runtime/debug/stack.go:24 +0x65
github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithin({0x54dd058, 0xc000ce11e0}, 0x0, 0x406aa60)
github.com/cockroachdb/cockroach/pkg/testutils/soon.go:60 +0x5f
github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon({0x54dd058, 0xc000ce11e0}, 0xc006a11710)
github.com/cockroachdb/cockroach/pkg/testutils/soon.go:41 +0x4a
github.com/cockroachdb/cockroach/pkg/server_test.TestClusterVersionUpgrade(0xc000ce11e0)
github.com/cockroachdb/cockroach/pkg/server_test/pkg/server/version_cluster_test.go:271 +0x896
testing.tRunner(0xc000ce11e0, 0x4494a48)
GOROOT/src/testing/testing.go:1259 +0x102
created by testing.(*T).Run
GOROOT/src/testing/testing.go:1306 +0x35a
panic.go:642: -- test log scope end --
--- FAIL: TestClusterVersionUpgrade (46.93s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #79162 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-21.2]
- #78126 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-22.1]
</p>
</details>
/cc @cockroachdb/server
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestClusterVersionUpgrade.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_priority | server testclusterversionupgrade failed server testclusterversionupgrade with on release run testclusterversionupgrade test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline version cluster test go condition failed to evaluate within active true wanted false goroutine runtime debug stack goroot src runtime debug stack go github com cockroachdb cockroach pkg testutils succeedswithin github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg testutils succeedssoon github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg server test testclusterversionupgrade github com cockroachdb cockroach pkg server test pkg server version cluster test go testing trunner goroot src testing testing go created by testing t run goroot src testing testing go panic go test log scope end fail testclusterversionupgrade help see also parameters in this failure tags bazel gss same failure on other branches server testclusterversionupgrade failed server testclusterversionupgrade failed cc cockroachdb server | 0 |
107,149 | 16,751,637,914 | IssuesEvent | 2021-06-12 01:34:59 | gms-ws-demo/nibrs | https://api.github.com/repos/gms-ws-demo/nibrs | opened | CVE-2018-14718 (High) detected in multiple libraries | security vulnerability | ## CVE-2018-14718 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.8.10.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-flatfile/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-staging-data/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.8.0.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.10.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-demo/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718>CVE-2018-14718</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14718">https://nvd.nist.gov/vuln/detail/CVE-2018-14718</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-staging-data/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-14718","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-14718 (High) detected in multiple libraries - ## CVE-2018-14718 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.8.10.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-flatfile/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-staging-data/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.8.0.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.10.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-demo/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718>CVE-2018-14718</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14718">https://nvd.nist.gov/vuln/detail/CVE-2018-14718</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-staging-data/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-14718","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs flatfile pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs staging data pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar nibrs web nibrs web target nibrs web web inf lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs fbi service pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar nibrs tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the ext class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the ext class from polymorphic deserialization vulnerabilityurl | 0 |
93 | 2,502,435,182 | IssuesEvent | 2015-01-09 08:56:27 | dhowe/AdNauseam | https://api.github.com/repos/dhowe/AdNauseam | closed | Things to finish for before video screencasts | 1 - Awaiting Code PRIORITY: High | @dhowe these are the things we still need to fix before I can make the screencasts:
- [x] release a fix to the borked #132
- [x] implement popup #70
- [ ] display text ads #112
- [ ] add (non-tracking) share buttons to the site #134
- [ ] use a placeholder image for banners #122
- [ ] show the ad visiting animation #116
- [x] implement new logo #96
- [x] finish ADN site #33 (almost)
| 1.0 | Things to finish for before video screencasts - @dhowe these are the things we still need to fix before I can make the screencasts:
- [x] release a fix to the borked #132
- [x] implement popup #70
- [ ] display text ads #112
- [ ] add (non-tracking) share buttons to the site #134
- [ ] use a placeholder image for banners #122
- [ ] show the ad visiting animation #116
- [x] implement new logo #96
- [x] finish ADN site #33 (almost)
| non_priority | things to finish for before video screencasts dhowe these are the things we still need to fix before i can make the screencasts release a fix to the borked implement popup display text ads add non tracking share buttons to the site use a placeholder image for banners show the ad visiting animation implement new logo finish adn site almost | 0 |
20,862 | 14,221,132,893 | IssuesEvent | 2020-11-17 15:20:36 | spdk/spdk | https://api.github.com/repos/spdk/spdk | closed | Vagrant VM startup failure (libvirt) | Infrastructure Intermittent Failure | Failure log shows:
00:00:31.690 ==> default: -> value=-drive,
00:00:31.690 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-zero.img,if=none,id=nvme-2,
00:00:31.690 ==> default: -> value=-device,
00:00:31.690 ==> default: -> value=nvme,drive=nvme-2,serial=12342,namespaces=0,
00:00:31.691 ==> default: -> value=-drive,
00:00:31.691 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img,if=none,id=ocssd-3,
00:00:31.692 ==> default: -> value=-device,
00:00:31.692 ==> default: -> value=nvme,drive=ocssd-3,serial=deadbeef,oacs=0,namespaces=1,lver=2,lba_index=4,mdts=10,lnum_lun=8,lnum_pln=2,lpgs_per_blk=1536,lsecs_per_pg=2,lblks_per_pln=45,metadata=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_md,nsdatafile=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_blknvme.ns,laer_thread_sleep=3000,stride=4,
00:00:32.501 ==> default: Creating shared folders metadata...
00:00:32.536 ==> default: Starting domain.
00:00:32.943 There was an error talking to Libvirt. The error message is shown
00:00:32.943 below:
00:00:32.943
00:00:32.943 Call to virDomainCreateWithFlags failed: internal error: process exited while connecting to monitor: qemu-system-x86_64: -realtime mlock=off: warning: '-realtime mlock=...' is deprecated, please use '-overcommit mem-lock=...' instead
00:00:32.943 2020-10-20T17:47:47.443254Z qemu-system-x86_64: unable to map backing store for guest RAM: Cannot allocate memory
00:00:32.963 + cd fedora31-libvirt
00:00:32.963 + vagrant destroy -f
00:00:35.969 ==> default: Removing domain...
00:00:36.032 Build step 'Execute shell' marked build as failure
00:00:36.102 [PostBuildScript] - [INFO] Executing post build scripts.
00:00:36.108 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins9062853704202237667.sh
00:00:36.112 + [[ -n fedora31 ]]
00:00:36.112 + mv /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/fedora31-libvirt/output /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/output
00:00:36.129 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins7126983447848921706.sh
| 1.0 | Vagrant VM startup failure (libvirt) - Failure log shows:
00:00:31.690 ==> default: -> value=-drive,
00:00:31.690 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-zero.img,if=none,id=nvme-2,
00:00:31.690 ==> default: -> value=-device,
00:00:31.690 ==> default: -> value=nvme,drive=nvme-2,serial=12342,namespaces=0,
00:00:31.691 ==> default: -> value=-drive,
00:00:31.691 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img,if=none,id=ocssd-3,
00:00:31.692 ==> default: -> value=-device,
00:00:31.692 ==> default: -> value=nvme,drive=ocssd-3,serial=deadbeef,oacs=0,namespaces=1,lver=2,lba_index=4,mdts=10,lnum_lun=8,lnum_pln=2,lpgs_per_blk=1536,lsecs_per_pg=2,lblks_per_pln=45,metadata=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_md,nsdatafile=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_blknvme.ns,laer_thread_sleep=3000,stride=4,
00:00:32.501 ==> default: Creating shared folders metadata...
00:00:32.536 ==> default: Starting domain.
00:00:32.943 There was an error talking to Libvirt. The error message is shown
00:00:32.943 below:
00:00:32.943
00:00:32.943 Call to virDomainCreateWithFlags failed: internal error: process exited while connecting to monitor: qemu-system-x86_64: -realtime mlock=off: warning: '-realtime mlock=...' is deprecated, please use '-overcommit mem-lock=...' instead
00:00:32.943 2020-10-20T17:47:47.443254Z qemu-system-x86_64: unable to map backing store for guest RAM: Cannot allocate memory
00:00:32.963 + cd fedora31-libvirt
00:00:32.963 + vagrant destroy -f
00:00:35.969 ==> default: Removing domain...
00:00:36.032 Build step 'Execute shell' marked build as failure
00:00:36.102 [PostBuildScript] - [INFO] Executing post build scripts.
00:00:36.108 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins9062853704202237667.sh
00:00:36.112 + [[ -n fedora31 ]]
00:00:36.112 + mv /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/fedora31-libvirt/output /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/output
00:00:36.129 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins7126983447848921706.sh
| non_priority | vagrant vm startup failure libvirt failure log shows default value drive default value format raw file var lib libvirt images backends nvme zero img if none id nvme default value device default value nvme drive nvme serial namespaces default value drive default value format raw file var lib libvirt images backends nvme ocssd img if none id ocssd default value device default value nvme drive ocssd serial deadbeef oacs namespaces lver lba index mdts lnum lun lnum pln lpgs per blk lsecs per pg lblks per pln metadata var lib libvirt images backends nvme ocssd img ocssd md nsdatafile var lib libvirt images backends nvme ocssd img ocssd blknvme ns laer thread sleep stride default creating shared folders metadata default starting domain there was an error talking to libvirt the error message is shown below call to virdomaincreatewithflags failed internal error process exited while connecting to monitor qemu system realtime mlock off warning realtime mlock is deprecated please use overcommit mem lock instead qemu system unable to map backing store for guest ram cannot allocate memory cd libvirt vagrant destroy f default removing domain build step execute shell marked build as failure executing post build scripts usr bin env bash tmp sh mv var jenkins workspace nvmf tcp uring vg autotest libvirt output var jenkins workspace nvmf tcp uring vg autotest output usr bin env bash tmp sh | 0 |
34,140 | 6,299,691,103 | IssuesEvent | 2017-07-21 00:09:35 | dgraph-io/dgraph | https://api.github.com/repos/dgraph-io/dgraph | closed | Update Go client documentation | documentation in progress | The client has a new API. We should update the documentation at docs.dgraph.io and in the Godoc. | 1.0 | Update Go client documentation - The client has a new API. We should update the documentation at docs.dgraph.io and in the Godoc. | non_priority | update go client documentation the client has a new api we should update the documentation at docs dgraph io and in the godoc | 0 |
517,703 | 15,018,708,058 | IssuesEvent | 2021-02-01 12:36:14 | buidl-labs/crypto-code-school-inside-tezos | https://api.github.com/repos/buidl-labs/crypto-code-school-inside-tezos | opened | Auth flow | Priority: Medium Type: Enhancement | Edge case that needs to be handled:
1. If the wallet is Uninstall after log in. Don't allow users to login.
2. Update copy for creating account in case where If no account is found. | 1.0 | Auth flow - Edge case that needs to be handled:
1. If the wallet is Uninstall after log in. Don't allow users to login.
2. Update copy for creating account in case where If no account is found. | priority | auth flow edge case that needs to be handled if the wallet is uninstall after log in don t allow users to login update copy for creating account in case where if no account is found | 1 |
397,285 | 27,158,460,109 | IssuesEvent | 2023-02-17 09:53:21 | appsmithorg/appsmith-docs | https://api.github.com/repos/appsmithorg/appsmith-docs | closed | [Docs]: Button Widget | Documentation Doc Rehaul User Education Pod | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Documentation Link
https://docs.appsmith.com/reference/widgets/button
### Discord/slack/intercom Link
_No response_
### Describe the problem
Restructuring
### Describe the improvement
Rehaul | 1.0 | [Docs]: Button Widget - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Documentation Link
https://docs.appsmith.com/reference/widgets/button
### Discord/slack/intercom Link
_No response_
### Describe the problem
Restructuring
### Describe the improvement
Rehaul | non_priority | button widget is there an existing issue for this i have searched the existing issues documentation link discord slack intercom link no response describe the problem restructuring describe the improvement rehaul | 0 |
161,824 | 20,156,384,692 | IssuesEvent | 2022-02-09 16:49:49 | ioana-nicolae/renovate2 | https://api.github.com/repos/ioana-nicolae/renovate2 | opened | CVE-2017-16137 (Medium) detected in debug-2.2.0.tgz | security vulnerability | ## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/renovate2/commit/9d0f2692d355dfff0d0ba18825cd40f2074a31f9">9d0f2692d355dfff0d0ba18825cd40f2074a31f9</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-16137 (Medium) detected in debug-2.2.0.tgz - ## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/renovate2/commit/9d0f2692d355dfff0d0ba18825cd40f2074a31f9">9d0f2692d355dfff0d0ba18825cd40f2074a31f9</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in debug tgz cve medium severity vulnerability vulnerable library debug tgz small debugging utility library home page a href path to dependency file package json path to vulnerable library node modules debug package json dependency hierarchy x debug tgz vulnerable library found in head commit a href found in base branch main vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree debug isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue vulnerabilityurl | 0 |
250,292 | 7,974,832,021 | IssuesEvent | 2018-07-17 07:23:42 | poanetwork/token-wizard | https://api.github.com/repos/poanetwork/token-wizard | reopened | (Bug) Contribution page:countdown timer has incorrect status after finalization | Step6-Contribute UI bug high priority migration to auth-os |
Steps:
1. Create crowdsale with any parameter, ex. :Minted, 1 tier, duration 2 min, minCap
2. Wait until crowdsale ended
3. As an owner open manage page and finalize
4. Open contribution page and observe countdown timer.
Expected result:
-timer's status is "CROWDSALE HAS BEEN FINALIZED""
Actual result:
-timer's status is "CROWDSALE HAS ENDED"
If you are reporting a problem with Token Wizard, please include the following information:
### Which network did you use? (Mainnet, Kovan, Rinkeby, etc.)
*localhost*
### If you were able to create it, what is the URL of your crowdsale?
*n/a*
### Do you have screenshots showing the problem?
<img width="1440" alt="screen shot 2018-07-11 at 00 08 37" src="https://user-images.githubusercontent.com/14067096/42556205-b78fd908-849f-11e8-9538-5f621802fc91.png">
### Do you see errors in the dev console? If yes, please include a screenshot
No
---
| 1.0 | (Bug) Contribution page:countdown timer has incorrect status after finalization -
Steps:
1. Create crowdsale with any parameter, ex. :Minted, 1 tier, duration 2 min, minCap
2. Wait until crowdsale ended
3. As an owner open manage page and finalize
4. Open contribution page and observe countdown timer.
Expected result:
-timer's status is "CROWDSALE HAS BEEN FINALIZED""
Actual result:
-timer's status is "CROWDSALE HAS ENDED"
If you are reporting a problem with Token Wizard, please include the following information:
### Which network did you use? (Mainnet, Kovan, Rinkeby, etc.)
*localhost*
### If you were able to create it, what is the URL of your crowdsale?
*n/a*
### Do you have screenshots showing the problem?
<img width="1440" alt="screen shot 2018-07-11 at 00 08 37" src="https://user-images.githubusercontent.com/14067096/42556205-b78fd908-849f-11e8-9538-5f621802fc91.png">
### Do you see errors in the dev console? If yes, please include a screenshot
No
---
| priority | bug contribution page countdown timer has incorrect status after finalization steps create crowdsale with any parameter ex minted tier duration min mincap wait until crowdsale ended as an owner open manage page and finalize open contribution page and observe countdown timer expected result timer s status is crowdsale has been finalized actual result timer s status is crowdsale has ended if you are reporting a problem with token wizard please include the following information which network did you use mainnet kovan rinkeby etc localhost if you were able to create it what is the url of your crowdsale n a do you have screenshots showing the problem img width alt screen shot at src do you see errors in the dev console if yes please include a screenshot no | 1 |
3,176 | 9,160,856,330 | IssuesEvent | 2019-03-01 08:52:24 | akvo/akvo-platform | https://api.github.com/repos/akvo/akvo-platform | closed | Revisit Authz | architecture in progress to do | * RSR Authz
Research the possibility of outsourcing the authorization of request to
Keycloak when moving the granularity of authorization from organization
to project.
* Lumen Authz
Work with Lumen team in implementing the fine-grained authorization in
the application
* Flow Authz
Revisit the document created by Dan (1 year ago) on the feasibility of
modeling Flow authorization in Keycloak Authz services
| 1.0 | Revisit Authz - * RSR Authz
Research the possibility of outsourcing the authorization of request to
Keycloak when moving the granularity of authorization from organization
to project.
* Lumen Authz
Work with Lumen team in implementing the fine-grained authorization in
the application
* Flow Authz
Revisit the document created by Dan (1 year ago) on the feasibility of
modeling Flow authorization in Keycloak Authz services
| non_priority | revisit authz rsr authz research the possibility of outsourcing the authorization of request to keycloak when moving the granularity of authorization from organization to project lumen authz work with lumen team in implementing the fine grained authorization in the application flow authz revisit the document created by dan year ago on the feasibility of modeling flow authorization in keycloak authz services | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.