Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
759,916
26,618,489,710
IssuesEvent
2023-01-24 09:24:10
rpm-software-management/dnf5
https://api.github.com/repos/rpm-software-management/dnf5
closed
Add highlevel API for usage of "install_specs"
Priority: HIGH
dnf4 (`install_specs()` method of the `Base` class) has a feature on API that enables usage of so-called "install_specs" for transaction operations. Install spec is a string that describes any artifact the user wants to include into transaction: - rpm from repository - local rpm file - URL to rpm file - group, environmental group - module New API method should - accept install spec on the input - be able to parse install spec and decide whether the artifact is a package, a group, or a module - download required package from given URL (for downloading the https://github.com/rpm-software-management/dnf5/issues/203 functionality is required) - de-duplicate command line packages, inserting the same rpm file to libsolv pool should be avoided (would behave like multiple independent packages and possibly cause troubles) - all downloads for given transaction are performed once during goal resolution Implementation of this feature will enable us to move parsing of packages specified on the command line from client apps (dnf5, dnf5daemon-client) to the common place in the libdnf5 library.
1.0
Add highlevel API for usage of "install_specs" - dnf4 (`install_specs()` method of the `Base` class) has a feature on API that enables usage of so-called "install_specs" for transaction operations. Install spec is a string that describes any artifact the user wants to include into transaction: - rpm from repository - local rpm file - URL to rpm file - group, environmental group - module New API method should - accept install spec on the input - be able to parse install spec and decide whether the artifact is a package, a group, or a module - download required package from given URL (for downloading the https://github.com/rpm-software-management/dnf5/issues/203 functionality is required) - de-duplicate command line packages, inserting the same rpm file to libsolv pool should be avoided (would behave like multiple independent packages and possibly cause troubles) - all downloads for given transaction are performed once during goal resolution Implementation of this feature will enable us to move parsing of packages specified on the command line from client apps (dnf5, dnf5daemon-client) to the common place in the libdnf5 library.
priority
add highlevel api for usage of install specs install specs method of the base class has a feature on api that enables usage of so called install specs for transaction operations install spec is a string that describes any artifact the user wants to include into transaction rpm from repository local rpm file url to rpm file group environmental group module new api method should accept install spec on the input be able to parse install spec and decide whether the artifact is a package a group or a module download required package from given url for downloading the functionality is required de duplicate command line packages inserting the same rpm file to libsolv pool should be avoided would behave like multiple independent packages and possibly cause troubles all downloads for given transaction are performed once during goal resolution implementation of this feature will enable us to move parsing of packages specified on the command line from client apps client to the common place in the library
1
143,965
11,589,311,876
IssuesEvent
2020-02-24 01:30:57
urapadmin/kiosk
https://api.github.com/repos/urapadmin/kiosk
closed
tag analysis has a stupid default value
C: before or during next season bug filemaker needs testing
tag analysis - Find a way to reset the standard input in tag analysis (that's why Wheeler always pops up)
1.0
tag analysis has a stupid default value - tag analysis - Find a way to reset the standard input in tag analysis (that's why Wheeler always pops up)
non_priority
tag analysis has a stupid default value tag analysis find a way to reset the standard input in tag analysis that s why wheeler always pops up
0
608,634
18,844,814,113
IssuesEvent
2021-11-11 13:51:38
weaveio/woll-forum
https://api.github.com/repos/weaveio/woll-forum
closed
Weekly average dose rate per Morning/Noon/ Afternoon (週別の服用率 - test)
Fixed Priority1
Please create a following graph. User "週別の服用率 - TEST" for this ticket. <br> <br> Mockup: <img width="799" alt="スクリーンショット 2021-10-25 12 25 17" src="https://user-images.githubusercontent.com/61481039/138630770-f199fcd8-a741-4622-ab79-784095fab2f4.png"> It will look similar to what we have now, But Monday/Tuesday.....instead of month <img width="1178" alt="スクリーンショット 2021-10-25 12 35 21" src="https://user-images.githubusercontent.com/61481039/138631127-9ab7dece-cb19-4512-a873-323dd5017bb7.png">
1.0
Weekly average dose rate per Morning/Noon/ Afternoon (週別の服用率 - test) - Please create a following graph. User "週別の服用率 - TEST" for this ticket. <br> <br> Mockup: <img width="799" alt="スクリーンショット 2021-10-25 12 25 17" src="https://user-images.githubusercontent.com/61481039/138630770-f199fcd8-a741-4622-ab79-784095fab2f4.png"> It will look similar to what we have now, But Monday/Tuesday.....instead of month <img width="1178" alt="スクリーンショット 2021-10-25 12 35 21" src="https://user-images.githubusercontent.com/61481039/138631127-9ab7dece-cb19-4512-a873-323dd5017bb7.png">
priority
weekly average dose rate per morning noon afternoon 週別の服用率 test please create a following graph user 週別の服用率 test for this ticket mockup img width alt スクリーンショット src it will look similar to what we have now but monday tuesday instead of month img width alt スクリーンショット src
1
39,118
2,851,134,727
IssuesEvent
2015-06-01 02:48:48
afollestad/cabinet-issue-tracker
https://api.github.com/repos/afollestad/cabinet-issue-tracker
closed
Use new support design library's widgets
enhancement high priority in progress
https://developer.android.com/tools/support-library/index.html `NavigationView`: navigation drawer `FloatingActionButton`: Unfortunately doesn't have menu right now (I don't think), but we should keep an eye out for that. `Snackbar`: Perhaps we can use this for error messages or just to replace the `Toast`s `AppBarLayout`: Responds to scrolling (like how we already have implemented)
1.0
Use new support design library's widgets - https://developer.android.com/tools/support-library/index.html `NavigationView`: navigation drawer `FloatingActionButton`: Unfortunately doesn't have menu right now (I don't think), but we should keep an eye out for that. `Snackbar`: Perhaps we can use this for error messages or just to replace the `Toast`s `AppBarLayout`: Responds to scrolling (like how we already have implemented)
priority
use new support design library s widgets navigationview navigation drawer floatingactionbutton unfortunately doesn t have menu right now i don t think but we should keep an eye out for that snackbar perhaps we can use this for error messages or just to replace the toast s appbarlayout responds to scrolling like how we already have implemented
1
4,897
4,707,272,746
IssuesEvent
2016-10-13 19:38:37
zulu-zsh/zulu
https://api.github.com/repos/zulu-zsh/zulu
opened
Improve performance of upgrade command
enhancement performance
Currently the `upgrade` command is very slow, as it has to perform a fetch on each package repository before it can check for changes. One possibility might be asynchronously fetching from each of the repositories on init, and keeping a log of the last fetch timestamp for each of them, then only performing the fetch again if the fetch is older than x amount of time (probably an hour should be enough). Any ideas welcome.
True
Improve performance of upgrade command - Currently the `upgrade` command is very slow, as it has to perform a fetch on each package repository before it can check for changes. One possibility might be asynchronously fetching from each of the repositories on init, and keeping a log of the last fetch timestamp for each of them, then only performing the fetch again if the fetch is older than x amount of time (probably an hour should be enough). Any ideas welcome.
non_priority
improve performance of upgrade command currently the upgrade command is very slow as it has to perform a fetch on each package repository before it can check for changes one possibility might be asynchronously fetching from each of the repositories on init and keeping a log of the last fetch timestamp for each of them then only performing the fetch again if the fetch is older than x amount of time probably an hour should be enough any ideas welcome
0
191,588
15,298,249,525
IssuesEvent
2021-02-24 09:29:17
SAP/luigi
https://api.github.com/repos/SAP/luigi
closed
Blog post with Luigi updates
documentation
Write a Luigi blog post explaining web components, new Luigi features and releases
1.0
Blog post with Luigi updates - Write a Luigi blog post explaining web components, new Luigi features and releases
non_priority
blog post with luigi updates write a luigi blog post explaining web components new luigi features and releases
0
810,648
30,252,527,329
IssuesEvent
2023-07-06 21:58:47
apcountryman/picolibrary
https://api.github.com/repos/apcountryman/picolibrary
closed
Add socket flags to WIZnet W5500 IP network stack
priority-normal status-awaiting_review type-enhancement
Add socket flags to WIZnet W5500 IP network stack (`::picolibrary::WIZnet::W5500::IP::Network_Stack`). - [x] The `Network_Stack` class should support the following new operations: - [x] `void set_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Set a socket's flags - [x] `void clear_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Clear a socket's flags - [x] `auto socket_flags( Socket_ID socket_id ) const noexcept -> std::uint_fast8_t;`: Get a socket's flags - [x] Documentation
1.0
Add socket flags to WIZnet W5500 IP network stack - Add socket flags to WIZnet W5500 IP network stack (`::picolibrary::WIZnet::W5500::IP::Network_Stack`). - [x] The `Network_Stack` class should support the following new operations: - [x] `void set_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Set a socket's flags - [x] `void clear_socket_flags( Socket_ID socket_id, std::uint_fast8_t mask ) noexcept;`: Clear a socket's flags - [x] `auto socket_flags( Socket_ID socket_id ) const noexcept -> std::uint_fast8_t;`: Get a socket's flags - [x] Documentation
priority
add socket flags to wiznet ip network stack add socket flags to wiznet ip network stack picolibrary wiznet ip network stack the network stack class should support the following new operations void set socket flags socket id socket id std uint t mask noexcept set a socket s flags void clear socket flags socket id socket id std uint t mask noexcept clear a socket s flags auto socket flags socket id socket id const noexcept std uint t get a socket s flags documentation
1
381,295
26,445,547,194
IssuesEvent
2023-01-16 06:52:24
Chillandchat/mobile
https://api.github.com/repos/Chillandchat/mobile
closed
Bugs list
bug documentation enhancement help wanted need review
### Bug when creating a room The error screen will be deployed when the user creates a new room. A error message will be also displayed in the console of the app, which read: ``` ERROR [TypeError: undefined is not an object (evaluating 'returnedMessages[returnedMessages.length - 1].user')] ``` #### Reproduction steps 1. Sign in 2. Tap plus icon 3. Tap create room 4. Enter details 5. Click create ### Styling bug A bug with the styling in the ```Menu``` screen was found. Specifically, in the ```RoomList``` component where the latest message will overflow to the outside of the screen. This prevents the user from reading it and is not aesthetcally pleasing as a contained text. #### Reproduction steps 1. Sign in That's All!
1.0
Bugs list - ### Bug when creating a room The error screen will be deployed when the user creates a new room. A error message will be also displayed in the console of the app, which read: ``` ERROR [TypeError: undefined is not an object (evaluating 'returnedMessages[returnedMessages.length - 1].user')] ``` #### Reproduction steps 1. Sign in 2. Tap plus icon 3. Tap create room 4. Enter details 5. Click create ### Styling bug A bug with the styling in the ```Menu``` screen was found. Specifically, in the ```RoomList``` component where the latest message will overflow to the outside of the screen. This prevents the user from reading it and is not aesthetcally pleasing as a contained text. #### Reproduction steps 1. Sign in That's All!
non_priority
bugs list bug when creating a room the error screen will be deployed when the user creates a new room a error message will be also displayed in the console of the app which read error user reproduction steps sign in tap plus icon tap create room enter details click create styling bug a bug with the styling in the menu screen was found specifically in the roomlist component where the latest message will overflow to the outside of the screen this prevents the user from reading it and is not aesthetcally pleasing as a contained text reproduction steps sign in that s all
0
373,689
11,047,375,230
IssuesEvent
2019-12-09 18:50:34
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Cannot disable default storage class
area/storage kind/bug priority/important-longterm
**Steps to reproduce**: 1. Start minikube (from scratch) 2. Disable addon : ``` > minikube addons disable default-storageclass ✅ "default-storageclass" was successfully disabled ``` 3. List storage classes: ``` > kubectl get storageclasses.storage.k8s.io NAME PROVISIONER AGE standard (default) k8s.io/minikube-hostpath 4m14s ``` **Bug**: `standard` storage class is still marked as _default_. Actually, the storage class is then annotated with both `storageclass.kubernetes.io/is-default-class: "true"` (from the addon) and `storageclass.beta.kubernetes.io/is-default-class: "false"` (added with `minikube addons disable default-storageclass`): ``` apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: annotations: storageclass.beta.kubernetes.io/is-default-class: "false" storageclass.kubernetes.io/is-default-class: "true" creationTimestamp: "2019-11-20T12:40:29Z" labels: addonmanager.kubernetes.io/mode: EnsureExists name: standard resourceVersion: "600" selfLink: /apis/storage.k8s.io/v1/storageclasses/standard uid: cdef4691-5dc5-42c5-8e02-db45e4b4265c provisioner: k8s.io/minikube-hostpath reclaimPolicy: Delete volumeBindingMode: Immediate ``` **Minikube version**: v1.5.2 **Kubernetes version**: v1.16.2
1.0
Cannot disable default storage class - **Steps to reproduce**: 1. Start minikube (from scratch) 2. Disable addon : ``` > minikube addons disable default-storageclass ✅ "default-storageclass" was successfully disabled ``` 3. List storage classes: ``` > kubectl get storageclasses.storage.k8s.io NAME PROVISIONER AGE standard (default) k8s.io/minikube-hostpath 4m14s ``` **Bug**: `standard` storage class is still marked as _default_. Actually, the storage class is then annotated with both `storageclass.kubernetes.io/is-default-class: "true"` (from the addon) and `storageclass.beta.kubernetes.io/is-default-class: "false"` (added with `minikube addons disable default-storageclass`): ``` apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: annotations: storageclass.beta.kubernetes.io/is-default-class: "false" storageclass.kubernetes.io/is-default-class: "true" creationTimestamp: "2019-11-20T12:40:29Z" labels: addonmanager.kubernetes.io/mode: EnsureExists name: standard resourceVersion: "600" selfLink: /apis/storage.k8s.io/v1/storageclasses/standard uid: cdef4691-5dc5-42c5-8e02-db45e4b4265c provisioner: k8s.io/minikube-hostpath reclaimPolicy: Delete volumeBindingMode: Immediate ``` **Minikube version**: v1.5.2 **Kubernetes version**: v1.16.2
priority
cannot disable default storage class steps to reproduce start minikube from scratch disable addon minikube addons disable default storageclass ✅ default storageclass was successfully disabled list storage classes kubectl get storageclasses storage io name provisioner age standard default io minikube hostpath bug standard storage class is still marked as default actually the storage class is then annotated with both storageclass kubernetes io is default class true from the addon and storageclass beta kubernetes io is default class false added with minikube addons disable default storageclass apiversion storage io kind storageclass metadata annotations storageclass beta kubernetes io is default class false storageclass kubernetes io is default class true creationtimestamp labels addonmanager kubernetes io mode ensureexists name standard resourceversion selflink apis storage io storageclasses standard uid provisioner io minikube hostpath reclaimpolicy delete volumebindingmode immediate minikube version kubernetes version
1
509,994
14,767,114,421
IssuesEvent
2021-01-10 04:39:28
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Not able to create a speaker due to Error "Field may not be null"
Priority: High Priority: Urgent bug
Trying to create a speaker on next months Codeheat event shows an error message "Field may not be null" even though I filled in all fields. Compare https://eventyay.com/events/f4fc71b8/speakers/create ![Screenshot from 2021-01-10 05-38-24](https://user-images.githubusercontent.com/1583873/104114569-26280c00-5306-11eb-8581-1a56d44c823c.png)
2.0
Not able to create a speaker due to Error "Field may not be null" - Trying to create a speaker on next months Codeheat event shows an error message "Field may not be null" even though I filled in all fields. Compare https://eventyay.com/events/f4fc71b8/speakers/create ![Screenshot from 2021-01-10 05-38-24](https://user-images.githubusercontent.com/1583873/104114569-26280c00-5306-11eb-8581-1a56d44c823c.png)
priority
not able to create a speaker due to error field may not be null trying to create a speaker on next months codeheat event shows an error message field may not be null even though i filled in all fields compare
1
79,701
3,541,590,382
IssuesEvent
2016-01-19 02:13:18
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
Provide ability to ignore free space check on deployment
enhancement forreview priority
Some users have reported issues when trying to deploy to SSH endpoints where either [mono is reporting incorrect space] (https://github.com/OctopusDeploy/Calamari/blob/94e883c84e8605a7b032492e23e813e369811733/source/Calamari/Integration/FileSystem/NixPhysicalFileSystem.cs#L14) or, being on a linux machine, they just don't have or need 500Mb free on their drive. Alternatively they may be deploying the package ultimately to a seperate drive with different free space anyway and so the check is really not serving its real purpose. For this reason it has been suggested that we keep the current default, but provide a mechanism by way of special variables to ignore the space check and/or provide a different threshold to the currently hard-coded 500Mb [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L458) and [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L469). Since this calculation takes place within calamari this is a platform independent change and therefore will be available to Windows users too.
1.0
Provide ability to ignore free space check on deployment - Some users have reported issues when trying to deploy to SSH endpoints where either [mono is reporting incorrect space] (https://github.com/OctopusDeploy/Calamari/blob/94e883c84e8605a7b032492e23e813e369811733/source/Calamari/Integration/FileSystem/NixPhysicalFileSystem.cs#L14) or, being on a linux machine, they just don't have or need 500Mb free on their drive. Alternatively they may be deploying the package ultimately to a seperate drive with different free space anyway and so the check is really not serving its real purpose. For this reason it has been suggested that we keep the current default, but provide a mechanism by way of special variables to ignore the space check and/or provide a different threshold to the currently hard-coded 500Mb [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L458) and [here](https://github.com/OctopusDeploy/Calamari/blob/0b9b329bbede777cf2912cfb27fb54fc12f21188/source/Calamari/Integration/FileSystem/CalamariPhysicalFileSystem.cs#L469). Since this calculation takes place within calamari this is a platform independent change and therefore will be available to Windows users too.
priority
provide ability to ignore free space check on deployment some users have reported issues when trying to deploy to ssh endpoints where either or being on a linux machine they just don t have or need free on their drive alternatively they may be deploying the package ultimately to a seperate drive with different free space anyway and so the check is really not serving its real purpose for this reason it has been suggested that we keep the current default but provide a mechanism by way of special variables to ignore the space check and or provide a different threshold to the currently hard coded and since this calculation takes place within calamari this is a platform independent change and therefore will be available to windows users too
1
290,836
8,908,281,880
IssuesEvent
2019-01-18 00:06:33
nluedtke/linux_kernel_cves
https://api.github.com/repos/nluedtke/linux_kernel_cves
closed
Upgrade Webpack
Web Frontend high priority in progress
Upgrade webpack-dev-server to version 3.1.11 or later to fix vulnerable dependency.
1.0
Upgrade Webpack - Upgrade webpack-dev-server to version 3.1.11 or later to fix vulnerable dependency.
priority
upgrade webpack upgrade webpack dev server to version or later to fix vulnerable dependency
1
22,064
10,722,644,831
IssuesEvent
2019-10-27 13:32:32
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 51: libtiff-4.0.9
1.severity: security
## libtiff-4.0.9: 3 advisories [search](https://search.nix.gsc.io/?q=libtiff&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libtiff+in%3Apath&type=Code) - [ ] [CVE-2018-17000](https://nvd.nist.gov/vuln/detail/CVE-2018-17000) (unstable, 18.09, 18.03) - [ ] [CVE-2018-17100](https://nvd.nist.gov/vuln/detail/CVE-2018-17100) (unstable, 18.09, 18.03) - [ ] [CVE-2018-17101](https://nvd.nist.gov/vuln/detail/CVE-2018-17101) (unstable, 18.09, 18.03)
True
Vulnerability roundup 51: libtiff-4.0.9 - ## libtiff-4.0.9: 3 advisories [search](https://search.nix.gsc.io/?q=libtiff&i=fosho&repos=nixos-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=libtiff+in%3Apath&type=Code) - [ ] [CVE-2018-17000](https://nvd.nist.gov/vuln/detail/CVE-2018-17000) (unstable, 18.09, 18.03) - [ ] [CVE-2018-17100](https://nvd.nist.gov/vuln/detail/CVE-2018-17100) (unstable, 18.09, 18.03) - [ ] [CVE-2018-17101](https://nvd.nist.gov/vuln/detail/CVE-2018-17101) (unstable, 18.09, 18.03)
non_priority
vulnerability roundup libtiff libtiff advisories unstable unstable unstable
0
544,463
15,893,814,660
IssuesEvent
2021-04-11 07:43:45
googleapis/nodejs-proto-files
https://api.github.com/repos/googleapis/nodejs-proto-files
closed
Synthesis failed for nodejs-proto-files
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate nodejs-proto-files. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the nodejs-proto-files API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` MPLATE/support_request.md .github/PULL_REQUEST_TEMPLATE.md .github/release-please.yml .github/workflows/ci.yaml .kokoro/.gitattributes .kokoro/common.cfg .kokoro/continuous/node10/common.cfg .kokoro/continuous/node10/docs.cfg .kokoro/continuous/node10/test.cfg .kokoro/continuous/node12/common.cfg .kokoro/continuous/node12/lint.cfg .kokoro/continuous/node12/samples-test.cfg .kokoro/continuous/node12/system-test.cfg .kokoro/continuous/node12/test.cfg .kokoro/docs.sh .kokoro/lint.sh .kokoro/populate-secrets.sh .kokoro/presubmit/node10/common.cfg .kokoro/presubmit/node12/common.cfg .kokoro/presubmit/node12/samples-test.cfg .kokoro/presubmit/node12/system-test.cfg .kokoro/presubmit/node12/test.cfg .kokoro/publish.sh .kokoro/release/docs-devsite.cfg .kokoro/release/docs-devsite.sh .kokoro/release/docs.cfg .kokoro/release/docs.sh .kokoro/release/publish.cfg .kokoro/samples-test.sh .kokoro/system-test.sh .kokoro/test.bat .kokoro/test.sh .kokoro/trampoline.sh .kokoro/trampoline_v2.sh .mocharc.js .nycrc .prettierignore .prettierrc.js .trampolinerc CODE_OF_CONDUCT.md CONTRIBUTING.md LICENSE README.md api-extractor.json renovate.json samples/README.md 2021-04-09 00:42:57,323 synthtool [DEBUG] > Installing dependencies... DEBUG:synthtool:Installing dependencies... npm WARN deprecated sinon@10.0.1: Breaking change found in this patch version npm WARN deprecated ts-simple-ast@12.4.0: NOTICE: ts-simple-ast has been renamed to ts-morph and version reset to 1.0.0. Switch at your leisure... npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies. npm WARN deprecated opn@6.0.0: The package has been renamed to `open` npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2. npm WARN deprecated core-js@2.6.12: core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to the actual version of core-js@3. npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated > core-js@2.6.12 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/core-js > node -e "try{require('./postinstall')}catch(e){}" Thank you for using core-js ( https://github.com/zloirock/core-js ) for polyfilling JavaScript standard library! The project needs your help! Please consider supporting of core-js on Open Collective or Patreon:  > https://opencollective.com/core-js  > https://www.patreon.com/zloirock  Also, the author of core-js ( https://github.com/zloirock ) is looking for a good job -) > @compodoc/compodoc@1.1.11 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@compodoc/compodoc > opencollective-postinstall || exit 0 Thank you for using @compodoc/compodoc! If you rely on this package, please consider supporting our open collective: > https://opencollective.com/compodoc/donate > protobufjs@6.10.2 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/protobufjs > node scripts/postinstall > google-proto-files@2.4.0 prepare /home/kbuilder/.cache/synthtool/nodejs-proto-files > npm run compile && node ./build/tools/prepublish.js > google-proto-files@2.4.0 precompile /home/kbuilder/.cache/synthtool/nodejs-proto-files > gts clean version: 14 Removing build ... > google-proto-files@2.4.0 compile /home/kbuilder/.cache/synthtool/nodejs-proto-files > tsc -p . node_modules/@types/sinon/index.d.ts:778:36 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'TimerId'. 778 type SinonTimerId = FakeTimers.TimerId;    ~~~~~~~ node_modules/@types/sinon/index.d.ts:780:39 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'InstalledMethods'. 780 type SinonFakeTimers = FakeTimers.InstalledMethods &    ~~~~~~~~~~~~~~~~ node_modules/@types/sinon/index.d.ts:781:20 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'NodeClock'. 781 FakeTimers.NodeClock &    ~~~~~~~~~ node_modules/@types/sinon/index.d.ts:782:20 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'BrowserClock'. 782 FakeTimers.BrowserClock & {    ~~~~~~~~~~~~ Found 4 errors. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! google-proto-files@2.4.0 compile: `tsc -p .` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the google-proto-files@2.4.0 compile script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_479Z-debug.log npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! google-proto-files@2.4.0 prepare: `npm run compile && node ./build/tools/prepublish.js` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the google-proto-files@2.4.0 prepare script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_547Z-debug.log 2021-04-09 00:43:26,576 synthtool [ERROR] > Failed executing npm install: None ERROR:synthtool:Failed executing npm install: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/nodejs-proto-files/synth.py", line 27, in <module> node.install() File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 171, in install shell.run(["npm", "install"], hide_output=hide_output) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1. 2021-04-09 00:43:26,626 autosynth [ERROR] > Synthesis failed 2021-04-09 00:43:26,627 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 6ffe0d7 chore(deps): update dependency sinon to v10 (#337) 2021-04-09 00:43:26,634 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2021-04-09 00:43:26,639 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Removing node_modules/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/7ee53bfe-efd4-4c1a-ab7f-3a67ce2765e5/targets/github%2Fsynthtool;config=default/tests;query=nodejs-proto-files;failed=false).
1.0
Synthesis failed for nodejs-proto-files - Hello! Autosynth couldn't regenerate nodejs-proto-files. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the nodejs-proto-files API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` MPLATE/support_request.md .github/PULL_REQUEST_TEMPLATE.md .github/release-please.yml .github/workflows/ci.yaml .kokoro/.gitattributes .kokoro/common.cfg .kokoro/continuous/node10/common.cfg .kokoro/continuous/node10/docs.cfg .kokoro/continuous/node10/test.cfg .kokoro/continuous/node12/common.cfg .kokoro/continuous/node12/lint.cfg .kokoro/continuous/node12/samples-test.cfg .kokoro/continuous/node12/system-test.cfg .kokoro/continuous/node12/test.cfg .kokoro/docs.sh .kokoro/lint.sh .kokoro/populate-secrets.sh .kokoro/presubmit/node10/common.cfg .kokoro/presubmit/node12/common.cfg .kokoro/presubmit/node12/samples-test.cfg .kokoro/presubmit/node12/system-test.cfg .kokoro/presubmit/node12/test.cfg .kokoro/publish.sh .kokoro/release/docs-devsite.cfg .kokoro/release/docs-devsite.sh .kokoro/release/docs.cfg .kokoro/release/docs.sh .kokoro/release/publish.cfg .kokoro/samples-test.sh .kokoro/system-test.sh .kokoro/test.bat .kokoro/test.sh .kokoro/trampoline.sh .kokoro/trampoline_v2.sh .mocharc.js .nycrc .prettierignore .prettierrc.js .trampolinerc CODE_OF_CONDUCT.md CONTRIBUTING.md LICENSE README.md api-extractor.json renovate.json samples/README.md 2021-04-09 00:42:57,323 synthtool [DEBUG] > Installing dependencies... DEBUG:synthtool:Installing dependencies... npm WARN deprecated sinon@10.0.1: Breaking change found in this patch version npm WARN deprecated ts-simple-ast@12.4.0: NOTICE: ts-simple-ast has been renamed to ts-morph and version reset to 1.0.0. Switch at your leisure... npm WARN deprecated chokidar@2.1.8: Chokidar 2 will break on node v14+. Upgrade to chokidar 3 with 15x less dependencies. npm WARN deprecated opn@6.0.0: The package has been renamed to `open` npm WARN deprecated fsevents@1.2.13: fsevents 1 will break on node v14+ and could be using insecure binaries. Upgrade to fsevents 2. npm WARN deprecated core-js@2.6.12: core-js@<3 is no longer maintained and not recommended for usage due to the number of issues. Please, upgrade your dependencies to the actual version of core-js@3. npm WARN deprecated resolve-url@0.2.1: https://github.com/lydell/resolve-url#deprecated npm WARN deprecated urix@0.1.0: Please see https://github.com/lydell/urix#deprecated > core-js@2.6.12 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/core-js > node -e "try{require('./postinstall')}catch(e){}" Thank you for using core-js ( https://github.com/zloirock/core-js ) for polyfilling JavaScript standard library! The project needs your help! Please consider supporting of core-js on Open Collective or Patreon:  > https://opencollective.com/core-js  > https://www.patreon.com/zloirock  Also, the author of core-js ( https://github.com/zloirock ) is looking for a good job -) > @compodoc/compodoc@1.1.11 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@compodoc/compodoc > opencollective-postinstall || exit 0 Thank you for using @compodoc/compodoc! If you rely on this package, please consider supporting our open collective: > https://opencollective.com/compodoc/donate > protobufjs@6.10.2 postinstall /home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/protobufjs > node scripts/postinstall > google-proto-files@2.4.0 prepare /home/kbuilder/.cache/synthtool/nodejs-proto-files > npm run compile && node ./build/tools/prepublish.js > google-proto-files@2.4.0 precompile /home/kbuilder/.cache/synthtool/nodejs-proto-files > gts clean version: 14 Removing build ... > google-proto-files@2.4.0 compile /home/kbuilder/.cache/synthtool/nodejs-proto-files > tsc -p . node_modules/@types/sinon/index.d.ts:778:36 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'TimerId'. 778 type SinonTimerId = FakeTimers.TimerId;    ~~~~~~~ node_modules/@types/sinon/index.d.ts:780:39 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'InstalledMethods'. 780 type SinonFakeTimers = FakeTimers.InstalledMethods &    ~~~~~~~~~~~~~~~~ node_modules/@types/sinon/index.d.ts:781:20 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'NodeClock'. 781 FakeTimers.NodeClock &    ~~~~~~~~~ node_modules/@types/sinon/index.d.ts:782:20 - error TS2694: Namespace '"/home/kbuilder/.cache/synthtool/nodejs-proto-files/node_modules/@sinonjs/fake-timers/types/fake-timers-src"' has no exported member 'BrowserClock'. 782 FakeTimers.BrowserClock & {    ~~~~~~~~~~~~ Found 4 errors. npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! google-proto-files@2.4.0 compile: `tsc -p .` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the google-proto-files@2.4.0 compile script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_479Z-debug.log npm ERR! code ELIFECYCLE npm ERR! errno 1 npm ERR! google-proto-files@2.4.0 prepare: `npm run compile && node ./build/tools/prepublish.js` npm ERR! Exit status 1 npm ERR! npm ERR! Failed at the google-proto-files@2.4.0 prepare script. npm ERR! This is probably not a problem with npm. There is likely additional logging output above. npm ERR! A complete log of this run can be found in: npm ERR! /home/kbuilder/.npm/_logs/2021-04-09T07_43_26_547Z-debug.log 2021-04-09 00:43:26,576 synthtool [ERROR] > Failed executing npm install: None ERROR:synthtool:Failed executing npm install: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/nodejs-proto-files/synth.py", line 27, in <module> node.install() File "/tmpfs/src/github/synthtool/synthtool/languages/node.py", line 171, in install shell.run(["npm", "install"], hide_output=hide_output) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['npm', 'install']' returned non-zero exit status 1. 2021-04-09 00:43:26,626 autosynth [ERROR] > Synthesis failed 2021-04-09 00:43:26,627 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 6ffe0d7 chore(deps): update dependency sinon to v10 (#337) 2021-04-09 00:43:26,634 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2021-04-09 00:43:26,639 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Removing node_modules/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/7ee53bfe-efd4-4c1a-ab7f-3a67ce2765e5/targets/github%2Fsynthtool;config=default/tests;query=nodejs-proto-files;failed=false).
priority
synthesis failed for nodejs proto files hello autosynth couldn t regenerate nodejs proto files broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the nodejs proto files api and the library grows stale see for trouble shooting tips here s the output from running synth py mplate support request md github pull request template md github release please yml github workflows ci yaml kokoro gitattributes kokoro common cfg kokoro continuous common cfg kokoro continuous docs cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous lint cfg kokoro continuous samples test cfg kokoro continuous system test cfg kokoro continuous test cfg kokoro docs sh kokoro lint sh kokoro populate secrets sh kokoro presubmit common cfg kokoro presubmit common cfg kokoro presubmit samples test cfg kokoro presubmit system test cfg kokoro presubmit test cfg kokoro publish sh kokoro release docs devsite cfg kokoro release docs devsite sh kokoro release docs cfg kokoro release docs sh kokoro release publish cfg kokoro samples test sh kokoro system test sh kokoro test bat kokoro test sh kokoro trampoline sh kokoro trampoline sh mocharc js nycrc prettierignore prettierrc js trampolinerc code of conduct md contributing md license readme md api extractor json renovate json samples readme md synthtool installing dependencies debug synthtool installing dependencies npm warn deprecated sinon breaking change found in this patch version npm warn deprecated ts simple ast notice ts simple ast has been renamed to ts morph and version reset to switch at your leisure npm warn deprecated chokidar chokidar will break on node upgrade to chokidar with less dependencies npm warn deprecated opn the package has been renamed to open npm warn deprecated fsevents fsevents will break on node and could be using insecure binaries upgrade to fsevents npm warn deprecated core js core js is no longer maintained and not recommended for usage due to the number of issues please upgrade your dependencies to the actual version of core js npm warn deprecated resolve url npm warn deprecated urix please see core js postinstall home kbuilder cache synthtool nodejs proto files node modules core js node e try require postinstall catch e  you for using core js   for polyfilling javascript standard library   project needs your help please consider supporting of core js on open collective or patreon         the author of core js   is looking for a good job  compodoc compodoc postinstall home kbuilder cache synthtool nodejs proto files node modules compodoc compodoc opencollective postinstall exit   you for using compodoc compodoc     you rely on this package please consider supporting our open collective    protobufjs postinstall home kbuilder cache synthtool nodejs proto files node modules protobufjs node scripts postinstall google proto files prepare home kbuilder cache synthtool nodejs proto files npm run compile node build tools prepublish js google proto files precompile home kbuilder cache synthtool nodejs proto files gts clean version removing build google proto files compile home kbuilder cache synthtool nodejs proto files tsc p  modules types sinon index d ts         home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member timerid   type sinontimerid faketimers timerid      modules types sinon index d ts         home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member installedmethods   type sinonfaketimers faketimers installedmethods      modules types sinon index d ts         home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member nodeclock   faketimers nodeclock      modules types sinon index d ts         home kbuilder cache synthtool nodejs proto files node modules sinonjs fake timers types fake timers src has no exported member browserclock   faketimers browserclock     found errors npm err code elifecycle npm err errno npm err google proto files compile tsc p npm err exit status npm err npm err failed at the google proto files compile script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log npm err code elifecycle npm err errno npm err google proto files prepare npm run compile node build tools prepublish js npm err exit status npm err npm err failed at the google proto files prepare script npm err this is probably not a problem with npm there is likely additional logging output above npm err a complete log of this run can be found in npm err home kbuilder npm logs debug log synthtool failed executing npm install none error synthtool failed executing npm install none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool nodejs proto files synth py line in node install file tmpfs src github synthtool synthtool languages node py line in install shell run hide output hide output file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at chore deps update dependency sinon to autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache removing node modules traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth toolbox py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
1
8,817
3,220,225,124
IssuesEvent
2015-10-08 14:01:22
jqPlot/jqPlot
https://api.github.com/repos/jqPlot/jqPlot
closed
bar-charts on the examples website
bug documentation
hi, i wondered if this is the same thing. why does it not load any of the graphs? http://www.jqplot.com/tests/bar-charts.php
1.0
bar-charts on the examples website - hi, i wondered if this is the same thing. why does it not load any of the graphs? http://www.jqplot.com/tests/bar-charts.php
non_priority
bar charts on the examples website hi i wondered if this is the same thing why does it not load any of the graphs
0
3,975
2,610,085,101
IssuesEvent
2015-02-26 18:25:54
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳彩光治痤疮
auto-migrated Priority-Medium Type-Defect
``` 深圳彩光治痤疮【深圳韩方科颜全国热线400-869-1818,24小时QQ4 008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:02
1.0
深圳彩光治痤疮 - ``` 深圳彩光治痤疮【深圳韩方科颜全国热线400-869-1818,24小时QQ4 008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:02
non_priority
深圳彩光治痤疮 深圳彩光治痤疮【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� original issue reported on code google com by szft com on may at
0
493,911
14,240,929,200
IssuesEvent
2020-11-18 22:28:53
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
`nim js --gc:arc` gives bad error: undeclared identifier: '+!'
ARC/ORC Error messages JS Low Priority
Function `echo` outputs the wrong string. ### Example ```nim # can be an empty file ``` ### Current Output nim js --gc:arc --skipparentcfg --skipusercfg main.nim lib/system/arc.nim(84, 42) Error: undeclared identifier: '+!' ### Expected Output * either make `--gc:arc` a noop (and code should work) * or give error `--gc:arc` not supported with js backend ### Additional Information devel 1.5.1 87d3e5331a7838aa255deca8a840c2080b5424ce this (or fixing it by making it not work with js backend) would prevent things like this in testament: ```nim discard """ targets: "c cpp js" matrix: ";--gc:arc" """ ``` but testament could also be made smart about this and skip `js --gc:arc` combination, although this isn't trivial to do it cleanly ### note curiously, this works: ```nim # main2.nim: echo compileOption("gc", "arc") # cmd: nim r -b:js --gc:arc main2.nim false ``` likewise with `nim js -r -b:js --gc:arc main2.nim` but this doesn't: ``` nim js -r --gc:arc main2.nim Error: undeclared identifier: '+!' ```
1.0
`nim js --gc:arc` gives bad error: undeclared identifier: '+!' - Function `echo` outputs the wrong string. ### Example ```nim # can be an empty file ``` ### Current Output nim js --gc:arc --skipparentcfg --skipusercfg main.nim lib/system/arc.nim(84, 42) Error: undeclared identifier: '+!' ### Expected Output * either make `--gc:arc` a noop (and code should work) * or give error `--gc:arc` not supported with js backend ### Additional Information devel 1.5.1 87d3e5331a7838aa255deca8a840c2080b5424ce this (or fixing it by making it not work with js backend) would prevent things like this in testament: ```nim discard """ targets: "c cpp js" matrix: ";--gc:arc" """ ``` but testament could also be made smart about this and skip `js --gc:arc` combination, although this isn't trivial to do it cleanly ### note curiously, this works: ```nim # main2.nim: echo compileOption("gc", "arc") # cmd: nim r -b:js --gc:arc main2.nim false ``` likewise with `nim js -r -b:js --gc:arc main2.nim` but this doesn't: ``` nim js -r --gc:arc main2.nim Error: undeclared identifier: '+!' ```
priority
nim js gc arc gives bad error undeclared identifier function echo outputs the wrong string example nim can be an empty file current output nim js gc arc skipparentcfg skipusercfg main nim lib system arc nim error undeclared identifier expected output either make gc arc a noop and code should work or give error gc arc not supported with js backend additional information devel this or fixing it by making it not work with js backend would prevent things like this in testament nim discard targets c cpp js matrix gc arc but testament could also be made smart about this and skip js gc arc combination although this isn t trivial to do it cleanly note curiously this works nim nim echo compileoption gc arc cmd nim r b js gc arc nim false likewise with nim js r b js gc arc nim but this doesn t nim js r gc arc nim error undeclared identifier
1
30,443
4,209,838,484
IssuesEvent
2016-06-29 07:37:36
gandm/language-babel
https://api.github.com/repos/gandm/language-babel
closed
'react' incompatibility
as-designed
I'm getting an error saying that the package 'react' is not compatible with 'language-babel'. This happens after migrating to language-babel@2.29.0.
1.0
'react' incompatibility - I'm getting an error saying that the package 'react' is not compatible with 'language-babel'. This happens after migrating to language-babel@2.29.0.
non_priority
react incompatibility i m getting an error saying that the package react is not compatible with language babel this happens after migrating to language babel
0
3,047
5,207,647,206
IssuesEvent
2017-01-25 00:21:22
saenzramiro/rambox
https://api.github.com/repos/saenzramiro/rambox
closed
Amium shows browser unsupported message
service-bug
<!-- Please complete the follow issue template. If is not an issue, feel free to remove the steps and behaviors --> ### Steps to reproduce 1. Login to Amium ### Expected behavior Should detect the user agent as Chrome or Firefox instead of electron. <!-- Write here --> ### Actual behavior Shows this message "Oops, this browser isn't supported yet!" <!-- Write here --> ### ENVIRONMENT <!-- Complete your environment --> > Rambox: v0.4.4 > OS: Ubuntu 16.04 > Arch: x64
1.0
Amium shows browser unsupported message - <!-- Please complete the follow issue template. If is not an issue, feel free to remove the steps and behaviors --> ### Steps to reproduce 1. Login to Amium ### Expected behavior Should detect the user agent as Chrome or Firefox instead of electron. <!-- Write here --> ### Actual behavior Shows this message "Oops, this browser isn't supported yet!" <!-- Write here --> ### ENVIRONMENT <!-- Complete your environment --> > Rambox: v0.4.4 > OS: Ubuntu 16.04 > Arch: x64
non_priority
amium shows browser unsupported message steps to reproduce login to amium expected behavior should detect the user agent as chrome or firefox instead of electron actual behavior shows this message oops this browser isn t supported yet environment rambox os ubuntu arch
0
343,555
10,332,507,093
IssuesEvent
2019-09-03 00:32:07
HackGT/bolt
https://api.github.com/repos/HackGT/bolt
closed
Logout is broken
component / auth priority / high type / bug
When you click the logout button, you get the following error ```js TypeError: "listener" argument must be a function at ClientRequest.once (events.js:340:11) at new ClientRequest (_http_client.js:164:10) at Object.request (http.js:38:10) at Object.request (https.js:239:15) at exports.authRoutes.all (C:\Users\Evan\Documents\GitHub\bolt\server\build\auth\auth.js:97:19) at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5) at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:137:13) at Route.dispatch (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:112:3) at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5) at C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:281:22 at Function.process_params (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:335:12) at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:275:10) at Function.handle (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:174:3) at router (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:47:12) at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5) at trim_prefix (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:317:13)```
1.0
Logout is broken - When you click the logout button, you get the following error ```js TypeError: "listener" argument must be a function at ClientRequest.once (events.js:340:11) at new ClientRequest (_http_client.js:164:10) at Object.request (http.js:38:10) at Object.request (https.js:239:15) at exports.authRoutes.all (C:\Users\Evan\Documents\GitHub\bolt\server\build\auth\auth.js:97:19) at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5) at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:137:13) at Route.dispatch (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\route.js:112:3) at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5) at C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:281:22 at Function.process_params (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:335:12) at next (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:275:10) at Function.handle (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:174:3) at router (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:47:12) at Layer.handle [as handle_request] (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\layer.js:95:5) at trim_prefix (C:\Users\Evan\Documents\GitHub\bolt\server\node_modules\express\lib\router\index.js:317:13)```
priority
logout is broken when you click the logout button you get the following error js typeerror listener argument must be a function at clientrequest once events js at new clientrequest http client js at object request http js at object request https js at exports authroutes all c users evan documents github bolt server build auth auth js at layer handle c users evan documents github bolt server node modules express lib router layer js at next c users evan documents github bolt server node modules express lib router route js at route dispatch c users evan documents github bolt server node modules express lib router route js at layer handle c users evan documents github bolt server node modules express lib router layer js at c users evan documents github bolt server node modules express lib router index js at function process params c users evan documents github bolt server node modules express lib router index js at next c users evan documents github bolt server node modules express lib router index js at function handle c users evan documents github bolt server node modules express lib router index js at router c users evan documents github bolt server node modules express lib router index js at layer handle c users evan documents github bolt server node modules express lib router layer js at trim prefix c users evan documents github bolt server node modules express lib router index js
1
191,816
14,596,491,050
IssuesEvent
2020-12-20 16:04:07
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
hello-mr-code/terraform-oci: oci/identity_network_source_test.go; 16 LoC
fresh small test
Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/identity_network_source_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to networkSourceId is reassigned at line 281 [Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292) <details> <summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary> ```go for _, networkSourceId := range networkSourceIds { if ok := SweeperDefaultResourceId[networkSourceId]; !ok { deleteNetworkSourceRequest := oci_identity.DeleteNetworkSourceRequest{} deleteNetworkSourceRequest.NetworkSourceId = &networkSourceId deleteNetworkSourceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "identity") _, error := identityClient.DeleteNetworkSource(context.Background(), deleteNetworkSourceRequest) if error != nil { fmt.Printf("Error deleting NetworkSource %s %s, It is possible that the resource is already deleted. Please verify manually \n", networkSourceId, error) continue } waitTillCondition(testAccProvider, &networkSourceId, networkSourceSweepWaitCondition, time.Duration(3*time.Minute), networkSourceSweepResponseFetchOperation, "identity", true) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
1.0
hello-mr-code/terraform-oci: oci/identity_network_source_test.go; 16 LoC - Found a possible issue in [hello-mr-code/terraform-oci](https://www.github.com/hello-mr-code/terraform-oci) at [oci/identity_network_source_test.go](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > reference to networkSourceId is reassigned at line 281 [Click here to see the code in its original context.](https://github.com/hello-mr-code/terraform-oci/blob/2f6aa93ef8643328af454512a5fe78ab006697f0/oci/identity_network_source_test.go#L277-L292) <details> <summary>Click here to show the 16 line(s) of Go which triggered the analyzer.</summary> ```go for _, networkSourceId := range networkSourceIds { if ok := SweeperDefaultResourceId[networkSourceId]; !ok { deleteNetworkSourceRequest := oci_identity.DeleteNetworkSourceRequest{} deleteNetworkSourceRequest.NetworkSourceId = &networkSourceId deleteNetworkSourceRequest.RequestMetadata.RetryPolicy = getRetryPolicy(true, "identity") _, error := identityClient.DeleteNetworkSource(context.Background(), deleteNetworkSourceRequest) if error != nil { fmt.Printf("Error deleting NetworkSource %s %s, It is possible that the resource is already deleted. Please verify manually \n", networkSourceId, error) continue } waitTillCondition(testAccProvider, &networkSourceId, networkSourceSweepWaitCondition, time.Duration(3*time.Minute), networkSourceSweepResponseFetchOperation, "identity", true) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 2f6aa93ef8643328af454512a5fe78ab006697f0
non_priority
hello mr code terraform oci oci identity network source test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message reference to networksourceid is reassigned at line click here to show the line s of go which triggered the analyzer go for networksourceid range networksourceids if ok sweeperdefaultresourceid ok deletenetworksourcerequest oci identity deletenetworksourcerequest deletenetworksourcerequest networksourceid networksourceid deletenetworksourcerequest requestmetadata retrypolicy getretrypolicy true identity error identityclient deletenetworksource context background deletenetworksourcerequest if error nil fmt printf error deleting networksource s s it is possible that the resource is already deleted please verify manually n networksourceid error continue waittillcondition testaccprovider networksourceid networksourcesweepwaitcondition time duration time minute networksourcesweepresponsefetchoperation identity true leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
432,586
30,288,544,900
IssuesEvent
2023-07-09 01:28:12
aigic8/warmlight
https://api.github.com/repos/aigic8/warmlight
opened
document adding bot to Bot Father
documentation
It should be specified in the docs how to exactly add the bot to the Bot Father, especially bot commands, so they can be copied and sent to Bot Father
1.0
document adding bot to Bot Father - It should be specified in the docs how to exactly add the bot to the Bot Father, especially bot commands, so they can be copied and sent to Bot Father
non_priority
document adding bot to bot father it should be specified in the docs how to exactly add the bot to the bot father especially bot commands so they can be copied and sent to bot father
0
42,704
17,267,417,850
IssuesEvent
2021-07-22 15:17:00
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Support for EnableAdvancedFilteringOnArrays on Event Grid Subscription
enhancement service/event-grid
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The current [event grid subscription](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/eventgrid_event_subscription#advanced_filter) terraform does not support the property `EnableAdvancedFilteringOnArrays` ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_eventgrid_event_subscription ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_eventgrid_event_subscription" "default" { name = "defaultEventSubscription" scope = azurerm_resource_group.default.id storage_queue_endpoint { storage_account_id = azurerm_storage_account.default.id queue_name = azurerm_storage_queue.default.name } advanced_filtering_on_arrays_enabled = true } ``` ### References https://stackoverflow.com/questions/68345240/event-grid-subscription-terraform-missing-enable-advanced-filtering-on-arrays/
1.0
Support for EnableAdvancedFilteringOnArrays on Event Grid Subscription - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description The current [event grid subscription](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/eventgrid_event_subscription#advanced_filter) terraform does not support the property `EnableAdvancedFilteringOnArrays` ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_eventgrid_event_subscription ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_eventgrid_event_subscription" "default" { name = "defaultEventSubscription" scope = azurerm_resource_group.default.id storage_queue_endpoint { storage_account_id = azurerm_storage_account.default.id queue_name = azurerm_storage_queue.default.name } advanced_filtering_on_arrays_enabled = true } ``` ### References https://stackoverflow.com/questions/68345240/event-grid-subscription-terraform-missing-enable-advanced-filtering-on-arrays/
non_priority
support for enableadvancedfilteringonarrays on event grid subscription community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description the current terraform does not support the property enableadvancedfilteringonarrays new or affected resource s azurerm eventgrid event subscription potential terraform configuration hcl resource azurerm eventgrid event subscription default name defaulteventsubscription scope azurerm resource group default id storage queue endpoint storage account id azurerm storage account default id queue name azurerm storage queue default name advanced filtering on arrays enabled true references
0
314,176
26,980,012,445
IssuesEvent
2023-02-09 12:25:43
astronomer/astronomer-providers
https://api.github.com/repos/astronomer/astronomer-providers
closed
Install astronomer-provider with Airflow constraint file
feature pri/high testing
For example astronomer-providers-1.0.0 with airflow 2.4.0 and Python3.8 ``` pip install "astronomer-providers[all]==1.0.0" -c https://raw.githubusercontent.com/apache/airflow/constraints-2.4.0/constraints-3.8.txt ```
1.0
Install astronomer-provider with Airflow constraint file - For example astronomer-providers-1.0.0 with airflow 2.4.0 and Python3.8 ``` pip install "astronomer-providers[all]==1.0.0" -c https://raw.githubusercontent.com/apache/airflow/constraints-2.4.0/constraints-3.8.txt ```
non_priority
install astronomer provider with airflow constraint file for example astronomer providers with airflow and pip install astronomer providers c
0
97,771
28,497,475,101
IssuesEvent
2023-04-18 15:05:57
deephaven/deephaven-core
https://api.github.com/repos/deephaven/deephaven-core
opened
Add build information to jars
feature request build
For benchmarking and other provenance use cases, it would be useful to have more build information as part of the jars. The most useful pieces of information would probably be: * git commit hash (-dirty) * Is this a release build? * Build date / timestamp (this may come into conflict with development workflows which may favor cacheability, so ostensibly may only be desired on release builds) * Other build environment details The most canonical place to add this information would be in the jar manifest `META-INF/MANIFEST.MF`: https://docs.oracle.com/en/java/javase/11/docs/specs/jar/jar.html#jar-manifest. Here is an example of our current manifest: ``` Manifest-Version: 1.0 Implementation-Title: Deephaven Implementation-Version: 0.23.0 Provider: gradle ``` Note: regardless of adding more build information, we may want to change `Implementation-Title` to actually be the archive base name instead of "Deephaven". We _could_ add build information to a separate file in the jar, but unless there is a strong compelling reason why we _shouldn't_ add it to the manifest, the manifest is likely the best choice. As an added bonus, this would give us an opportunity to add information that automating tooling likes to look for (license, PURL info, ...). We recently did this for https://github.com/deephaven/deephaven-csv/pull/93: ``` Manifest-Version: 1.0 Specification-Title: deephaven-csv Specification-Version: 0.10.0 Specification-Vendor: Deephaven Data Labs Implementation-Title: deephaven-csv Implementation-Version: 0.10.0 Implementation-Vendor: Deephaven Data Labs Implementation-Vendor-Id: io.deephaven Implementation-URL: https://github.com/deephaven/deephaven-csv Bundle-License: https://www.apache.org/licenses/LICENSE-2.0.txt ```
1.0
Add build information to jars - For benchmarking and other provenance use cases, it would be useful to have more build information as part of the jars. The most useful pieces of information would probably be: * git commit hash (-dirty) * Is this a release build? * Build date / timestamp (this may come into conflict with development workflows which may favor cacheability, so ostensibly may only be desired on release builds) * Other build environment details The most canonical place to add this information would be in the jar manifest `META-INF/MANIFEST.MF`: https://docs.oracle.com/en/java/javase/11/docs/specs/jar/jar.html#jar-manifest. Here is an example of our current manifest: ``` Manifest-Version: 1.0 Implementation-Title: Deephaven Implementation-Version: 0.23.0 Provider: gradle ``` Note: regardless of adding more build information, we may want to change `Implementation-Title` to actually be the archive base name instead of "Deephaven". We _could_ add build information to a separate file in the jar, but unless there is a strong compelling reason why we _shouldn't_ add it to the manifest, the manifest is likely the best choice. As an added bonus, this would give us an opportunity to add information that automating tooling likes to look for (license, PURL info, ...). We recently did this for https://github.com/deephaven/deephaven-csv/pull/93: ``` Manifest-Version: 1.0 Specification-Title: deephaven-csv Specification-Version: 0.10.0 Specification-Vendor: Deephaven Data Labs Implementation-Title: deephaven-csv Implementation-Version: 0.10.0 Implementation-Vendor: Deephaven Data Labs Implementation-Vendor-Id: io.deephaven Implementation-URL: https://github.com/deephaven/deephaven-csv Bundle-License: https://www.apache.org/licenses/LICENSE-2.0.txt ```
non_priority
add build information to jars for benchmarking and other provenance use cases it would be useful to have more build information as part of the jars the most useful pieces of information would probably be git commit hash dirty is this a release build build date timestamp this may come into conflict with development workflows which may favor cacheability so ostensibly may only be desired on release builds other build environment details the most canonical place to add this information would be in the jar manifest meta inf manifest mf here is an example of our current manifest manifest version implementation title deephaven implementation version provider gradle note regardless of adding more build information we may want to change implementation title to actually be the archive base name instead of deephaven we could add build information to a separate file in the jar but unless there is a strong compelling reason why we shouldn t add it to the manifest the manifest is likely the best choice as an added bonus this would give us an opportunity to add information that automating tooling likes to look for license purl info we recently did this for manifest version specification title deephaven csv specification version specification vendor deephaven data labs implementation title deephaven csv implementation version implementation vendor deephaven data labs implementation vendor id io deephaven implementation url bundle license
0
312,013
26,831,659,970
IssuesEvent
2023-02-02 16:24:23
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] GeoIpDownloaderCliIT testGeoIpDatabasesDownloadNoGeoipProcessors failing
>test-failure :Data Management/Ingest Node Team:Data Management
Failed because of a NPE in the test itself. **Build scan:** https://gradle-enterprise.elastic.co/s/zztkk4y3bucwu/tests/:modules:ingest-geoip:internalClusterTest/org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT/testGeoIpDatabasesDownloadNoGeoipProcessors **Reproduction line:** ``` ./gradlew ':modules:ingest-geoip:internalClusterTest' --tests "org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT.testGeoIpDatabasesDownloadNoGeoipProcessors" -Dtests.seed=59DC63CBE0AC5A4C -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Grand_Turk -Druntime.java=17 ``` **Applicable branches:** main **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT&tests.test=testGeoIpDatabasesDownloadNoGeoipProcessors **Failure excerpt:** ``` java.lang.NullPointerException: Cannot invoke "org.elasticsearch.persistent.PersistentTasksCustomMetadata$PersistentTask.getState()" because the return value of "org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.getTask()" is null at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.lambda$testGeoIpDatabasesDownloadNoGeoipProcessors$14(GeoIpDownloaderIT.java:315) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1109) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1094) at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.testGeoIpDatabasesDownloadNoGeoipProcessors(GeoIpDownloaderIT.java:313) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:48) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
1.0
[CI] GeoIpDownloaderCliIT testGeoIpDatabasesDownloadNoGeoipProcessors failing - Failed because of a NPE in the test itself. **Build scan:** https://gradle-enterprise.elastic.co/s/zztkk4y3bucwu/tests/:modules:ingest-geoip:internalClusterTest/org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT/testGeoIpDatabasesDownloadNoGeoipProcessors **Reproduction line:** ``` ./gradlew ':modules:ingest-geoip:internalClusterTest' --tests "org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT.testGeoIpDatabasesDownloadNoGeoipProcessors" -Dtests.seed=59DC63CBE0AC5A4C -Dtests.locale=sr-Latn-BA -Dtests.timezone=America/Grand_Turk -Druntime.java=17 ``` **Applicable branches:** main **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.ingest.geoip.GeoIpDownloaderCliIT&tests.test=testGeoIpDatabasesDownloadNoGeoipProcessors **Failure excerpt:** ``` java.lang.NullPointerException: Cannot invoke "org.elasticsearch.persistent.PersistentTasksCustomMetadata$PersistentTask.getState()" because the return value of "org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.getTask()" is null at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.lambda$testGeoIpDatabasesDownloadNoGeoipProcessors$14(GeoIpDownloaderIT.java:315) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1109) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1094) at org.elasticsearch.ingest.geoip.GeoIpDownloaderIT.testGeoIpDatabasesDownloadNoGeoipProcessors(GeoIpDownloaderIT.java:313) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:568) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:48) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.tests.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.tests.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43) at org.apache.lucene.tests.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44) at org.apache.lucene.tests.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60) at org.apache.lucene.tests.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:833) ```
non_priority
geoipdownloadercliit testgeoipdatabasesdownloadnogeoipprocessors failing failed because of a npe in the test itself build scan reproduction line gradlew modules ingest geoip internalclustertest tests org elasticsearch ingest geoip geoipdownloadercliit testgeoipdatabasesdownloadnogeoipprocessors dtests seed dtests locale sr latn ba dtests timezone america grand turk druntime java applicable branches main reproduces locally no failure history failure excerpt java lang nullpointerexception cannot invoke org elasticsearch persistent persistenttaskscustommetadata persistenttask getstate because the return value of org elasticsearch ingest geoip geoipdownloaderit gettask is null at org elasticsearch ingest geoip geoipdownloaderit lambda testgeoipdatabasesdownloadnogeoipprocessors geoipdownloaderit java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch ingest geoip geoipdownloaderit testgeoipdatabasesdownloadnogeoipprocessors geoipdownloaderit java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene tests util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene tests util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene tests util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene tests util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene tests util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java
0
53,624
13,261,989,026
IssuesEvent
2020-08-20 20:54:10
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
millipede / taupede: error message "Assertion `alpha >= 0' failed" leads to abortion (Trac #1773)
Migrated from Trac combo reconstruction defect
When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better? A test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2 I am using a monopod-based seed, from "CscdL4_Monopod_noBD_4" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1773">https://code.icecube.wisc.edu/projects/icecube/ticket/1773</a>, reported by jstachurskaand owned by jbraun</em></summary> <p> ```json { "status": "closed", "changetime": "2016-07-12T16:09:01", "_ts": "1468339741343204", "description": "When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better?\n\nA test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2 \nI am using a monopod-based seed, from \"CscdL4_Monopod_noBD_4\" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed.", "reporter": "jstachurska", "cc": "", "resolution": "fixed", "time": "2016-07-06T19:12:52", "component": "combo reconstruction", "summary": "millipede / taupede: error message \"Assertion `alpha >= 0' failed\" leads to abortion", "priority": "normal", "keywords": "", "milestone": "", "owner": "jbraun", "type": "defect" } ``` </p> </details>
1.0
millipede / taupede: error message "Assertion `alpha >= 0' failed" leads to abortion (Trac #1773) - When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better? A test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2 I am using a monopod-based seed, from "CscdL4_Monopod_noBD_4" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1773">https://code.icecube.wisc.edu/projects/icecube/ticket/1773</a>, reported by jstachurskaand owned by jbraun</em></summary> <p> ```json { "status": "closed", "changetime": "2016-07-12T16:09:01", "_ts": "1468339741343204", "description": "When running Taupede (icerec trunc version) reconstruction, the above problem causes the whole job to abort. Maybe returning a warning and continuing with the next frame would be better?\n\nA test file with 2 unproblematic events and 2 events causing abortion can be found here: /data/user/jstachurska/NuTauAna/trunk_icerec/Test002-3253+3.i3.bz2 \nI am using a monopod-based seed, from \"CscdL4_Monopod_noBD_4\" and setting the seed.length = max(50 * I3Units.m * p.energy / I3Units.PeV, 10*I3Units.m) for a positive-length seed, and the negative of that for a negative-length seed.", "reporter": "jstachurska", "cc": "", "resolution": "fixed", "time": "2016-07-06T19:12:52", "component": "combo reconstruction", "summary": "millipede / taupede: error message \"Assertion `alpha >= 0' failed\" leads to abortion", "priority": "normal", "keywords": "", "milestone": "", "owner": "jbraun", "type": "defect" } ``` </p> </details>
non_priority
millipede taupede error message assertion alpha failed leads to abortion trac when running taupede icerec trunc version reconstruction the above problem causes the whole job to abort maybe returning a warning and continuing with the next frame would be better a test file with unproblematic events and events causing abortion can be found here data user jstachurska nutauana trunk icerec i am using a monopod based seed from monopod nobd and setting the seed length max m p energy pev m for a positive length seed and the negative of that for a negative length seed migrated from json status closed changetime ts description when running taupede icerec trunc version reconstruction the above problem causes the whole job to abort maybe returning a warning and continuing with the next frame would be better n na test file with unproblematic events and events causing abortion can be found here data user jstachurska nutauana trunk icerec ni am using a monopod based seed from monopod nobd and setting the seed length max m p energy pev m for a positive length seed and the negative of that for a negative length seed reporter jstachurska cc resolution fixed time component combo reconstruction summary millipede taupede error message assertion alpha failed leads to abortion priority normal keywords milestone owner jbraun type defect
0
60,008
3,117,669,176
IssuesEvent
2015-09-04 03:58:19
framingeinstein/issues-test
https://api.github.com/repos/framingeinstein/issues-test
closed
SPK-41: Blog Category: Remove More Categories from page.
priority:low resolution:will-not-fix
Hi Andrew, Let's just remove the "More Categories" functionality from the blog category to save time. See screenshot. [~Spencer] - We can move this to 1.1 if needbe.
1.0
SPK-41: Blog Category: Remove More Categories from page. - Hi Andrew, Let's just remove the "More Categories" functionality from the blog category to save time. See screenshot. [~Spencer] - We can move this to 1.1 if needbe.
priority
spk blog category remove more categories from page hi andrew let s just remove the more categories functionality from the blog category to save time see screenshot we can move this to if needbe
1
1,129
3,381,145,652
IssuesEvent
2015-11-26 00:11:05
BCDevExchange/BCDevExchange-app
https://api.github.com/repos/BCDevExchange/BCDevExchange-app
opened
As a user I would like to test against the API methods so that I can determine if the API service delivers the information I need.
API Services Bulls Eye Center
Users need to be able to test the listed APIs to see if the API delivers information that is salient to the business need. Created by Paul R DOD 1. Users can test the API according to the spec. 2. Users can test against modern programming languages (JAVA, NODE, PHP etc) Connect with Peter, Todd, Fred, Leo
1.0
As a user I would like to test against the API methods so that I can determine if the API service delivers the information I need. - Users need to be able to test the listed APIs to see if the API delivers information that is salient to the business need. Created by Paul R DOD 1. Users can test the API according to the spec. 2. Users can test against modern programming languages (JAVA, NODE, PHP etc) Connect with Peter, Todd, Fred, Leo
non_priority
as a user i would like to test against the api methods so that i can determine if the api service delivers the information i need users need to be able to test the listed apis to see if the api delivers information that is salient to the business need created by paul r dod users can test the api according to the spec users can test against modern programming languages java node php etc connect with peter todd fred leo
0
197,923
22,617,748,194
IssuesEvent
2022-06-30 01:04:38
turkdevops/sanity
https://api.github.com/repos/turkdevops/sanity
opened
CVE-2022-2216 (High) detected in parse-url-5.0.2.tgz
security vulnerability
## CVE-2022-2216 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/parse-url/package.json</p> <p> Dependency Hierarchy: - lerna-3.22.1.tgz (Root Library) - version-3.22.1.tgz - github-client-3.22.0.tgz - git-url-parse-11.4.4.tgz - git-up-4.0.2.tgz - :x: **parse-url-5.0.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>next</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 7.0.0. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2216>CVE-2022-2216</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/">https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution: parse-url - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-2216 (High) detected in parse-url-5.0.2.tgz - ## CVE-2022-2216 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/parse-url/package.json</p> <p> Dependency Hierarchy: - lerna-3.22.1.tgz (Root Library) - version-3.22.1.tgz - github-client-3.22.0.tgz - git-url-parse-11.4.4.tgz - git-up-4.0.2.tgz - :x: **parse-url-5.0.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>next</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 7.0.0. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2216>CVE-2022-2216</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/">https://huntr.dev/bounties/505a3d39-2723-4a06-b1f7-9b2d133c92e1/</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution: parse-url - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in parse url tgz cve high severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href path to dependency file package json path to vulnerable library node modules parse url package json dependency hierarchy lerna tgz root library version tgz github client tgz git url parse tgz git up tgz x parse url tgz vulnerable library found in base branch next vulnerability details server side request forgery ssrf in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution parse url step up your open source security game with mend
0
84,407
15,720,901,073
IssuesEvent
2021-03-29 01:33:55
benchmarkdebricked/generator-jhipster
https://api.github.com/repos/benchmarkdebricked/generator-jhipster
closed
CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed
security vulnerability
## CVE-2020-5398 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library) - :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input. <p>Publish Date: 2020-01-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p> <p>Release Date: 2020-01-17</p> <p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-5398 (High) detected in spring-web-5.0.6.RELEASE.jar - autoclosed - ## CVE-2020-5398 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.0.6.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /tmp/ws-scm/generator-jhipster/test/templates/ci-cd/maven-ngx-yarn/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-web/5.0.6.RELEASE/spring-web-5.0.6.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.2.RELEASE.jar (Root Library) - :x: **spring-web-5.0.6.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/benchmarkdebricked/generator-jhipster/commit/56521ff393b8726bd648cc9bf6e2fd9552b7cc38">56521ff393b8726bd648cc9bf6e2fd9552b7cc38</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.3, versions 5.1.x prior to 5.1.13, and versions 5.0.x prior to 5.0.16, an application is vulnerable to a reflected file download (RFD) attack when it sets a "Content-Disposition" header in the response where the filename attribute is derived from user supplied input. <p>Publish Date: 2020-01-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-5398>CVE-2020-5398</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pivotal.io/security/cve-2020-5398">https://pivotal.io/security/cve-2020-5398</a></p> <p>Release Date: 2020-01-17</p> <p>Fix Resolution: org.springframework:spring-web:5.0.16.RELEASE,org.springframework:spring-web:5.1.13.RELEASE,org.springframework:spring-web:5.2.3.RELEASE</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in spring web release jar autoclosed cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file tmp ws scm generator jhipster test templates ci cd maven ngx yarn pom xml path to vulnerable library root repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in head commit a href vulnerability details in spring framework versions x prior to versions x prior to and versions x prior to an application is vulnerable to a reflected file download rfd attack when it sets a content disposition header in the response where the filename attribute is derived from user supplied input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web release org springframework spring web release org springframework spring web release step up your open source security game with whitesource
0
64,349
6,900,338,494
IssuesEvent
2017-11-24 17:59:21
emfoundation/ce100-app
https://api.github.com/repos/emfoundation/ce100-app
closed
filter organistions by tags (from the API)
bug DAM-API please-test T2h
Organisations can be filtered by tags and locations: - [x] update the list of tags and location - [x] update the queries to search for tag or location
1.0
filter organistions by tags (from the API) - Organisations can be filtered by tags and locations: - [x] update the list of tags and location - [x] update the queries to search for tag or location
non_priority
filter organistions by tags from the api organisations can be filtered by tags and locations update the list of tags and location update the queries to search for tag or location
0
89,157
3,790,282,569
IssuesEvent
2016-03-21 20:54:27
phetsims/fraction-matcher
https://api.github.com/repos/phetsims/fraction-matcher
closed
Shapes are not correctly positioned in the match collection
priority:2-high type:bug type:performance
The position of representations (circles, squares, etc...) in their gray box will be incorrect if they are included in the last match of a perfectly scored level. This was seen to happen on Witten (iPad2 9.2.1), diSessa (iPad3 9.2.1), and Tycho (iPad2 0.3 Beta 6). I'm labeling this as a performance issue since the iPad Air 2 (Leibniz) does not show this problem. The Air blows the other iPads out of the water with its performance. There is a noticeable lag between pressing "OK" on the last match and when the reward cascade begins. During this lag is when the match animates towards the gray box and comes to rest incorrectly. ![20160315_122119](https://cloud.githubusercontent.com/assets/5863899/13789227/a9201858-eaa8-11e5-902a-d9a3746b1455.jpg) Zoomed: ![20160315_122130](https://cloud.githubusercontent.com/assets/5863899/13789233/ade548ae-eaa8-11e5-9a12-d327df1cd06d.jpg) Assigning to @ariel-phet to figure out who will deal with this. For phetsims/tasks/issues/541.
1.0
Shapes are not correctly positioned in the match collection - The position of representations (circles, squares, etc...) in their gray box will be incorrect if they are included in the last match of a perfectly scored level. This was seen to happen on Witten (iPad2 9.2.1), diSessa (iPad3 9.2.1), and Tycho (iPad2 0.3 Beta 6). I'm labeling this as a performance issue since the iPad Air 2 (Leibniz) does not show this problem. The Air blows the other iPads out of the water with its performance. There is a noticeable lag between pressing "OK" on the last match and when the reward cascade begins. During this lag is when the match animates towards the gray box and comes to rest incorrectly. ![20160315_122119](https://cloud.githubusercontent.com/assets/5863899/13789227/a9201858-eaa8-11e5-902a-d9a3746b1455.jpg) Zoomed: ![20160315_122130](https://cloud.githubusercontent.com/assets/5863899/13789233/ade548ae-eaa8-11e5-9a12-d327df1cd06d.jpg) Assigning to @ariel-phet to figure out who will deal with this. For phetsims/tasks/issues/541.
priority
shapes are not correctly positioned in the match collection the position of representations circles squares etc in their gray box will be incorrect if they are included in the last match of a perfectly scored level this was seen to happen on witten disessa and tycho beta i m labeling this as a performance issue since the ipad air leibniz does not show this problem the air blows the other ipads out of the water with its performance there is a noticeable lag between pressing ok on the last match and when the reward cascade begins during this lag is when the match animates towards the gray box and comes to rest incorrectly zoomed assigning to ariel phet to figure out who will deal with this for phetsims tasks issues
1
216,184
7,301,965,700
IssuesEvent
2018-02-27 08:00:00
Motoxpro/WorldCupStatsSite
https://api.github.com/repos/Motoxpro/WorldCupStatsSite
closed
Add Timed Training Overall
Medium Priority Data Issue Medium Priority Feature MySQL
Create a timed training overall by adding up all of the finishes and taking the lowest score. If a rider hasn't done a timed training round, their finish for that round is last place or just the number of riders in the session. Don't bother breaking ties
2.0
Add Timed Training Overall - Create a timed training overall by adding up all of the finishes and taking the lowest score. If a rider hasn't done a timed training round, their finish for that round is last place or just the number of riders in the session. Don't bother breaking ties
priority
add timed training overall create a timed training overall by adding up all of the finishes and taking the lowest score if a rider hasn t done a timed training round their finish for that round is last place or just the number of riders in the session don t bother breaking ties
1
301,748
9,223,542,809
IssuesEvent
2019-03-12 03:58:30
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
net: icmpv4: Zephyr drops valid echo request
area: Conformance area: Networking bug priority: medium
Zephyr drops echo request with valid checksum. Zephyr must respond to valid echo request. [icmpv4-valid-chksum2.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610165/icmpv4-valid-chksum2.pcap.gz) [icmpv4-valid-chksum1.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610166/icmpv4-valid-chksum1.pcap.gz)
1.0
net: icmpv4: Zephyr drops valid echo request - Zephyr drops echo request with valid checksum. Zephyr must respond to valid echo request. [icmpv4-valid-chksum2.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610165/icmpv4-valid-chksum2.pcap.gz) [icmpv4-valid-chksum1.pcap.gz](https://github.com/zephyrproject-rtos/zephyr/files/2610166/icmpv4-valid-chksum1.pcap.gz)
priority
net zephyr drops valid echo request zephyr drops echo request with valid checksum zephyr must respond to valid echo request
1
36,340
14,979,486,378
IssuesEvent
2021-01-28 12:20:01
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Drilldowns] Enable Drilldowns in edit mode for faster testing
Feature:Dashboard Feature:Drilldowns Team:AppServices enhancement
Currently, it is very tedious to test changes in a drilldown as you have to: 1. save a drilldown 2. save the dashboard 3. try the drilldown We intentionally disabled drilldowns in edit mode, because we were afraid the users would lose unsaved changes when navigating away. But after [this awesome pr](https://github.com/elastic/kibana/pull/86939) that stores drafts in session storage this isn't a problem anymore. So it seems like we can safely enable drilldowns in the edit mode and remove one step in between editing and trying-out the drilldown. We should also remove text from drilldown toasts that mention "save you dashboard before testing" cc @streamich @elastic-jb
1.0
[Drilldowns] Enable Drilldowns in edit mode for faster testing - Currently, it is very tedious to test changes in a drilldown as you have to: 1. save a drilldown 2. save the dashboard 3. try the drilldown We intentionally disabled drilldowns in edit mode, because we were afraid the users would lose unsaved changes when navigating away. But after [this awesome pr](https://github.com/elastic/kibana/pull/86939) that stores drafts in session storage this isn't a problem anymore. So it seems like we can safely enable drilldowns in the edit mode and remove one step in between editing and trying-out the drilldown. We should also remove text from drilldown toasts that mention "save you dashboard before testing" cc @streamich @elastic-jb
non_priority
enable drilldowns in edit mode for faster testing currently it is very tedious to test changes in a drilldown as you have to save a drilldown save the dashboard try the drilldown we intentionally disabled drilldowns in edit mode because we were afraid the users would lose unsaved changes when navigating away but after that stores drafts in session storage this isn t a problem anymore so it seems like we can safely enable drilldowns in the edit mode and remove one step in between editing and trying out the drilldown we should also remove text from drilldown toasts that mention save you dashboard before testing cc streamich elastic jb
0
443,384
12,793,665,996
IssuesEvent
2020-07-02 04:46:39
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
closed
Cannot delete conversation when Notification component is deactivated
Has-PR bug priority: medium status: needs review
**Describe the bug** Cannot delete conversation when Notification component is deactivated **To Reproduce** Steps to reproduce the behavior: 1. Deactivate notification component 2. Try to delete a conversation, notice that it will not be deleted. **Screencast** https://drive.google.com/file/d/1KuPuBwPM7Av406m2KrErIHH36c3PMC0R/view **Support ticket links** https://secure.helpscout.net/conversation/1205558198/79889
1.0
Cannot delete conversation when Notification component is deactivated - **Describe the bug** Cannot delete conversation when Notification component is deactivated **To Reproduce** Steps to reproduce the behavior: 1. Deactivate notification component 2. Try to delete a conversation, notice that it will not be deleted. **Screencast** https://drive.google.com/file/d/1KuPuBwPM7Av406m2KrErIHH36c3PMC0R/view **Support ticket links** https://secure.helpscout.net/conversation/1205558198/79889
priority
cannot delete conversation when notification component is deactivated describe the bug cannot delete conversation when notification component is deactivated to reproduce steps to reproduce the behavior deactivate notification component try to delete a conversation notice that it will not be deleted screencast support ticket links
1
338,493
10,230,066,965
IssuesEvent
2019-08-17 18:03:43
simplitech/simpli-web-sdk
https://api.github.com/repos/simplitech/simpli-web-sdk
closed
input components should validate and validation and on schema should be a bind property instead of separate avj info
bug highpriority
selects are not currently validable
1.0
input components should validate and validation and on schema should be a bind property instead of separate avj info - selects are not currently validable
priority
input components should validate and validation and on schema should be a bind property instead of separate avj info selects are not currently validable
1
384,251
11,386,054,434
IssuesEvent
2020-01-29 12:28:26
NukkitX/Nukkit
https://api.github.com/repos/NukkitX/Nukkit
closed
Speedbug
[Priority] Medium [Status] Unconfirmed [Type] Bug
### Actual Behavior <!--- What actually happened --> I had a problem on my server using Nukkit X last version. Players spawn 9/10 times with a speed effect (not visible on effect menu) very high. Please fix this bug ! ### Steps to Reproduce <!--- Reliable steps which someone can use to reproduce the issue. Please do not create issues for non reproducible bug! --> I dont know.... sorry :C
1.0
Speedbug - ### Actual Behavior <!--- What actually happened --> I had a problem on my server using Nukkit X last version. Players spawn 9/10 times with a speed effect (not visible on effect menu) very high. Please fix this bug ! ### Steps to Reproduce <!--- Reliable steps which someone can use to reproduce the issue. Please do not create issues for non reproducible bug! --> I dont know.... sorry :C
priority
speedbug actual behavior i had a problem on my server using nukkit x last version players spawn times with a speed effect not visible on effect menu very high please fix this bug steps to reproduce i dont know sorry c
1
30,367
8,526,463,699
IssuesEvent
2018-11-02 16:19:30
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
build with default cuda options fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable"
type:build/install
**System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 16.04.5 x86_64 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): master branch from 10/31/18 - - Python version: - Bazel version (if compiling from source): 0.18.0 - GCC/Compiler version (if compiling from source): 5.4.0 20160609 - CUDA/cuDNN version: 9.0/7 - GPU model and memory: 2x Tesla P100 16 GB **Describe the current behavior** Configuring TF to build with CUDA support and choosing all the default option for all the other questions in ./configure including the default for NCCL "Please specify the locally installed NCCL version you want to use. [Default is to use https://github.com/nvidia/nccl]:", the build fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable" This has been broken since this commit: https://github.com/tensorflow/tensorflow/commit/fc6cd33c334f88759ce637e29e1586733076e094 Running in the docker container: nvidia/cuda:9.0-cudnn7-devel (for ease of setup) export TF_NEED_CUDA=1 "" | ./configure bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package failed with: ERROR: /root/.cache/bazel/_bazel_root/68a62076e91007a7908bc42a32e4cff9/external/nccl_archive/BUILD.bazel:139:1: Executing genrule @nccl_archive//:device_code_fatbin_h failed (Exit 1) /bin/bash: CUDA_TOOLKIT_PATH: unbound variable Target //tensorflow/tools/pip_package:build_pip_package failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 267.057s, Critical Path: 93.66s INFO: 3606 processes: 3606 local. FAILED: Build did NOT complete successfully **Describe the expected behavior** Build should succeed. **Code to reproduce the issue** export TF_NEED_CUDA=1 "" | ./configure bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package **Other info / logs** Fails on both x86 and ppc64le, tried with both bazel 0.15.0 and 0.18.0
1.0
build with default cuda options fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable" - **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): No - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 16.04.5 x86_64 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: N/A - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): master branch from 10/31/18 - - Python version: - Bazel version (if compiling from source): 0.18.0 - GCC/Compiler version (if compiling from source): 5.4.0 20160609 - CUDA/cuDNN version: 9.0/7 - GPU model and memory: 2x Tesla P100 16 GB **Describe the current behavior** Configuring TF to build with CUDA support and choosing all the default option for all the other questions in ./configure including the default for NCCL "Please specify the locally installed NCCL version you want to use. [Default is to use https://github.com/nvidia/nccl]:", the build fails with "/bin/bash: CUDA_TOOLKIT_PATH: unbound variable" This has been broken since this commit: https://github.com/tensorflow/tensorflow/commit/fc6cd33c334f88759ce637e29e1586733076e094 Running in the docker container: nvidia/cuda:9.0-cudnn7-devel (for ease of setup) export TF_NEED_CUDA=1 "" | ./configure bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package failed with: ERROR: /root/.cache/bazel/_bazel_root/68a62076e91007a7908bc42a32e4cff9/external/nccl_archive/BUILD.bazel:139:1: Executing genrule @nccl_archive//:device_code_fatbin_h failed (Exit 1) /bin/bash: CUDA_TOOLKIT_PATH: unbound variable Target //tensorflow/tools/pip_package:build_pip_package failed to build Use --verbose_failures to see the command lines of failed build steps. INFO: Elapsed time: 267.057s, Critical Path: 93.66s INFO: 3606 processes: 3606 local. FAILED: Build did NOT complete successfully **Describe the expected behavior** Build should succeed. **Code to reproduce the issue** export TF_NEED_CUDA=1 "" | ./configure bazel build --config=opt --config=cuda //tensorflow/tools/pip_package:build_pip_package **Other info / logs** Fails on both x86 and ppc64le, tried with both bazel 0.15.0 and 0.18.0
non_priority
build with default cuda options fails with bin bash cuda toolkit path unbound variable system information have i written custom code as opposed to using a stock example script provided in tensorflow no os platform and distribution e g linux ubuntu linux ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device n a tensorflow installed from source or binary source tensorflow version use command below master branch from python version bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory tesla gb describe the current behavior configuring tf to build with cuda support and choosing all the default option for all the other questions in configure including the default for nccl please specify the locally installed nccl version you want to use the build fails with bin bash cuda toolkit path unbound variable this has been broken since this commit running in the docker container nvidia cuda devel for ease of setup export tf need cuda configure bazel build config opt config cuda tensorflow tools pip package build pip package failed with error root cache bazel bazel root external nccl archive build bazel executing genrule nccl archive device code fatbin h failed exit bin bash cuda toolkit path unbound variable target tensorflow tools pip package build pip package failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes local failed build did not complete successfully describe the expected behavior build should succeed code to reproduce the issue export tf need cuda configure bazel build config opt config cuda tensorflow tools pip package build pip package other info logs fails on both and tried with both bazel and
0
827,353
31,767,098,978
IssuesEvent
2023-09-12 09:26:00
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
[Playlist] Brave crashes when `Remove from Playlist` is clicked in the address bar playlist + button dropdown menu.
crash OS/Windows priority/P3 QA/No release-notes/exclude features/playlist OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> @sangwoo108 Clicking on `Remove from Playlist` is in the address bar is crashing the browser, it works fine when removing from the sidebar Playlist panel. I also noticed the 'Open in playlist' doesn't work, but doesn't crash the browser. 1.60.7 Chromium: 117.0.5938.48 (Official Build) nightly (64-bit) Windows 11 22H2 https://github.com/brave/brave-browser/assets/122518587/20a51737-9390-433b-bc0a-7be85c7979b8 Have a good day! BTW, something unrelated to this issue but about Playlist. In the past, you could manually add links to playlist, will that come back at some point? But more important, will Brave Playlist support grabbing many links/videos at once and using something like Youtube Playlists/Mixes? There is the `More media on this site` button, for other type of sites, but I thought I would ask, since it would be nice to have like a way to just paste Playlists links or get them from a playlist link, just to get videos added more easily, instead of manually go one by one; also, that would be useful to to avoid going to Youtube to get videos in the playlist, like if someone got the the link from the search engine or something like that.
1.0
[Playlist] Brave crashes when `Remove from Playlist` is clicked in the address bar playlist + button dropdown menu. - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> @sangwoo108 Clicking on `Remove from Playlist` is in the address bar is crashing the browser, it works fine when removing from the sidebar Playlist panel. I also noticed the 'Open in playlist' doesn't work, but doesn't crash the browser. 1.60.7 Chromium: 117.0.5938.48 (Official Build) nightly (64-bit) Windows 11 22H2 https://github.com/brave/brave-browser/assets/122518587/20a51737-9390-433b-bc0a-7be85c7979b8 Have a good day! BTW, something unrelated to this issue but about Playlist. In the past, you could manually add links to playlist, will that come back at some point? But more important, will Brave Playlist support grabbing many links/videos at once and using something like Youtube Playlists/Mixes? There is the `More media on this site` button, for other type of sites, but I thought I would ask, since it would be nice to have like a way to just paste Playlists links or get them from a playlist link, just to get videos added more easily, instead of manually go one by one; also, that would be useful to to avoid going to Youtube to get videos in the playlist, like if someone got the the link from the search engine or something like that.
priority
brave crashes when remove from playlist is clicked in the address bar playlist button dropdown menu have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description clicking on remove from playlist is in the address bar is crashing the browser it works fine when removing from the sidebar playlist panel i also noticed the open in playlist doesn t work but doesn t crash the browser chromium official build nightly bit windows have a good day btw something unrelated to this issue but about playlist in the past you could manually add links to playlist will that come back at some point but more important will brave playlist support grabbing many links videos at once and using something like youtube playlists mixes there is the more media on this site button for other type of sites but i thought i would ask since it would be nice to have like a way to just paste playlists links or get them from a playlist link just to get videos added more easily instead of manually go one by one also that would be useful to to avoid going to youtube to get videos in the playlist like if someone got the the link from the search engine or something like that
1
618,379
19,433,775,646
IssuesEvent
2021-12-21 14:52:35
BIDMCDigitalPsychiatry/LAMP-platform
https://api.github.com/repos/BIDMCDigitalPsychiatry/LAMP-platform
closed
User Hierarchy Update
bug 2day frontend priority 1
Can you please add the permission for User Admin roles to have access to the manage credentials button in the top left corner when they click on their listed role (where it says User Administrator in the corner). Additionally, make it so they can only assign the roles of user administrator and practice lead, not system administrator
1.0
User Hierarchy Update - Can you please add the permission for User Admin roles to have access to the manage credentials button in the top left corner when they click on their listed role (where it says User Administrator in the corner). Additionally, make it so they can only assign the roles of user administrator and practice lead, not system administrator
priority
user hierarchy update can you please add the permission for user admin roles to have access to the manage credentials button in the top left corner when they click on their listed role where it says user administrator in the corner additionally make it so they can only assign the roles of user administrator and practice lead not system administrator
1
297,400
25,727,678,935
IssuesEvent
2022-12-07 17:47:23
liberland/liberland_substrate
https://api.github.com/repos/liberland/liberland_substrate
opened
LLM - multiple calls lack proper permission checks
tests-blocker
Following calls can be executed by any signed origin: * `fake_send` - we should only have it enabled on dev. Having it on/off should be part of chain_spec, it's not acceptable to have to change code when releasing to mainnet. * `freeze_llm` - this call allows freezing LLM for any account, as long as origin is signed. It should either require specific origin or only apply to account of origin * `createllm` - is this even needed? * `mint_llm` - I'm not sure here, as there are some cap checks inside, so maybe it's OK for anyone to call it anytime... but isn't it already handled by `try_mint` run on block initialize? Maybe this should be removed?
1.0
LLM - multiple calls lack proper permission checks - Following calls can be executed by any signed origin: * `fake_send` - we should only have it enabled on dev. Having it on/off should be part of chain_spec, it's not acceptable to have to change code when releasing to mainnet. * `freeze_llm` - this call allows freezing LLM for any account, as long as origin is signed. It should either require specific origin or only apply to account of origin * `createllm` - is this even needed? * `mint_llm` - I'm not sure here, as there are some cap checks inside, so maybe it's OK for anyone to call it anytime... but isn't it already handled by `try_mint` run on block initialize? Maybe this should be removed?
non_priority
llm multiple calls lack proper permission checks following calls can be executed by any signed origin fake send we should only have it enabled on dev having it on off should be part of chain spec it s not acceptable to have to change code when releasing to mainnet freeze llm this call allows freezing llm for any account as long as origin is signed it should either require specific origin or only apply to account of origin createllm is this even needed mint llm i m not sure here as there are some cap checks inside so maybe it s ok for anyone to call it anytime but isn t it already handled by try mint run on block initialize maybe this should be removed
0
103,312
8,895,997,387
IssuesEvent
2019-01-16 10:13:57
tracim/tracim
https://api.github.com/repos/tracim/tracim
opened
Cypress test: test not working good all the time tiny_mce__can_type_content_spec.js
backend test
**At present** The test "app_html-document/tiny_mce__can_type_content_spec.js" is not working correctly each time is starting. Sometime you just need to restart this test and its working **Expected** - [ ] If this test not working good all the time, we need to disabled this test.
1.0
Cypress test: test not working good all the time tiny_mce__can_type_content_spec.js - **At present** The test "app_html-document/tiny_mce__can_type_content_spec.js" is not working correctly each time is starting. Sometime you just need to restart this test and its working **Expected** - [ ] If this test not working good all the time, we need to disabled this test.
non_priority
cypress test test not working good all the time tiny mce can type content spec js at present the test app html document tiny mce can type content spec js is not working correctly each time is starting sometime you just need to restart this test and its working expected if this test not working good all the time we need to disabled this test
0
414,550
27,991,906,811
IssuesEvent
2023-03-27 05:00:28
CarsOk/castillo_del_rey
https://api.github.com/repos/CarsOk/castillo_del_rey
closed
Script HTML, CSS
documentation
# Historias de usuario ## Yo como Aprendiz sena # Quiero llevar a cabo la realización de las interfaces de mi proyecto con su orden en el index # Para que de esta forma la avancemos con nuestro aplicativo. # Criterios de aceptación: - [] Debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente. - [] la información debe ser verídica y completa.
1.0
Script HTML, CSS - # Historias de usuario ## Yo como Aprendiz sena # Quiero llevar a cabo la realización de las interfaces de mi proyecto con su orden en el index # Para que de esta forma la avancemos con nuestro aplicativo. # Criterios de aceptación: - [] Debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente. - [] la información debe ser verídica y completa.
non_priority
script html css historias de usuario yo como aprendiz sena quiero llevar a cabo la realización de las interfaces de mi proyecto con su orden en el index para que de esta forma la avancemos con nuestro aplicativo criterios de aceptación debe cumplir con las herramientas de uso ya planteadas por nuestro grupo y el diseño correspondiente la información debe ser verídica y completa
0
7,737
5,176,455,958
IssuesEvent
2017-01-19 00:46:53
geneontology/go-site
https://api.github.com/repos/geneontology/go-site
closed
Make sure that users.yaml is the sole source for user metadata
bug (B: affects usability)
Now that users.yaml (and its schema) have been merged (#106) added to the repo, no other user metadata should be edited or used. Talking with @cmungall, some of this will be taken care of by documentation sooner (#105), but at some point we will have to do one of the following: 1. remove the files: GO.curators_dbxrefs, GO.user_data.json (TermGenie), termgenie-user-permissions.json (TermGenie) 2. make the listed files read-only versions that are produced by the users.yaml master (like dbxrefs.yaml) This is a ticket to pick back up on this later when we have more bandwidth.
True
Make sure that users.yaml is the sole source for user metadata - Now that users.yaml (and its schema) have been merged (#106) added to the repo, no other user metadata should be edited or used. Talking with @cmungall, some of this will be taken care of by documentation sooner (#105), but at some point we will have to do one of the following: 1. remove the files: GO.curators_dbxrefs, GO.user_data.json (TermGenie), termgenie-user-permissions.json (TermGenie) 2. make the listed files read-only versions that are produced by the users.yaml master (like dbxrefs.yaml) This is a ticket to pick back up on this later when we have more bandwidth.
non_priority
make sure that users yaml is the sole source for user metadata now that users yaml and its schema have been merged added to the repo no other user metadata should be edited or used talking with cmungall some of this will be taken care of by documentation sooner but at some point we will have to do one of the following remove the files go curators dbxrefs go user data json termgenie termgenie user permissions json termgenie make the listed files read only versions that are produced by the users yaml master like dbxrefs yaml this is a ticket to pick back up on this later when we have more bandwidth
0
151,968
13,440,196,295
IssuesEvent
2020-09-08 00:00:24
InstituteforDiseaseModeling/covasim
https://api.github.com/repos/InstituteforDiseaseModeling/covasim
closed
WebApp: Tooltips for testing and tracing intervention should mention that people are isolated.
CovasimUI documentation enhancement
Covasim, WebApp, click to expand Interventions Bar, hover over the intervention tabs Current descriptions for Testing and Tracing interventions are: "Testing rates for people with symptoms" and "Contact Tracing of diagnosed cases (requires testing intervention to be set)" Nothing in this indicates how that reduces the spread. I propose to expand the descriptions to "Testing rates and, if found positive isolating, for people with symptoms" and "Contact tracing of diagnosed cases, those found positive are isolated (requires testing intervention to be set)"
1.0
WebApp: Tooltips for testing and tracing intervention should mention that people are isolated. - Covasim, WebApp, click to expand Interventions Bar, hover over the intervention tabs Current descriptions for Testing and Tracing interventions are: "Testing rates for people with symptoms" and "Contact Tracing of diagnosed cases (requires testing intervention to be set)" Nothing in this indicates how that reduces the spread. I propose to expand the descriptions to "Testing rates and, if found positive isolating, for people with symptoms" and "Contact tracing of diagnosed cases, those found positive are isolated (requires testing intervention to be set)"
non_priority
webapp tooltips for testing and tracing intervention should mention that people are isolated covasim webapp click to expand interventions bar hover over the intervention tabs current descriptions for testing and tracing interventions are testing rates for people with symptoms and contact tracing of diagnosed cases requires testing intervention to be set nothing in this indicates how that reduces the spread i propose to expand the descriptions to testing rates and if found positive isolating for people with symptoms and contact tracing of diagnosed cases those found positive are isolated requires testing intervention to be set
0
95,070
10,865,387,877
IssuesEvent
2019-11-14 18:53:27
AIR-FOI-HR/AIR1925
https://api.github.com/repos/AIR-FOI-HR/AIR1925
closed
Izrada korisničkih zahtjeva
documentation
Identificiranje te popisivanje korisničkih zahtjeva sa razrađenim opisom i scenarijima.
1.0
Izrada korisničkih zahtjeva - Identificiranje te popisivanje korisničkih zahtjeva sa razrađenim opisom i scenarijima.
non_priority
izrada korisničkih zahtjeva identificiranje te popisivanje korisničkih zahtjeva sa razrađenim opisom i scenarijima
0
148,282
19,529,190,873
IssuesEvent
2021-12-30 13:41:27
developerone12/WebGoat-WhiteSource-Bolt
https://api.github.com/repos/developerone12/WebGoat-WhiteSource-Bolt
opened
CVE-2019-10768 (High) detected in angular-1.2.16.js, angular-1.2.16.min.js
security vulnerability
## CVE-2019-10768 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.2.16.js</b>, <b>angular-1.2.16.min.js</b></p></summary> <p> <details><summary><b>angular-1.2.16.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js</a></p> <p>Path to vulnerable library: /src/main/webapp/js/angular/angular.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.16.js** (Vulnerable Library) </details> <details><summary><b>angular-1.2.16.min.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js</a></p> <p>Path to vulnerable library: /src/main/webapp/js/angular/angular.min.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.16.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/developerone12/WebGoat-WhiteSource-Bolt/commit/c42e663814e4b88294ff90339ad577ca1afcf531">c42e663814e4b88294ff90339ad577ca1afcf531</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In AngularJS before 1.7.9 the function `merge()` could be tricked into adding or modifying properties of `Object.prototype` using a `__proto__` payload. <p>Publish Date: 2019-11-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10768>CVE-2019-10768</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a">https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a</a></p> <p>Release Date: 2019-11-19</p> <p>Fix Resolution: angularjs - 1.7.8,1.7.8.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10768 (High) detected in angular-1.2.16.js, angular-1.2.16.min.js - ## CVE-2019-10768 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.2.16.js</b>, <b>angular-1.2.16.min.js</b></p></summary> <p> <details><summary><b>angular-1.2.16.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.js</a></p> <p>Path to vulnerable library: /src/main/webapp/js/angular/angular.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.16.js** (Vulnerable Library) </details> <details><summary><b>angular-1.2.16.min.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.2.16/angular.min.js</a></p> <p>Path to vulnerable library: /src/main/webapp/js/angular/angular.min.js</p> <p> Dependency Hierarchy: - :x: **angular-1.2.16.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/developerone12/WebGoat-WhiteSource-Bolt/commit/c42e663814e4b88294ff90339ad577ca1afcf531">c42e663814e4b88294ff90339ad577ca1afcf531</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In AngularJS before 1.7.9 the function `merge()` could be tricked into adding or modifying properties of `Object.prototype` using a `__proto__` payload. <p>Publish Date: 2019-11-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10768>CVE-2019-10768</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a">https://github.com/angular/angular.js/commit/726f49dcf6c23106ddaf5cfd5e2e592841db743a</a></p> <p>Release Date: 2019-11-19</p> <p>Fix Resolution: angularjs - 1.7.8,1.7.8.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in angular js angular min js cve high severity vulnerability vulnerable libraries angular js angular min js angular js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to vulnerable library src main webapp js angular angular js dependency hierarchy x angular js vulnerable library angular min js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to vulnerable library src main webapp js angular angular min js dependency hierarchy x angular min js vulnerable library found in head commit a href found in base branch master vulnerability details in angularjs before the function merge could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angularjs step up your open source security game with whitesource
0
476,317
13,736,755,215
IssuesEvent
2020-10-05 12:15:00
wso2/streaming-integrator-tooling
https://api.github.com/repos/wso2/streaming-integrator-tooling
opened
ETL Wizard - Multiple ETL wizards simultaneously do not work
Priority/High Severity/Critical Type/Bug
**Description:** When multiple ETL wizards are opened, they are not working properly. **Affected Product Version:** SI-Tooling 1.1.0
1.0
ETL Wizard - Multiple ETL wizards simultaneously do not work - **Description:** When multiple ETL wizards are opened, they are not working properly. **Affected Product Version:** SI-Tooling 1.1.0
priority
etl wizard multiple etl wizards simultaneously do not work description when multiple etl wizards are opened they are not working properly affected product version si tooling
1
6,001
8,674,599,627
IssuesEvent
2018-11-30 08:16:12
FundacionParaguaya/MentorApp
https://api.github.com/repos/FundacionParaguaya/MentorApp
reopened
Validation on day and year of birth date
UX Requirement bug
- [ ] Validation of Day and Year in the dropdown its delayed - An error message only appears when the user clicks on the next input field - When the user clicks on the next field it hides the invalid message - thus creating a usability issue - [ ] Validation on year is incorrect - This allows any number even those such as 4444
1.0
Validation on day and year of birth date - - [ ] Validation of Day and Year in the dropdown its delayed - An error message only appears when the user clicks on the next input field - When the user clicks on the next field it hides the invalid message - thus creating a usability issue - [ ] Validation on year is incorrect - This allows any number even those such as 4444
non_priority
validation on day and year of birth date validation of day and year in the dropdown its delayed an error message only appears when the user clicks on the next input field when the user clicks on the next field it hides the invalid message thus creating a usability issue validation on year is incorrect this allows any number even those such as
0
44,558
11,461,490,745
IssuesEvent
2020-02-07 12:04:08
Qiskit/qiskit-aer
https://api.github.com/repos/Qiskit/qiskit-aer
closed
Pass user specified directories to the build system
build system enhancement
<!-- ⚠️ If you do not respect this template, your issue will be closed --> <!-- ⚠️ Make sure to browse the opened and closed issues --> ### Informations - **Qiskit AER version**: 0.1 - **Operating system**: OSX ### What is the current behavior? The make system does not find the OpenBLAS libraries in their default location in OSX. Instead, the Apple Accelerate libs are always used. ### Steps to reproduce the problem ### What is the expected behavior? The installer should find the OpenBLAS libs if they are installed in the default dirs. ### Suggested solutions
1.0
Pass user specified directories to the build system - <!-- ⚠️ If you do not respect this template, your issue will be closed --> <!-- ⚠️ Make sure to browse the opened and closed issues --> ### Informations - **Qiskit AER version**: 0.1 - **Operating system**: OSX ### What is the current behavior? The make system does not find the OpenBLAS libraries in their default location in OSX. Instead, the Apple Accelerate libs are always used. ### Steps to reproduce the problem ### What is the expected behavior? The installer should find the OpenBLAS libs if they are installed in the default dirs. ### Suggested solutions
non_priority
pass user specified directories to the build system informations qiskit aer version operating system osx what is the current behavior the make system does not find the openblas libraries in their default location in osx instead the apple accelerate libs are always used steps to reproduce the problem what is the expected behavior the installer should find the openblas libs if they are installed in the default dirs suggested solutions
0
241,428
7,811,899,804
IssuesEvent
2018-06-12 11:44:09
our-city-app/oca-backend
https://api.github.com/repos/our-city-app/oca-backend
closed
Missing translation in sidebar?
priority_critical state_verification type_bug
The configuration looks the same as Lochristi, but in Niel it's called "Regional/Local News" instead of "Regionaal/Lokaal Nieuws". Are there other apps which have the same problem? ![image-1](https://user-images.githubusercontent.com/17762105/40789807-99d5224a-64f3-11e8-874a-467e7a62de88.jpg) ![image-1](https://user-images.githubusercontent.com/17762105/40789888-d241a6ee-64f3-11e8-9297-60f1dd6d0a04.jpg)
1.0
Missing translation in sidebar? - The configuration looks the same as Lochristi, but in Niel it's called "Regional/Local News" instead of "Regionaal/Lokaal Nieuws". Are there other apps which have the same problem? ![image-1](https://user-images.githubusercontent.com/17762105/40789807-99d5224a-64f3-11e8-874a-467e7a62de88.jpg) ![image-1](https://user-images.githubusercontent.com/17762105/40789888-d241a6ee-64f3-11e8-9297-60f1dd6d0a04.jpg)
priority
missing translation in sidebar the configuration looks the same as lochristi but in niel it s called regional local news instead of regionaal lokaal nieuws are there other apps which have the same problem
1
19,467
5,888,929,298
IssuesEvent
2017-05-17 11:32:32
mrr0088/Python_DataClassification
https://api.github.com/repos/mrr0088/Python_DataClassification
closed
Funcionalidad Clasificación (Parte 3)
code enhancement
Conseguir enlazar los botones de Jinja con el código Python para ejecutar en backend la clasificación de la noticia.
1.0
Funcionalidad Clasificación (Parte 3) - Conseguir enlazar los botones de Jinja con el código Python para ejecutar en backend la clasificación de la noticia.
non_priority
funcionalidad clasificación parte conseguir enlazar los botones de jinja con el código python para ejecutar en backend la clasificación de la noticia
0
117,430
9,935,037,769
IssuesEvent
2019-07-02 15:39:02
somaliaims/Somali-AIMSUI
https://api.github.com/repos/somaliaims/Somali-AIMSUI
closed
If set start and end year of report to 2019 - no projects
july testing
This setting should show all projects active at any point in 2019
1.0
If set start and end year of report to 2019 - no projects - This setting should show all projects active at any point in 2019
non_priority
if set start and end year of report to no projects this setting should show all projects active at any point in
0
763,379
26,754,702,661
IssuesEvent
2023-01-30 22:49:53
nv-morpheus/Morpheus
https://api.github.com/repos/nv-morpheus/Morpheus
reopened
[FEA] Remove Direct `cuda-python` Dependency
feature request Priority 2
**Is your feature request related to a problem? Please describe.** Found in issue #250, the update of `cuda-python` from `11.7.0` to `11.7.1` has broken both CI and the conda build. The main cause is the switch from the conda package `cudatoolkit` to `cuda-toolkit`. The new package doesnt seem to include the necessary driver stubs that are used during linking. To fix this issue, we needed to pin `cuda-python <=11.7.0`. However, this isnt a direct dependency of ours and should be handled by `rmm` and `cudf`. **Describe the solution you'd like** Ideally, once a clear strategy for how to handle the different conda packages `cudatoolkit` and `cuda-toolkit` emerges, RMM and cuDF would handle setting the correct `cuda-python` requirements. When that happens, we should remove our pinned versions of `cuda-python`. **Describe alternatives you've considered** Alternatively, we could add the `cuda-driver-dev` package which does include the CUDA Driver stubs. However, this doesnt solve the issue of which conda package we should be using.
1.0
[FEA] Remove Direct `cuda-python` Dependency - **Is your feature request related to a problem? Please describe.** Found in issue #250, the update of `cuda-python` from `11.7.0` to `11.7.1` has broken both CI and the conda build. The main cause is the switch from the conda package `cudatoolkit` to `cuda-toolkit`. The new package doesnt seem to include the necessary driver stubs that are used during linking. To fix this issue, we needed to pin `cuda-python <=11.7.0`. However, this isnt a direct dependency of ours and should be handled by `rmm` and `cudf`. **Describe the solution you'd like** Ideally, once a clear strategy for how to handle the different conda packages `cudatoolkit` and `cuda-toolkit` emerges, RMM and cuDF would handle setting the correct `cuda-python` requirements. When that happens, we should remove our pinned versions of `cuda-python`. **Describe alternatives you've considered** Alternatively, we could add the `cuda-driver-dev` package which does include the CUDA Driver stubs. However, this doesnt solve the issue of which conda package we should be using.
priority
remove direct cuda python dependency is your feature request related to a problem please describe found in issue the update of cuda python from to has broken both ci and the conda build the main cause is the switch from the conda package cudatoolkit to cuda toolkit the new package doesnt seem to include the necessary driver stubs that are used during linking to fix this issue we needed to pin cuda python however this isnt a direct dependency of ours and should be handled by rmm and cudf describe the solution you d like ideally once a clear strategy for how to handle the different conda packages cudatoolkit and cuda toolkit emerges rmm and cudf would handle setting the correct cuda python requirements when that happens we should remove our pinned versions of cuda python describe alternatives you ve considered alternatively we could add the cuda driver dev package which does include the cuda driver stubs however this doesnt solve the issue of which conda package we should be using
1
713,400
24,527,460,056
IssuesEvent
2022-10-11 14:05:34
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
When reading the properties of `RpcException`, an unrelated exception may be thrown
kind/bug lang/C# area/api priority/P2
I found this in 2.36.4 but it appears to still be an issue in the latest. When reading the properties of `RcpException`, an unrelated exception can be thrown. Specifically, `RpcException.Trailers` is of type `Metadata` and `Metadata.Value` can throw an exception from its getter. This is generally bad practice, and is recommended against by Microsoft. The fact that exceptions are often serialized for logging exacerbates the issue, as this would obfuscate the initial exception. Whether or not to call this a "bug" is subjective since this appears to be by design, but I think it's clearly a flaw that should be addressed either way. From a more general design perspective, it's generally a bad idea to throw an exception inside a property getter or setter. If the behavior involved could explicitly throw an exception, it's better to use a method instead of a property. Microsoft's guidance on this [can be found here](https://docs.microsoft.com/en-us/dotnet/standard/design-guidelines/property?redirectedfrom=MSDN). Relevant excerpt (emphasis mine): > ❌ AVOID throwing exceptions from property getters. > > Property getters should be simple operations and should not have any preconditions. If a getter can throw an exception, **it should probably be redesigned to be a method**. Notice that this rule does not apply to indexers, where we do expect exceptions as a result of validating the arguments. `RpcException.Trailers` is of type `Metadata`. `Metadata.Value` is a property as follows: ``` public string Value { get { GrpcPreconditions.CheckState(!IsBinary, "Cannot access string value of a binary metadata entry"); return value; } } ``` When the metadata is recognized as binary, an exception is thrown. This can interrupt exception handling for the RpcException and make it difficult to diagnose the issue. In our case, when our logger attempts to parse one of these during serialization, the resulting exception obfuscates the real exception. One option would be to change it to a getter method such as `bool TryGetValue(out string value)` - Prevents it from being read automatically by serializers - Method name makes it clear to the caller that it may fail - Fails safely (return `false` instead of throwing an exception)
1.0
When reading the properties of `RpcException`, an unrelated exception may be thrown - I found this in 2.36.4 but it appears to still be an issue in the latest. When reading the properties of `RcpException`, an unrelated exception can be thrown. Specifically, `RpcException.Trailers` is of type `Metadata` and `Metadata.Value` can throw an exception from its getter. This is generally bad practice, and is recommended against by Microsoft. The fact that exceptions are often serialized for logging exacerbates the issue, as this would obfuscate the initial exception. Whether or not to call this a "bug" is subjective since this appears to be by design, but I think it's clearly a flaw that should be addressed either way. From a more general design perspective, it's generally a bad idea to throw an exception inside a property getter or setter. If the behavior involved could explicitly throw an exception, it's better to use a method instead of a property. Microsoft's guidance on this [can be found here](https://docs.microsoft.com/en-us/dotnet/standard/design-guidelines/property?redirectedfrom=MSDN). Relevant excerpt (emphasis mine): > ❌ AVOID throwing exceptions from property getters. > > Property getters should be simple operations and should not have any preconditions. If a getter can throw an exception, **it should probably be redesigned to be a method**. Notice that this rule does not apply to indexers, where we do expect exceptions as a result of validating the arguments. `RpcException.Trailers` is of type `Metadata`. `Metadata.Value` is a property as follows: ``` public string Value { get { GrpcPreconditions.CheckState(!IsBinary, "Cannot access string value of a binary metadata entry"); return value; } } ``` When the metadata is recognized as binary, an exception is thrown. This can interrupt exception handling for the RpcException and make it difficult to diagnose the issue. In our case, when our logger attempts to parse one of these during serialization, the resulting exception obfuscates the real exception. One option would be to change it to a getter method such as `bool TryGetValue(out string value)` - Prevents it from being read automatically by serializers - Method name makes it clear to the caller that it may fail - Fails safely (return `false` instead of throwing an exception)
priority
when reading the properties of rpcexception an unrelated exception may be thrown i found this in but it appears to still be an issue in the latest when reading the properties of rcpexception an unrelated exception can be thrown specifically rpcexception trailers is of type metadata and metadata value can throw an exception from its getter this is generally bad practice and is recommended against by microsoft the fact that exceptions are often serialized for logging exacerbates the issue as this would obfuscate the initial exception whether or not to call this a bug is subjective since this appears to be by design but i think it s clearly a flaw that should be addressed either way from a more general design perspective it s generally a bad idea to throw an exception inside a property getter or setter if the behavior involved could explicitly throw an exception it s better to use a method instead of a property microsoft s guidance on this relevant excerpt emphasis mine ❌ avoid throwing exceptions from property getters property getters should be simple operations and should not have any preconditions if a getter can throw an exception it should probably be redesigned to be a method notice that this rule does not apply to indexers where we do expect exceptions as a result of validating the arguments rpcexception trailers is of type metadata metadata value is a property as follows public string value get grpcpreconditions checkstate isbinary cannot access string value of a binary metadata entry return value when the metadata is recognized as binary an exception is thrown this can interrupt exception handling for the rpcexception and make it difficult to diagnose the issue in our case when our logger attempts to parse one of these during serialization the resulting exception obfuscates the real exception one option would be to change it to a getter method such as bool trygetvalue out string value prevents it from being read automatically by serializers method name makes it clear to the caller that it may fail fails safely return false instead of throwing an exception
1
410,506
11,992,452,507
IssuesEvent
2020-04-08 10:08:15
PostHog/posthog
https://api.github.com/repos/PostHog/posthog
opened
Add SMTP credentials in UX
enhancement priority-6
**Is your feature request related to a problem? Please describe.** For users to have a password reset, they currently have to set env variables. This isn't user friendly. **Describe the solution you'd like** In the settings page of the app, be able to set: ``` EMAIL_HOST EMAIL_PORT EMAIL_HOST_USER EMAIL_HOST_PASSWORD EMAIL_USE_TLS EMAIL_USE_SSL DEFAULT_FROM_EMAIL ``` **Describe alternatives you've considered** Set environment variables.
1.0
Add SMTP credentials in UX - **Is your feature request related to a problem? Please describe.** For users to have a password reset, they currently have to set env variables. This isn't user friendly. **Describe the solution you'd like** In the settings page of the app, be able to set: ``` EMAIL_HOST EMAIL_PORT EMAIL_HOST_USER EMAIL_HOST_PASSWORD EMAIL_USE_TLS EMAIL_USE_SSL DEFAULT_FROM_EMAIL ``` **Describe alternatives you've considered** Set environment variables.
priority
add smtp credentials in ux is your feature request related to a problem please describe for users to have a password reset they currently have to set env variables this isn t user friendly describe the solution you d like in the settings page of the app be able to set email host email port email host user email host password email use tls email use ssl default from email describe alternatives you ve considered set environment variables
1
290,522
25,073,326,738
IssuesEvent
2022-11-07 13:49:56
jackmcgrath20088629/notes-app-2
https://api.github.com/repos/jackmcgrath20088629/notes-app-2
closed
New Functionality - Update a Note in the Notes Collection
enhancement test driven development
This new functionality should allow the user to update a note in the notes collection. The flow should be: - All notes [along with the index number in the ArrayList] should be printed to the console. - The user should be prompted to enter the index number of the note they wish to update. - The user is then asked to enter new data for the title, priority, and category. Note: Archive status will be managed in later versions. - The note is then updated - the user is informed of whether the update was successful or not (e.g if the user entered an index number that was not valid, this would result in a note not being updated.) The JUnit tests associated with this functionality should be completed as part of this issue.
1.0
New Functionality - Update a Note in the Notes Collection - This new functionality should allow the user to update a note in the notes collection. The flow should be: - All notes [along with the index number in the ArrayList] should be printed to the console. - The user should be prompted to enter the index number of the note they wish to update. - The user is then asked to enter new data for the title, priority, and category. Note: Archive status will be managed in later versions. - The note is then updated - the user is informed of whether the update was successful or not (e.g if the user entered an index number that was not valid, this would result in a note not being updated.) The JUnit tests associated with this functionality should be completed as part of this issue.
non_priority
new functionality update a note in the notes collection this new functionality should allow the user to update a note in the notes collection the flow should be all notes should be printed to the console the user should be prompted to enter the index number of the note they wish to update the user is then asked to enter new data for the title priority and category note archive status will be managed in later versions the note is then updated the user is informed of whether the update was successful or not e g if the user entered an index number that was not valid this would result in a note not being updated the junit tests associated with this functionality should be completed as part of this issue
0
733,306
25,299,828,766
IssuesEvent
2022-11-17 09:52:40
software-mansion/starknet.py
https://api.github.com/repos/software-mansion/starknet.py
closed
Investigate if methods using token_address must be updated
question priority: high
It seems like something related to token address was changed in general config: perhaps methods like get_balance need updating. ![image](https://user-images.githubusercontent.com/52135326/200853878-9727f54c-d8ca-4384-b6c9-15448f7b3d71.png)
1.0
Investigate if methods using token_address must be updated - It seems like something related to token address was changed in general config: perhaps methods like get_balance need updating. ![image](https://user-images.githubusercontent.com/52135326/200853878-9727f54c-d8ca-4384-b6c9-15448f7b3d71.png)
priority
investigate if methods using token address must be updated it seems like something related to token address was changed in general config perhaps methods like get balance need updating
1
6,870
15,686,961,734
IssuesEvent
2021-03-25 13:09:24
nikolagsiderov/APPartment
https://api.github.com/repos/nikolagsiderov/APPartment
closed
Introduce Areas in routing
General Architecture General UI
During development of #271 from @fr0wBtW, I figured it would be best if we separated the whole controllers' structure in areas...
1.0
Introduce Areas in routing - During development of #271 from @fr0wBtW, I figured it would be best if we separated the whole controllers' structure in areas...
non_priority
introduce areas in routing during development of from i figured it would be best if we separated the whole controllers structure in areas
0
33,982
2,774,133,775
IssuesEvent
2015-05-04 05:21:38
galenp/PROPERTYCOMPASS.ISSUES
https://api.github.com/repos/galenp/PROPERTYCOMPASS.ISSUES
closed
Payment details page - Credit Card number and Total field
1.3 Type-Suggestion 2.2 Priority-Low 3.5 Status-Resolved
Please refer attached image thanks, ![paymentdetails-carddetails and total field-29april2015](https://cloud.githubusercontent.com/assets/6513000/7390878/bd838976-ee99-11e4-9736-676411d6af5a.png)
1.0
Payment details page - Credit Card number and Total field - Please refer attached image thanks, ![paymentdetails-carddetails and total field-29april2015](https://cloud.githubusercontent.com/assets/6513000/7390878/bd838976-ee99-11e4-9736-676411d6af5a.png)
priority
payment details page credit card number and total field please refer attached image thanks
1
672,293
22,820,549,178
IssuesEvent
2022-07-12 01:32:27
LuanRT/YouTube.js
https://api.github.com/repos/LuanRT/YouTube.js
opened
Properly organize parser classes
enhancement good first issue priority: low
### Describe your suggestion The parser is mostly complete at this point, but I reckon the `classes` folder is a it messy. With f924a394091feda6f38168eb84ccb38ab1c701ce we should be able to put Renderers/DA classes in specific folders — that would allow us to find problematic code more easily. I already started doing that but only with files related to recently added features such as `LiveChat` and `Comments` ### Other details _No response_ ### Checklist - [X] I am running the latest version. - [X] I checked the documentation and found no answer. - [X] I have searched the existing issues and made sure this is not a duplicate. - [X] I have provided sufficient information.
1.0
Properly organize parser classes - ### Describe your suggestion The parser is mostly complete at this point, but I reckon the `classes` folder is a it messy. With f924a394091feda6f38168eb84ccb38ab1c701ce we should be able to put Renderers/DA classes in specific folders — that would allow us to find problematic code more easily. I already started doing that but only with files related to recently added features such as `LiveChat` and `Comments` ### Other details _No response_ ### Checklist - [X] I am running the latest version. - [X] I checked the documentation and found no answer. - [X] I have searched the existing issues and made sure this is not a duplicate. - [X] I have provided sufficient information.
priority
properly organize parser classes describe your suggestion the parser is mostly complete at this point but i reckon the classes folder is a it messy with we should be able to put renderers da classes in specific folders — that would allow us to find problematic code more easily i already started doing that but only with files related to recently added features such as livechat and comments other details no response checklist i am running the latest version i checked the documentation and found no answer i have searched the existing issues and made sure this is not a duplicate i have provided sufficient information
1
4,491
6,996,111,643
IssuesEvent
2017-12-15 22:26:44
amcnulty/Whoop-It-Up
https://api.github.com/repos/amcnulty/Whoop-It-Up
closed
Mobile User page
Requirement
Basic mobile layout for the landing page - bootstrap 4 - modify grunt file (if needed) to compile scss - basic JS, no animation needed yet - basic CSS/JS for avatar selection - "Update Profile" button require backend support
1.0
Mobile User page - Basic mobile layout for the landing page - bootstrap 4 - modify grunt file (if needed) to compile scss - basic JS, no animation needed yet - basic CSS/JS for avatar selection - "Update Profile" button require backend support
non_priority
mobile user page basic mobile layout for the landing page bootstrap modify grunt file if needed to compile scss basic js no animation needed yet basic css js for avatar selection update profile button require backend support
0
53,389
13,161,489,515
IssuesEvent
2020-08-10 19:38:32
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
FNX-4726 ⁃ Dependency substitution for android-components
E5 P2 eng:build
This will allow us to easily develop against a local checkout of `android-components`. Currently blocked on a-c moving to AndroidX (https://github.com/mozilla-mobile/android-components/issues/842). Also, see similar r-b issue: https://github.com/mozilla-mobile/reference-browser/issues/365
1.0
FNX-4726 ⁃ Dependency substitution for android-components - This will allow us to easily develop against a local checkout of `android-components`. Currently blocked on a-c moving to AndroidX (https://github.com/mozilla-mobile/android-components/issues/842). Also, see similar r-b issue: https://github.com/mozilla-mobile/reference-browser/issues/365
non_priority
fnx ⁃ dependency substitution for android components this will allow us to easily develop against a local checkout of android components currently blocked on a c moving to androidx also see similar r b issue
0
215,825
16,707,166,899
IssuesEvent
2021-06-09 11:24:55
ita-social-projects/EventsExpress
https://api.github.com/repos/ita-social-projects/EventsExpress
opened
Verify that the event owner can create an online event and add a link to that event
test
**Date Time** 06-09-2021 15:15 **Priority** Medium **Description** This test verifies if the event owner can provide a link to the online event. **Pre-requisites** 1. Move to https://eventsexpress-test.azurewebsites.net/home/events?page=1&status=active 2. Authorize as a user (E-Mail: eventsexpress.testuser3@gmail.com Password: 1qaz2wsx1qaz2wsx) 3. 'Home' page is opened **Test Cases** | S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated | |:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:| | 1<br> | Click on the 'Add event' button<br> | <br> | The ‘Edit event draft’ page is open<br> | <br> | Pass<br> | <br> | | 2<br> | Fill in the 'Title', 'Max Count Of Participants', 'Date', 'Description' and ' Hashtags' fields<br> | e.g. 'Нова онлайн подія', 35, 'Подія для провірки встановлення посилання на онлайн подію', '#travel'<br> | <br> | <br> | Pass<br> | <br> | | 3<br> | Select the 'Online' radio button<br> | <br> | The input field appeared under the 'Online' radio button<br> | <br> | Pass<br> | <br> | | 4<br> | Insert the event link in the input field<br> | e.g. https://www.google.com.ua<br> | <br> | <br> | Pass<br> | <br> | | 5<br> | Сlick 'Save' button<br> | <br> | In the lower left corner of the page a message about the successful saving of the event appears<br> | <br> | Pass<br> | <br> | **Environment:** - OS: [Windows 10 (64 bit)] - Browser [Chrome Version 91.0.4472.77] - Version [20210609.3]
1.0
Verify that the event owner can create an online event and add a link to that event - **Date Time** 06-09-2021 15:15 **Priority** Medium **Description** This test verifies if the event owner can provide a link to the online event. **Pre-requisites** 1. Move to https://eventsexpress-test.azurewebsites.net/home/events?page=1&status=active 2. Authorize as a user (E-Mail: eventsexpress.testuser3@gmail.com Password: 1qaz2wsx1qaz2wsx) 3. 'Home' page is opened **Test Cases** | S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated | |:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:| | 1<br> | Click on the 'Add event' button<br> | <br> | The ‘Edit event draft’ page is open<br> | <br> | Pass<br> | <br> | | 2<br> | Fill in the 'Title', 'Max Count Of Participants', 'Date', 'Description' and ' Hashtags' fields<br> | e.g. 'Нова онлайн подія', 35, 'Подія для провірки встановлення посилання на онлайн подію', '#travel'<br> | <br> | <br> | Pass<br> | <br> | | 3<br> | Select the 'Online' radio button<br> | <br> | The input field appeared under the 'Online' radio button<br> | <br> | Pass<br> | <br> | | 4<br> | Insert the event link in the input field<br> | e.g. https://www.google.com.ua<br> | <br> | <br> | Pass<br> | <br> | | 5<br> | Сlick 'Save' button<br> | <br> | In the lower left corner of the page a message about the successful saving of the event appears<br> | <br> | Pass<br> | <br> | **Environment:** - OS: [Windows 10 (64 bit)] - Browser [Chrome Version 91.0.4472.77] - Version [20210609.3]
non_priority
verify that the event owner can create an online event and add a link to that event date time priority medium description this test verifies if the event owner can provide a link to the online event pre requisites move to authorize as a user e mail eventsexpress gmail com password home page is opened test cases s action test data expected result actual result p f automated click on the add event button the ‘edit event draft’ page is open pass fill in the title max count of participants date description and hashtags fields e g нова онлайн подія подія для провірки встановлення посилання на онлайн подію travel pass select the online radio button the input field appeared under the online radio button pass insert the event link in the input field e g pass сlick save button in the lower left corner of the page a message about the successful saving of the event appears pass environment os browser version
0
329,621
10,022,309,203
IssuesEvent
2019-07-16 16:23:10
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
Authenticating a user with Google is throwing 500 on the 2nd login
priority: medium status: have to reproduce type: bug 🐛
<!-- ⚠️ If you do not respect this template your issue will be closed. --> <!-- =============================================================================== --> <!-- ⚠️ If you are not using the current Strapi release, you will be asked to update. --> <!-- Please see the wiki for guides on upgrading to the latest release. --> <!-- =============================================================================== --> <!-- ⚠️ Make sure to browse the opened and closed issues before submitting your issue. --> <!-- ⚠️ Before writing your issue make sure you are using:--> <!-- Node 10.x.x --> <!-- npm 6.x.x --> <!-- The latest version of Strapi. --> **Informations** - **Node.js version**: 10.13.0 - **npm version**: 6.4.1 - **Strapi version**: v3.0.0-alpha.14.5 - **Database**: mongodb 3.6.5 - **Operating system**: Win10 **What is the current behavior?** After the first successful authentication with google, additional attempts to authenticate, at least within the next minutes, fail at redirecting to GET /auth/google/callback as long as the user remains signed in with google in the browser. Response is 500 at http://localhost:4200/auth/callback/google?error%5Berror%5D=invalid_grant&error%5Berror_description%5D=Malformed%20auth%20code. Signing off from Google fixes this issue and returns jwt and user response after redirecting to GET /auth/google/callback **Steps to reproduce the problem** Authenticate with google Stay signed in with google Authenticate again with google **What is the expected behavior?** After the first authentication subsequent authentication attempts should also issue a new token. **Suggested solutions** I'm not quite sure why this happens since the initial redirect from accounts.google.com/o/oauth2/auth to /connect/google/callback looks pretty much the same on the first and subsequent calls, only difference i could recognize was that the first redirect is encoded while the next ones are not First attempt, working: http://localhost:1337/connect/google/callback?code=4%2FogD_bJ10kjNs7l8gqTF25hLsYQPUU-rkatS6jK5shBXcw-lLN0wlJpbSDbyx8zFP2yuDyVLDA1ScgSxLaZxxxxx&scope=openid+email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fuserinfo.email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fplus.me Next attempt, failing: http://localhost:1337/connect/google/callback?code=4/ogCWJEaBayzfL2PoY4NJ3rL-11Wu4FNjrPOxRlFG003qPuGCy7nV6qCnlX1vvh_Dt1cUmJZrFn5ESxxxxx&scope=openid+email+https://www.googleapis.com/auth/plus.me+https://www.googleapis.com/auth/userinfo.email
1.0
Authenticating a user with Google is throwing 500 on the 2nd login - <!-- ⚠️ If you do not respect this template your issue will be closed. --> <!-- =============================================================================== --> <!-- ⚠️ If you are not using the current Strapi release, you will be asked to update. --> <!-- Please see the wiki for guides on upgrading to the latest release. --> <!-- =============================================================================== --> <!-- ⚠️ Make sure to browse the opened and closed issues before submitting your issue. --> <!-- ⚠️ Before writing your issue make sure you are using:--> <!-- Node 10.x.x --> <!-- npm 6.x.x --> <!-- The latest version of Strapi. --> **Informations** - **Node.js version**: 10.13.0 - **npm version**: 6.4.1 - **Strapi version**: v3.0.0-alpha.14.5 - **Database**: mongodb 3.6.5 - **Operating system**: Win10 **What is the current behavior?** After the first successful authentication with google, additional attempts to authenticate, at least within the next minutes, fail at redirecting to GET /auth/google/callback as long as the user remains signed in with google in the browser. Response is 500 at http://localhost:4200/auth/callback/google?error%5Berror%5D=invalid_grant&error%5Berror_description%5D=Malformed%20auth%20code. Signing off from Google fixes this issue and returns jwt and user response after redirecting to GET /auth/google/callback **Steps to reproduce the problem** Authenticate with google Stay signed in with google Authenticate again with google **What is the expected behavior?** After the first authentication subsequent authentication attempts should also issue a new token. **Suggested solutions** I'm not quite sure why this happens since the initial redirect from accounts.google.com/o/oauth2/auth to /connect/google/callback looks pretty much the same on the first and subsequent calls, only difference i could recognize was that the first redirect is encoded while the next ones are not First attempt, working: http://localhost:1337/connect/google/callback?code=4%2FogD_bJ10kjNs7l8gqTF25hLsYQPUU-rkatS6jK5shBXcw-lLN0wlJpbSDbyx8zFP2yuDyVLDA1ScgSxLaZxxxxx&scope=openid+email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fuserinfo.email+https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fplus.me Next attempt, failing: http://localhost:1337/connect/google/callback?code=4/ogCWJEaBayzfL2PoY4NJ3rL-11Wu4FNjrPOxRlFG003qPuGCy7nV6qCnlX1vvh_Dt1cUmJZrFn5ESxxxxx&scope=openid+email+https://www.googleapis.com/auth/plus.me+https://www.googleapis.com/auth/userinfo.email
priority
authenticating a user with google is throwing on the login informations node js version npm version strapi version alpha database mongodb operating system what is the current behavior after the first successful authentication with google additional attempts to authenticate at least within the next minutes fail at redirecting to get auth google callback as long as the user remains signed in with google in the browser response is at signing off from google fixes this issue and returns jwt and user response after redirecting to get auth google callback steps to reproduce the problem authenticate with google stay signed in with google authenticate again with google what is the expected behavior after the first authentication subsequent authentication attempts should also issue a new token suggested solutions i m not quite sure why this happens since the initial redirect from accounts google com o auth to connect google callback looks pretty much the same on the first and subsequent calls only difference i could recognize was that the first redirect is encoded while the next ones are not first attempt working next attempt failing
1
260,463
22,622,957,629
IssuesEvent
2022-06-30 08:12:42
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.3.5
release test/4.3.5
Main issue gathering all the tasks regarding the preparation and validation of Wazuh version 4.3.5 release. Close the issue only after the GA release. ## Release info | | | |---------------------------------------|------------------------------------------------------| | Release name| Wazuh v4.3.5 | | Tag | https://github.com/wazuh/wazuh/tree/v4.3.5 | | Branch | https://github.com/wazuh/wazuh/tree/4.3 | | Changelog | https://github.com/wazuh/wazuh/blob/4.3/CHANGELOG.md | | Latest release candidate (GA version) | N/A | | ## Release candidates As usual, in order to publish, one or more release candidate versions are generated and fully tested, in the links below you can check the status and evolution of each version. | Version | Issue | Tag | Status | |---------|-------|-------------------------------------------------|-----------| | RC1 | #13966 | [v4.3.5-rc1](https://github.com/wazuh/wazuh/tree/v4.3.5-rc1) | Completed | - _In progress: Fixes are being implemented._ - _In review: All the fixes have been implemented. The RC is now tagged and the QA process kicks off._ - _Closed: Fixes and QA process has finished._ ## Tests All the tests results for this release can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1) ## Fixes/Known issues All the fixes and the issues detected in the testing phase that won't be solved in this version can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1)
1.0
Release 4.3.5 - Main issue gathering all the tasks regarding the preparation and validation of Wazuh version 4.3.5 release. Close the issue only after the GA release. ## Release info | | | |---------------------------------------|------------------------------------------------------| | Release name| Wazuh v4.3.5 | | Tag | https://github.com/wazuh/wazuh/tree/v4.3.5 | | Branch | https://github.com/wazuh/wazuh/tree/4.3 | | Changelog | https://github.com/wazuh/wazuh/blob/4.3/CHANGELOG.md | | Latest release candidate (GA version) | N/A | | ## Release candidates As usual, in order to publish, one or more release candidate versions are generated and fully tested, in the links below you can check the status and evolution of each version. | Version | Issue | Tag | Status | |---------|-------|-------------------------------------------------|-----------| | RC1 | #13966 | [v4.3.5-rc1](https://github.com/wazuh/wazuh/tree/v4.3.5-rc1) | Completed | - _In progress: Fixes are being implemented._ - _In review: All the fixes have been implemented. The RC is now tagged and the QA process kicks off._ - _Closed: Fixes and QA process has finished._ ## Tests All the tests results for this release can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1) ## Fixes/Known issues All the fixes and the issues detected in the testing phase that won't be solved in this version can be found in the [v4.3.5 project](https://github.com/orgs/wazuh/projects/13/views/1)
non_priority
release main issue gathering all the tasks regarding the preparation and validation of wazuh version release close the issue only after the ga release release info release name wazuh tag branch changelog latest release candidate ga version n a release candidates as usual in order to publish one or more release candidate versions are generated and fully tested in the links below you can check the status and evolution of each version version issue tag status completed in progress fixes are being implemented in review all the fixes have been implemented the rc is now tagged and the qa process kicks off closed fixes and qa process has finished tests all the tests results for this release can be found in the fixes known issues all the fixes and the issues detected in the testing phase that won t be solved in this version can be found in the
0
21,378
29,202,228,594
IssuesEvent
2023-05-21 00:36:56
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Product Manager na Coodesh
SALVADOR GESTÃO DE PROJETOS JIRA REQUISITOS REMOTO PROCESSOS GITHUB UMA POWER BI APIs NEGÓCIOS PRODUCT MANAGER Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Techsocial</strong> está em busca de <strong><ins>Product Manager</ins></strong> para compor seu time!</p> <p></p> <p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas. A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais.&nbsp;</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Entendimento dos módulos do nossa Plataforma;&nbsp;</li> <li>Conhecimento dos nossos negócios e suas peculiaridades;</li> <li>Identificação de gaps e oportunidades de melhoria;&nbsp;</li> <li>Auxílio no mapeamento do processo e levantamento de requisitos de mudanças;</li> <li>Confecção de tickets de melhoria alinhado ao time de processos;&nbsp;</li> <li>Apresentação e revisão de tickets junto ao time de desenvolvimento;&nbsp;</li> <li>Definição de prioridades de implementação;&nbsp;</li> <li>Identificação de interfaces do(s) módulo(s) que representa para que requisitos e regras não sejam alterados sem alinhamento e definição com outros módulos, etc.</li> </ul> ## Techsocial: <p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas.</p> <p>A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. Nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas.</p><a href='https://coodesh.com/empresas/techsocial-tecnologia-e-consultoria-ltda'>Veja mais no site</a> ## Habilidades: - JIRA - Análise de requisitos - Gestão e Negociação com Cliente ## Local: 100% Remoto ## Requisitos: - Conhecimento em Levantamento de requisitos; - Mapeamento de Processos; - Experiência em Gestão de Projetos; - Experiência em Gestão de Produtos. ## Diferenciais: - Power BI; - Experíência na Plataforma Jira. ## Benefícios: - Convênio Médico; - Trabalho Remoto. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Product Manager na Techsocial](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Gestão em TI
1.0
[Remoto] Product Manager na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Techsocial</strong> está em busca de <strong><ins>Product Manager</ins></strong> para compor seu time!</p> <p></p> <p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas. A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais.&nbsp;</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Entendimento dos módulos do nossa Plataforma;&nbsp;</li> <li>Conhecimento dos nossos negócios e suas peculiaridades;</li> <li>Identificação de gaps e oportunidades de melhoria;&nbsp;</li> <li>Auxílio no mapeamento do processo e levantamento de requisitos de mudanças;</li> <li>Confecção de tickets de melhoria alinhado ao time de processos;&nbsp;</li> <li>Apresentação e revisão de tickets junto ao time de desenvolvimento;&nbsp;</li> <li>Definição de prioridades de implementação;&nbsp;</li> <li>Identificação de interfaces do(s) módulo(s) que representa para que requisitos e regras não sejam alterados sem alinhamento e definição com outros módulos, etc.</li> </ul> ## Techsocial: <p>Somos uma empresa de Soluções Tecnológicas, que busca transformar os dados e informações de nossos clientes em resultados. Evoluímos a partir de consultoria em Gestão Empresarial, somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas.</p> <p>A Tech é uma empresa inovadora! Desenvolvemos e aportamos inteligência em softwares, aplicativos, RPAs, APIs entre outras soluções digitais. Nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas.</p><a href='https://coodesh.com/empresas/techsocial-tecnologia-e-consultoria-ltda'>Veja mais no site</a> ## Habilidades: - JIRA - Análise de requisitos - Gestão e Negociação com Cliente ## Local: 100% Remoto ## Requisitos: - Conhecimento em Levantamento de requisitos; - Mapeamento de Processos; - Experiência em Gestão de Projetos; - Experiência em Gestão de Produtos. ## Diferenciais: - Power BI; - Experíência na Plataforma Jira. ## Benefícios: - Convênio Médico; - Trabalho Remoto. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Product Manager na Techsocial](https://coodesh.com/vagas/product-manager-152815535?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Gestão em TI
non_priority
product manager na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a techsocial está em busca de product manager para compor seu time somos uma empresa de soluções tecnológicas que busca transformar os dados e informações de nossos clientes em resultados evoluímos a partir de consultoria em gestão empresarial somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas a tech é uma empresa inovadora desenvolvemos e aportamos inteligência em softwares aplicativos rpas apis entre outras soluções digitais nbsp responsabilidades entendimento dos módulos do nossa plataforma nbsp conhecimento dos nossos negócios e suas peculiaridades identificação de gaps e oportunidades de melhoria nbsp auxílio no mapeamento do processo e levantamento de requisitos de mudanças confecção de tickets de melhoria alinhado ao time de processos nbsp apresentação e revisão de tickets junto ao time de desenvolvimento nbsp definição de prioridades de implementação nbsp identificação de interfaces do s módulo s que representa para que requisitos e regras não sejam alterados sem alinhamento e definição com outros módulos etc techsocial somos uma empresa de soluções tecnológicas que busca transformar os dados e informações de nossos clientes em resultados evoluímos a partir de consultoria em gestão empresarial somando as múltiplas competências e experiência de nossos profissionais às inovações tecnológicas a tech é uma empresa inovadora desenvolvemos e aportamos inteligência em softwares aplicativos rpas apis entre outras soluções digitais nossa missão é simplificar os processos de nossos clientes por meio da tecnologia e estruturar grandes bancos de dados para garimparmos e lapidarmos as melhores informações para as empresas habilidades jira análise de requisitos gestão e negociação com cliente local remoto requisitos conhecimento em levantamento de requisitos mapeamento de processos experiência em gestão de projetos experiência em gestão de produtos diferenciais power bi experíência na plataforma jira benefícios convênio médico trabalho remoto como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria gestão em ti
0
517,461
15,013,967,458
IssuesEvent
2021-02-01 05:39:45
TerriaJS/nationalmap
https://api.github.com/repos/TerriaJS/nationalmap
closed
NationalMap - Jan 2021 release (v8)
High priority
**HIGH PRIORITY:** - [ ] AREMI catalogue transfer to NatMap https://github.com/TerriaJS/aremi-natmap/issues/437 - Data transfer mapping is [HERE ](https://csiroau.sharepoint.com/:w:/r/sites/WebDataPlatformsgroup/Shared%20Documents/General/Terria/AREMI/Data/Decommissioning%20-%20data%20transfer%20to%20NatMap/AREMI%20data%20transfer%20to%20NatMap%20-%20mapping%20to%20catalogue.docx?d=w807583b6b8684cfa85cbc4fefa064729&csf=1&web=1&e=DqTrGg) V8 porting related issues for AREMI data: - Token implementation for WA services: https://github.com/TerriaJS/aremi-natmap/issues/437 - ISF data - investigate CSS custom sheets and vector data handling: https://github.com/TerriaJS/aremi-natmap/issues/439 - Traffic volumes - NSW transport API - https://github.com/TerriaJS/aremi-natmap/issues/440 - AEMO data - csv polling - feature info disappears after 1 cycle / polling - insert ticket https://github.com/terriajs/terriajs/issues/4768 **MEDIUM PRIORITY:** - [ ] Help page (under About) needs to be deleted - [ ] update some GA layers: https://github.com/TerriaJS/nationalmap/issues/986 - [ ] NationalMap catalogue https://github.com/TerriaJS/nationalmap/issues/985 - needs SDMX fixes and https://github.com/TerriaJS/terriajs/pull/5085 merged - [ ] YourDataYourRegions https://github.com/TerriaJS/terriajs/issues/4861 - [ ] CKAN support for data.gov.au - https://github.com/terriajs/nationalmap/issues/987 (show data with no filter) - [ ] MAGDA API support for data.gov.au - search? - [ ] MAGDA integration _ https://github.com/TerriaJS/nationalmap/issues/981 and https://github.com/TerriaJS/TerriaMap/issues/489 - [ ] GeoJSON - [ ] KML/KMZ - [ ] Boundaries: e.g. http://geoserver.nationalmap.nicta.com.au/admin_bnds_abs/ows - [ ] region mapping - [ ] vector tiles - [ ] csw https://github.com/TerriaJS/terriajs/issues/4693 **LOW PRIORITY** - [ ] NSW decimal issue: https://github.com/TerriaJS/nationalmap/issues/980 - [ ] Update Mobile black spot data position in catalogue - https://github.com/TerriaJS/nationalmap/issues/979
1.0
NationalMap - Jan 2021 release (v8) - **HIGH PRIORITY:** - [ ] AREMI catalogue transfer to NatMap https://github.com/TerriaJS/aremi-natmap/issues/437 - Data transfer mapping is [HERE ](https://csiroau.sharepoint.com/:w:/r/sites/WebDataPlatformsgroup/Shared%20Documents/General/Terria/AREMI/Data/Decommissioning%20-%20data%20transfer%20to%20NatMap/AREMI%20data%20transfer%20to%20NatMap%20-%20mapping%20to%20catalogue.docx?d=w807583b6b8684cfa85cbc4fefa064729&csf=1&web=1&e=DqTrGg) V8 porting related issues for AREMI data: - Token implementation for WA services: https://github.com/TerriaJS/aremi-natmap/issues/437 - ISF data - investigate CSS custom sheets and vector data handling: https://github.com/TerriaJS/aremi-natmap/issues/439 - Traffic volumes - NSW transport API - https://github.com/TerriaJS/aremi-natmap/issues/440 - AEMO data - csv polling - feature info disappears after 1 cycle / polling - insert ticket https://github.com/terriajs/terriajs/issues/4768 **MEDIUM PRIORITY:** - [ ] Help page (under About) needs to be deleted - [ ] update some GA layers: https://github.com/TerriaJS/nationalmap/issues/986 - [ ] NationalMap catalogue https://github.com/TerriaJS/nationalmap/issues/985 - needs SDMX fixes and https://github.com/TerriaJS/terriajs/pull/5085 merged - [ ] YourDataYourRegions https://github.com/TerriaJS/terriajs/issues/4861 - [ ] CKAN support for data.gov.au - https://github.com/terriajs/nationalmap/issues/987 (show data with no filter) - [ ] MAGDA API support for data.gov.au - search? - [ ] MAGDA integration _ https://github.com/TerriaJS/nationalmap/issues/981 and https://github.com/TerriaJS/TerriaMap/issues/489 - [ ] GeoJSON - [ ] KML/KMZ - [ ] Boundaries: e.g. http://geoserver.nationalmap.nicta.com.au/admin_bnds_abs/ows - [ ] region mapping - [ ] vector tiles - [ ] csw https://github.com/TerriaJS/terriajs/issues/4693 **LOW PRIORITY** - [ ] NSW decimal issue: https://github.com/TerriaJS/nationalmap/issues/980 - [ ] Update Mobile black spot data position in catalogue - https://github.com/TerriaJS/nationalmap/issues/979
priority
nationalmap jan release high priority aremi catalogue transfer to natmap data transfer mapping is porting related issues for aremi data token implementation for wa services isf data investigate css custom sheets and vector data handling traffic volumes nsw transport api aemo data csv polling feature info disappears after cycle polling insert ticket medium priority help page under about needs to be deleted update some ga layers nationalmap catalogue needs sdmx fixes and merged yourdatayourregions ckan support for data gov au show data with no filter magda api support for data gov au search magda integration and geojson kml kmz boundaries e g region mapping vector tiles csw low priority nsw decimal issue update mobile black spot data position in catalogue
1
268,400
23,366,318,859
IssuesEvent
2022-08-10 15:38:56
cosmos/ibc-go
https://api.github.com/repos/cosmos/ibc-go
opened
E2E: Timeout IBC transfer
testing
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Word of caution: poorly thought-out proposals may be rejected v without deliberation ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Summary Add a test which times out IBC transfers ____ #### For Admin Use - [ ] Not duplicate issue - [ ] Appropriate labels applied - [ ] Appropriate contributors tagged/assigned
1.0
E2E: Timeout IBC transfer - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Word of caution: poorly thought-out proposals may be rejected v without deliberation ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Summary Add a test which times out IBC transfers ____ #### For Admin Use - [ ] Not duplicate issue - [ ] Appropriate labels applied - [ ] Appropriate contributors tagged/assigned
non_priority
timeout ibc transfer ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v word of caution poorly thought out proposals may be rejected v without deliberation ☺ summary add a test which times out ibc transfers for admin use not duplicate issue appropriate labels applied appropriate contributors tagged assigned
0
40,304
5,199,405,735
IssuesEvent
2017-01-23 20:47:51
azavea/cac-tripplanner
https://api.github.com/repos/azavea/cac-tripplanner
opened
Explore nav button has no effect on Learn page
bug redesign
On staging, go to the Learn page. Click the Explore link in the top nav. Nothing happens. Should go to Explore mode.
1.0
Explore nav button has no effect on Learn page - On staging, go to the Learn page. Click the Explore link in the top nav. Nothing happens. Should go to Explore mode.
non_priority
explore nav button has no effect on learn page on staging go to the learn page click the explore link in the top nav nothing happens should go to explore mode
0
3,925
6,845,689,064
IssuesEvent
2017-11-13 09:17:46
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
closed
Redirected volume removal times out and leaves stale metadata
priority_normal process_wontfix type_bug
Scenario: - config: - redirect timeout 5 secs - backend timeout 15 secs - volume V is running on node N - SCO write W to the backend is happening in the background - volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish - W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires => unlink call times out on N
1.0
Redirected volume removal times out and leaves stale metadata - Scenario: - config: - redirect timeout 5 secs - backend timeout 15 secs - volume V is running on node N - SCO write W to the backend is happening in the background - volume removal is started on node M -> redirect to node N, removal needs to wait for W to finish - W encounters backend issues, finishes (unsuccessfully) when the backend timeout fires => unlink call times out on N
non_priority
redirected volume removal times out and leaves stale metadata scenario config redirect timeout secs backend timeout secs volume v is running on node n sco write w to the backend is happening in the background volume removal is started on node m redirect to node n removal needs to wait for w to finish w encounters backend issues finishes unsuccessfully when the backend timeout fires unlink call times out on n
0
23,626
12,042,422,297
IssuesEvent
2020-04-14 10:32:04
flutter/gallery
https://api.github.com/repos/flutter/gallery
closed
[Performance] Use deferred loading to improve performance
performance
## Acceptance Criteria * The studies load asynchronously after the initial page load * Load non necessary locales asynchronously https://dart.dev/guides/language/language-tour#deferred-loading Migrated from https://github.com/material-components/material-components-flutter-gallery/issues/571
True
[Performance] Use deferred loading to improve performance - ## Acceptance Criteria * The studies load asynchronously after the initial page load * Load non necessary locales asynchronously https://dart.dev/guides/language/language-tour#deferred-loading Migrated from https://github.com/material-components/material-components-flutter-gallery/issues/571
non_priority
use deferred loading to improve performance acceptance criteria the studies load asynchronously after the initial page load load non necessary locales asynchronously migrated from
0
35,025
9,531,211,295
IssuesEvent
2019-04-29 15:29:58
Guestlogix/traveler-ios
https://api.github.com/repos/Guestlogix/traveler-ios
closed
Minor | Text in "Airline Name" and "Flight Number" field disappear when dismissing keyboard on those fields
:bug: bug Future Build iOS pm-review
**Steps:** 1. Launch App 2. Navigate to booking form 3. Click "Airline Name" or "Flight Number" 4. Dismiss keyboard by dragging down on the screen **Actual result:** Keyboard is dismissed but text inside that field also disappears **Expected result:** Keyboard dismisses without text disappearing **Screenshots:** ![ezgif com-video-to-gif 3](https://user-images.githubusercontent.com/25214171/53654134-f6dcea00-3c1a-11e9-8e2d-3f03708b1943.gif) Build 22 0.3(40)
1.0
Minor | Text in "Airline Name" and "Flight Number" field disappear when dismissing keyboard on those fields - **Steps:** 1. Launch App 2. Navigate to booking form 3. Click "Airline Name" or "Flight Number" 4. Dismiss keyboard by dragging down on the screen **Actual result:** Keyboard is dismissed but text inside that field also disappears **Expected result:** Keyboard dismisses without text disappearing **Screenshots:** ![ezgif com-video-to-gif 3](https://user-images.githubusercontent.com/25214171/53654134-f6dcea00-3c1a-11e9-8e2d-3f03708b1943.gif) Build 22 0.3(40)
non_priority
minor text in airline name and flight number field disappear when dismissing keyboard on those fields steps launch app navigate to booking form click airline name or flight number dismiss keyboard by dragging down on the screen actual result keyboard is dismissed but text inside that field also disappears expected result keyboard dismisses without text disappearing screenshots build
0
271,769
8,489,111,978
IssuesEvent
2018-10-26 18:48:00
wevote/WebApp
https://api.github.com/repos/wevote/WebApp
closed
Race Level & Decided State: Save in Store
Difficulty: Medium Priority: 1
### Please describe the issue (What happens? What do you expect?) When a voter dives in to look at a candidate or measure from the Ballot page, and then comes back to the ballot, the "Race Level" (Federal, State, Measure, Local), and the "Decided State" (Remaining Decisions, All Items, Items Decided) both get unset. Store these "states" in a Store so they remain the same when the voter returns to the Ballot.
1.0
Race Level & Decided State: Save in Store - ### Please describe the issue (What happens? What do you expect?) When a voter dives in to look at a candidate or measure from the Ballot page, and then comes back to the ballot, the "Race Level" (Federal, State, Measure, Local), and the "Decided State" (Remaining Decisions, All Items, Items Decided) both get unset. Store these "states" in a Store so they remain the same when the voter returns to the Ballot.
priority
race level decided state save in store please describe the issue what happens what do you expect when a voter dives in to look at a candidate or measure from the ballot page and then comes back to the ballot the race level federal state measure local and the decided state remaining decisions all items items decided both get unset store these states in a store so they remain the same when the voter returns to the ballot
1
26,516
26,903,976,433
IssuesEvent
2023-02-06 17:33:21
DCS-LCSR/SignStream3
https://api.github.com/repos/DCS-LCSR/SignStream3
opened
Changing Utterance name UI not working as expected
bug severity LOW usability concern
Changing the name of an utterance appears to have some bugs. From the User's Guide... > 4. To edit the label (i.e., assign or change the name of Utterance), click on the label > in the Utterance Bar of the Temporal Partition and hold the mouse down until > the text becomes editable (just as you would edit a filename in the Finder in Mac > OS). A visible text box for the label will appear. Edit the text of the label, and hit > Return. The label will also change for all associated Segment Bars, as illustrated > in Figure 29. But clicking on the name and holding in SS does not work as it does in Mac Finder. You can get it to pop up the name change box other ways, but it's a bit buggy.
True
Changing Utterance name UI not working as expected - Changing the name of an utterance appears to have some bugs. From the User's Guide... > 4. To edit the label (i.e., assign or change the name of Utterance), click on the label > in the Utterance Bar of the Temporal Partition and hold the mouse down until > the text becomes editable (just as you would edit a filename in the Finder in Mac > OS). A visible text box for the label will appear. Edit the text of the label, and hit > Return. The label will also change for all associated Segment Bars, as illustrated > in Figure 29. But clicking on the name and holding in SS does not work as it does in Mac Finder. You can get it to pop up the name change box other ways, but it's a bit buggy.
non_priority
changing utterance name ui not working as expected changing the name of an utterance appears to have some bugs from the user s guide to edit the label i e assign or change the name of utterance click on the label in the utterance bar of the temporal partition and hold the mouse down until the text becomes editable just as you would edit a filename in the finder in mac os a visible text box for the label will appear edit the text of the label and hit return the label will also change for all associated segment bars as illustrated in figure but clicking on the name and holding in ss does not work as it does in mac finder you can get it to pop up the name change box other ways but it s a bit buggy
0
631,357
20,151,151,704
IssuesEvent
2022-02-09 12:31:48
ita-social-projects/horondi_admin
https://api.github.com/repos/ita-social-projects/horondi_admin
closed
(SP:1)Material initial values don`t show on form
bug priority: high Admin
Steps to reproduce: 1) Go to https://horondi-admin-staging.azurewebsites.net/materials 2) Open material Actual result: ![image](https://user-images.githubusercontent.com/37375602/114228053-83bc0b00-997e-11eb-84c2-5ef46db94329.png) Expected result: Initial values should be shown in material form for each material
1.0
(SP:1)Material initial values don`t show on form - Steps to reproduce: 1) Go to https://horondi-admin-staging.azurewebsites.net/materials 2) Open material Actual result: ![image](https://user-images.githubusercontent.com/37375602/114228053-83bc0b00-997e-11eb-84c2-5ef46db94329.png) Expected result: Initial values should be shown in material form for each material
priority
sp material initial values don t show on form steps to reproduce go to open material actual result expected result initial values should be shown in material form for each material
1
108,028
4,325,558,759
IssuesEvent
2016-07-26 00:28:34
syb0rg/Khronos
https://api.github.com/repos/syb0rg/Khronos
closed
CMake Rebuild Problems
Priority: Medium Status: Completed Status: Pending Type: Bug
Upon rebuilding, CMake sometimes fails to download external dependencies. Deleting the project's folder in the `libs` solves the issue.
1.0
CMake Rebuild Problems - Upon rebuilding, CMake sometimes fails to download external dependencies. Deleting the project's folder in the `libs` solves the issue.
priority
cmake rebuild problems upon rebuilding cmake sometimes fails to download external dependencies deleting the project s folder in the libs solves the issue
1
120,540
15,776,445,503
IssuesEvent
2021-04-01 04:42:44
microsoft/react-native-windows
https://api.github.com/repos/microsoft/react-native-windows
closed
How to deal with NM breaking changes
Agenda Area: Native Modules enhancement needs dev design
Meta-point: we have breaking changes for NM across releases (C# for 0.63, e.g.). How do we deal with keeping community modules working despite breaking changes in the platform? Breaking changes in Native Modules should have a higher bar than other breaking changes. Do we need to make the breaking change now at all? _Originally posted by @asklar in https://github.com/microsoft/react-native-windows/issues/6444#issuecomment-724225605_
1.0
How to deal with NM breaking changes - Meta-point: we have breaking changes for NM across releases (C# for 0.63, e.g.). How do we deal with keeping community modules working despite breaking changes in the platform? Breaking changes in Native Modules should have a higher bar than other breaking changes. Do we need to make the breaking change now at all? _Originally posted by @asklar in https://github.com/microsoft/react-native-windows/issues/6444#issuecomment-724225605_
non_priority
how to deal with nm breaking changes meta point we have breaking changes for nm across releases c for e g how do we deal with keeping community modules working despite breaking changes in the platform breaking changes in native modules should have a higher bar than other breaking changes do we need to make the breaking change now at all originally posted by asklar in
0
54,636
7,904,502,742
IssuesEvent
2018-07-02 05:05:01
SalatielBairros/HomeLibrary
https://api.github.com/repos/SalatielBairros/HomeLibrary
opened
Terminar documentação inicial
documentation help wanted
Terminar documentação inicial do projeto para entrega do Marco Avaliativo 2.
1.0
Terminar documentação inicial - Terminar documentação inicial do projeto para entrega do Marco Avaliativo 2.
non_priority
terminar documentação inicial terminar documentação inicial do projeto para entrega do marco avaliativo
0
34,006
16,373,389,657
IssuesEvent
2021-05-15 15:59:27
Dart-Code/Dart-Code
https://api.github.com/repos/Dart-Code/Dart-Code
closed
Unusable in Large Monorepo
awaiting info in editor in lsp/analysis server is performance
We operate a large (~230k lines) monorepo consisting of several Flutter teams, several apps and several dozens of packages all using various Dart versions (null-safe and non-null-safe, though the entire monorepo uses Flutter 2.0.5). As the monorepo has continued to grow, intellisense, auto-complete, auto-import etc performance in VSCode has continued to degrade. We're very heavy users of Git sparse-checkout to remove apps and packages that aren't relevant for people working on specific teams. For awhile this actually solved all of our performance problems in VSCode. Now however, even the smallest team's sparse-checkout is ~18% of the monorepo and performance is as bad as ever. **To Reproduce** Have a big monorepo. **Expected behavior** Intellisense, auto-complete, auto-import should work. **Versions (please complete the following information):** - VS Code version: 1.56.0 - Dart extension version: 3.22.0 - Dart/Flutter SDK version: 2.0.6 Is there a way to capture a performance trace that can be submitted somewhere?
True
Unusable in Large Monorepo - We operate a large (~230k lines) monorepo consisting of several Flutter teams, several apps and several dozens of packages all using various Dart versions (null-safe and non-null-safe, though the entire monorepo uses Flutter 2.0.5). As the monorepo has continued to grow, intellisense, auto-complete, auto-import etc performance in VSCode has continued to degrade. We're very heavy users of Git sparse-checkout to remove apps and packages that aren't relevant for people working on specific teams. For awhile this actually solved all of our performance problems in VSCode. Now however, even the smallest team's sparse-checkout is ~18% of the monorepo and performance is as bad as ever. **To Reproduce** Have a big monorepo. **Expected behavior** Intellisense, auto-complete, auto-import should work. **Versions (please complete the following information):** - VS Code version: 1.56.0 - Dart extension version: 3.22.0 - Dart/Flutter SDK version: 2.0.6 Is there a way to capture a performance trace that can be submitted somewhere?
non_priority
unusable in large monorepo we operate a large lines monorepo consisting of several flutter teams several apps and several dozens of packages all using various dart versions null safe and non null safe though the entire monorepo uses flutter as the monorepo has continued to grow intellisense auto complete auto import etc performance in vscode has continued to degrade we re very heavy users of git sparse checkout to remove apps and packages that aren t relevant for people working on specific teams for awhile this actually solved all of our performance problems in vscode now however even the smallest team s sparse checkout is of the monorepo and performance is as bad as ever to reproduce have a big monorepo expected behavior intellisense auto complete auto import should work versions please complete the following information vs code version dart extension version dart flutter sdk version is there a way to capture a performance trace that can be submitted somewhere
0
105,160
9,037,227,639
IssuesEvent
2019-02-09 08:14:08
chartjs/chartjs-plugin-datalabels
https://api.github.com/repos/chartjs/chartjs-plugin-datalabels
closed
[ask] get error from script source
needs test case
hello, i try this plugin on my html, but when i put the script then reload the page, i get 2 error like this.. ![screenshot from 2019-02-06 10-25-52](https://user-images.githubusercontent.com/17553775/52318955-a39bc100-29f9-11e9-82d4-5bf06e083230.png) ![screenshot from 2019-02-06 10-18-49](https://user-images.githubusercontent.com/17553775/52318958-a5fe1b00-29f9-11e9-8937-7994e3253a85.png) do you guys know why this happen? Thanks!
1.0
[ask] get error from script source - hello, i try this plugin on my html, but when i put the script then reload the page, i get 2 error like this.. ![screenshot from 2019-02-06 10-25-52](https://user-images.githubusercontent.com/17553775/52318955-a39bc100-29f9-11e9-82d4-5bf06e083230.png) ![screenshot from 2019-02-06 10-18-49](https://user-images.githubusercontent.com/17553775/52318958-a5fe1b00-29f9-11e9-8937-7994e3253a85.png) do you guys know why this happen? Thanks!
non_priority
get error from script source hello i try this plugin on my html but when i put the script then reload the page i get error like this do you guys know why this happen thanks
0
139,372
12,855,294,750
IssuesEvent
2020-07-09 04:46:13
rnpgp/rnp
https://api.github.com/repos/rnpgp/rnp
opened
Update Fuzzer section of `develop.adoc` due to #1195
documentation
From @ni4 : > LGTM, except minor copyright year mismatch. > And it would be also good to update Fuzzer section of develop.adoc once this is merged.
1.0
Update Fuzzer section of `develop.adoc` due to #1195 - From @ni4 : > LGTM, except minor copyright year mismatch. > And it would be also good to update Fuzzer section of develop.adoc once this is merged.
non_priority
update fuzzer section of develop adoc due to from lgtm except minor copyright year mismatch and it would be also good to update fuzzer section of develop adoc once this is merged
0
67,393
12,953,373,853
IssuesEvent
2020-07-20 00:20:58
Abbassihraf/P-curiosity-LAB
https://api.github.com/repos/Abbassihraf/P-curiosity-LAB
closed
Programm details
Code In progress
### **All versions** - [x] Include nav - [x] Include hero header - [ ] Programms details section - [x] Include footer
1.0
Programm details - ### **All versions** - [x] Include nav - [x] Include hero header - [ ] Programms details section - [x] Include footer
non_priority
programm details all versions include nav include hero header programms details section include footer
0
171,995
6,497,755,605
IssuesEvent
2017-08-22 15:00:02
fossasia/susi_skill_cms
https://api.github.com/repos/fossasia/susi_skill_cms
closed
Follow web chat and use same icons in dropdown
bug Priority: High
![screenshot from 2017-08-02 14-21-26](https://user-images.githubusercontent.com/1583873/28873836-8c23ebc4-778f-11e7-9a33-514c1f60ae6c.png) Please implement and use the same icons.
1.0
Follow web chat and use same icons in dropdown - ![screenshot from 2017-08-02 14-21-26](https://user-images.githubusercontent.com/1583873/28873836-8c23ebc4-778f-11e7-9a33-514c1f60ae6c.png) Please implement and use the same icons.
priority
follow web chat and use same icons in dropdown please implement and use the same icons
1
102,088
11,274,343,838
IssuesEvent
2020-01-14 18:21:57
lxc/lxd
https://api.github.com/repos/lxc/lxd
closed
Multi-architecture clustering
API Documentation Easy Feature
LXD clustering can be used to turn multiple LXD servers into one large instance. Right now, this assumes that all servers in the cluster are of the same architecture. While that's certainly the common case, there are times where it would be useful to have a single LXD cluster which supports multiple architectures, usually a mix of Intel and Arm hardware. To make this possible, we'd need to: - Register the native architecture of each server in the `nodes` database table - Update the generated /1.0 output to advertise all architectures supported by the cluster - Update the default placement rule such that if asked to deploy an image specific to a given architecture, we find the least busy cluster member that supports that architecture
1.0
Multi-architecture clustering - LXD clustering can be used to turn multiple LXD servers into one large instance. Right now, this assumes that all servers in the cluster are of the same architecture. While that's certainly the common case, there are times where it would be useful to have a single LXD cluster which supports multiple architectures, usually a mix of Intel and Arm hardware. To make this possible, we'd need to: - Register the native architecture of each server in the `nodes` database table - Update the generated /1.0 output to advertise all architectures supported by the cluster - Update the default placement rule such that if asked to deploy an image specific to a given architecture, we find the least busy cluster member that supports that architecture
non_priority
multi architecture clustering lxd clustering can be used to turn multiple lxd servers into one large instance right now this assumes that all servers in the cluster are of the same architecture while that s certainly the common case there are times where it would be useful to have a single lxd cluster which supports multiple architectures usually a mix of intel and arm hardware to make this possible we d need to register the native architecture of each server in the nodes database table update the generated output to advertise all architectures supported by the cluster update the default placement rule such that if asked to deploy an image specific to a given architecture we find the least busy cluster member that supports that architecture
0
57,435
11,756,093,988
IssuesEvent
2020-03-13 10:50:20
fac19/week2-hklo
https://api.github.com/repos/fac19/week2-hklo
closed
G R E A T J O B 😍
code review compliment
I love the concept of the game - you've done an amazing job fetching and displaying the data on the page! Well done you, this is great! And... a super mega well done for working from home. I really admire what you've managed to create together working remotely. Wow! PS: I'm a big fan of your game :) ![](https://media.giphy.com/media/5mTwWeN3MoS08/giphy.gif)
1.0
G R E A T J O B 😍 - I love the concept of the game - you've done an amazing job fetching and displaying the data on the page! Well done you, this is great! And... a super mega well done for working from home. I really admire what you've managed to create together working remotely. Wow! PS: I'm a big fan of your game :) ![](https://media.giphy.com/media/5mTwWeN3MoS08/giphy.gif)
non_priority
g r e a t j o b 😍 i love the concept of the game you ve done an amazing job fetching and displaying the data on the page well done you this is great and a super mega well done for working from home i really admire what you ve managed to create together working remotely wow ps i m a big fan of your game
0
35,726
7,800,101,936
IssuesEvent
2018-06-09 04:49:46
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012 (Trac #491)
Migrated from Trac clubb_src defect roehl@uwm.edu
The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear. Steven, could you please investigate? Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/491 ```json { "status": "closed", "changetime": "2012-02-07T15:32:28", "description": "The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear.\n\nSteven, could you please investigate?", "reporter": "vlarson@uwm.edu", "cc": "vlarson@uwm.edu, dschanen@uwm.edu, connork@uwm.edu", "resolution": "fixed", "_ts": "1328628748000000", "component": "clubb_src", "summary": "Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012", "priority": "critical", "keywords": "", "time": "2012-01-30T12:21:11", "milestone": "", "owner": "roehl@uwm.edu", "type": "defect" } ```
1.0
Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012 (Trac #491) - The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear. Steven, could you please investigate? Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/491 ```json { "status": "closed", "changetime": "2012-02-07T15:32:28", "description": "The jpg images are missing when I try to display the nightly plots for CAM, GFDL, and SAM. However, CLUBB standalone plots do appear.\n\nSteven, could you please investigate?", "reporter": "vlarson@uwm.edu", "cc": "vlarson@uwm.edu, dschanen@uwm.edu, connork@uwm.edu", "resolution": "fixed", "_ts": "1328628748000000", "component": "clubb_src", "summary": "Nightly plots do not appear for CAM, GFDL, and SAM on 29 Jan 2012", "priority": "critical", "keywords": "", "time": "2012-01-30T12:21:11", "milestone": "", "owner": "roehl@uwm.edu", "type": "defect" } ```
non_priority
nightly plots do not appear for cam gfdl and sam on jan trac the jpg images are missing when i try to display the nightly plots for cam gfdl and sam however clubb standalone plots do appear steven could you please investigate attachments migrated from json status closed changetime description the jpg images are missing when i try to display the nightly plots for cam gfdl and sam however clubb standalone plots do appear n nsteven could you please investigate reporter vlarson uwm edu cc vlarson uwm edu dschanen uwm edu connork uwm edu resolution fixed ts component clubb src summary nightly plots do not appear for cam gfdl and sam on jan priority critical keywords time milestone owner roehl uwm edu type defect
0
366,310
10,819,566,544
IssuesEvent
2019-11-08 14:39:30
raz0red/wii-mednafen
https://api.github.com/repos/raz0red/wii-mednafen
closed
NTFS Support
Priority-Medium Type-Enhancement auto-migrated
``` Please add Support to NTFS Hard Disk! ``` Original issue reported on code.google.com by `cardelli...@gmail.com` on 3 Mar 2012 at 9:41
1.0
NTFS Support - ``` Please add Support to NTFS Hard Disk! ``` Original issue reported on code.google.com by `cardelli...@gmail.com` on 3 Mar 2012 at 9:41
priority
ntfs support please add support to ntfs hard disk original issue reported on code google com by cardelli gmail com on mar at
1
251,860
21,526,307,611
IssuesEvent
2022-04-28 18:49:29
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
server: TestClusterVersionUpgrade failed
C-test-failure O-robot GA-blocker T-server-and-security branch-release-22.1.0
server.TestClusterVersionUpgrade [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=artifacts#/) on release-22.1.0 @ [dc225064c5f50ae59bd2216283b5a1ee13adbd1d](https://github.com/cockroachdb/cockroach/commits/dc225064c5f50ae59bd2216283b5a1ee13adbd1d): ``` === RUN TestClusterVersionUpgrade test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/ab5462d2cb3989e2a33018db970f23de/logTestClusterVersionUpgrade2444000157 test_log_scope.go:80: use -show-logs to present logs inline version_cluster_test.go:271: condition failed to evaluate within 45s: 0: v21.2-112 active=true (wanted false) goroutine 520664 [running]: runtime/debug.Stack() GOROOT/src/runtime/debug/stack.go:24 +0x65 github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithin({0x54dd058, 0xc000ce11e0}, 0x0, 0x406aa60) github.com/cockroachdb/cockroach/pkg/testutils/soon.go:60 +0x5f github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon({0x54dd058, 0xc000ce11e0}, 0xc006a11710) github.com/cockroachdb/cockroach/pkg/testutils/soon.go:41 +0x4a github.com/cockroachdb/cockroach/pkg/server_test.TestClusterVersionUpgrade(0xc000ce11e0) github.com/cockroachdb/cockroach/pkg/server_test/pkg/server/version_cluster_test.go:271 +0x896 testing.tRunner(0xc000ce11e0, 0x4494a48) GOROOT/src/testing/testing.go:1259 +0x102 created by testing.(*T).Run GOROOT/src/testing/testing.go:1306 +0x35a panic.go:642: -- test log scope end -- --- FAIL: TestClusterVersionUpgrade (46.93s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss </p> </details> <details><summary>Same failure on other branches</summary> <p> - #79162 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-21.2] - #78126 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-22.1] </p> </details> /cc @cockroachdb/server <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestClusterVersionUpgrade.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
server: TestClusterVersionUpgrade failed - server.TestClusterVersionUpgrade [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4938477&tab=artifacts#/) on release-22.1.0 @ [dc225064c5f50ae59bd2216283b5a1ee13adbd1d](https://github.com/cockroachdb/cockroach/commits/dc225064c5f50ae59bd2216283b5a1ee13adbd1d): ``` === RUN TestClusterVersionUpgrade test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/ab5462d2cb3989e2a33018db970f23de/logTestClusterVersionUpgrade2444000157 test_log_scope.go:80: use -show-logs to present logs inline version_cluster_test.go:271: condition failed to evaluate within 45s: 0: v21.2-112 active=true (wanted false) goroutine 520664 [running]: runtime/debug.Stack() GOROOT/src/runtime/debug/stack.go:24 +0x65 github.com/cockroachdb/cockroach/pkg/testutils.SucceedsWithin({0x54dd058, 0xc000ce11e0}, 0x0, 0x406aa60) github.com/cockroachdb/cockroach/pkg/testutils/soon.go:60 +0x5f github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon({0x54dd058, 0xc000ce11e0}, 0xc006a11710) github.com/cockroachdb/cockroach/pkg/testutils/soon.go:41 +0x4a github.com/cockroachdb/cockroach/pkg/server_test.TestClusterVersionUpgrade(0xc000ce11e0) github.com/cockroachdb/cockroach/pkg/server_test/pkg/server/version_cluster_test.go:271 +0x896 testing.tRunner(0xc000ce11e0, 0x4494a48) GOROOT/src/testing/testing.go:1259 +0x102 created by testing.(*T).Run GOROOT/src/testing/testing.go:1306 +0x35a panic.go:642: -- test log scope end -- --- FAIL: TestClusterVersionUpgrade (46.93s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=bazel,gss </p> </details> <details><summary>Same failure on other branches</summary> <p> - #79162 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-21.2] - #78126 server: TestClusterVersionUpgrade failed [C-test-failure O-robot T-server-and-security branch-release-22.1] </p> </details> /cc @cockroachdb/server <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestClusterVersionUpgrade.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_priority
server testclusterversionupgrade failed server testclusterversionupgrade with on release run testclusterversionupgrade test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline version cluster test go condition failed to evaluate within active true wanted false goroutine runtime debug stack goroot src runtime debug stack go github com cockroachdb cockroach pkg testutils succeedswithin github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg testutils succeedssoon github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg server test testclusterversionupgrade github com cockroachdb cockroach pkg server test pkg server version cluster test go testing trunner goroot src testing testing go created by testing t run goroot src testing testing go panic go test log scope end fail testclusterversionupgrade help see also parameters in this failure tags bazel gss same failure on other branches server testclusterversionupgrade failed server testclusterversionupgrade failed cc cockroachdb server
0
107,149
16,751,637,914
IssuesEvent
2021-06-12 01:34:59
gms-ws-demo/nibrs
https://api.github.com/repos/gms-ws-demo/nibrs
opened
CVE-2018-14718 (High) detected in multiple libraries
security vulnerability
## CVE-2018-14718 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.8.10.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-flatfile/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-staging-data/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.8.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-demo/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization. <p>Publish Date: 2019-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718>CVE-2018-14718</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14718">https://nvd.nist.gov/vuln/detail/CVE-2018-14718</a></p> <p>Release Date: 2019-01-02</p> <p>Fix Resolution: 2.9.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-staging-data/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-14718","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-14718 (High) detected in multiple libraries - ## CVE-2018-14718 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.8.10.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-flatfile/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-staging-data/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.8.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar,nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-demo/nibrs/commit/9fb1c19bd26c2113d1961640de126a33eacdc946">9fb1c19bd26c2113d1961640de126a33eacdc946</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization. <p>Publish Date: 2019-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718>CVE-2018-14718</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14718">https://nvd.nist.gov/vuln/detail/CVE-2018-14718</a></p> <p>Release Date: 2019-01-02</p> <p>Fix Resolution: 2.9.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-staging-data/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-validation/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-14718","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the slf4j-ext class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14718","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs flatfile pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs staging data pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar nibrs web nibrs web target nibrs web web inf lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs tools nibrs fbi service pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar nibrs tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the ext class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the ext class from polymorphic deserialization vulnerabilityurl
0
93
2,502,435,182
IssuesEvent
2015-01-09 08:56:27
dhowe/AdNauseam
https://api.github.com/repos/dhowe/AdNauseam
closed
Things to finish for before video screencasts
1 - Awaiting Code PRIORITY: High
@dhowe these are the things we still need to fix before I can make the screencasts: - [x] release a fix to the borked #132 - [x] implement popup #70 - [ ] display text ads #112 - [ ] add (non-tracking) share buttons to the site #134 - [ ] use a placeholder image for banners #122 - [ ] show the ad visiting animation #116 - [x] implement new logo #96 - [x] finish ADN site #33 (almost)
1.0
Things to finish for before video screencasts - @dhowe these are the things we still need to fix before I can make the screencasts: - [x] release a fix to the borked #132 - [x] implement popup #70 - [ ] display text ads #112 - [ ] add (non-tracking) share buttons to the site #134 - [ ] use a placeholder image for banners #122 - [ ] show the ad visiting animation #116 - [x] implement new logo #96 - [x] finish ADN site #33 (almost)
non_priority
things to finish for before video screencasts dhowe these are the things we still need to fix before i can make the screencasts release a fix to the borked implement popup display text ads add non tracking share buttons to the site use a placeholder image for banners show the ad visiting animation implement new logo finish adn site almost
0
20,862
14,221,132,893
IssuesEvent
2020-11-17 15:20:36
spdk/spdk
https://api.github.com/repos/spdk/spdk
closed
Vagrant VM startup failure (libvirt)
Infrastructure Intermittent Failure
Failure log shows: 00:00:31.690 ==> default: -> value=-drive, 00:00:31.690 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-zero.img,if=none,id=nvme-2, 00:00:31.690 ==> default: -> value=-device, 00:00:31.690 ==> default: -> value=nvme,drive=nvme-2,serial=12342,namespaces=0, 00:00:31.691 ==> default: -> value=-drive, 00:00:31.691 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img,if=none,id=ocssd-3, 00:00:31.692 ==> default: -> value=-device, 00:00:31.692 ==> default: -> value=nvme,drive=ocssd-3,serial=deadbeef,oacs=0,namespaces=1,lver=2,lba_index=4,mdts=10,lnum_lun=8,lnum_pln=2,lpgs_per_blk=1536,lsecs_per_pg=2,lblks_per_pln=45,metadata=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_md,nsdatafile=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_blknvme.ns,laer_thread_sleep=3000,stride=4, 00:00:32.501 ==> default: Creating shared folders metadata... 00:00:32.536 ==> default: Starting domain. 00:00:32.943 There was an error talking to Libvirt. The error message is shown 00:00:32.943 below: 00:00:32.943 00:00:32.943 Call to virDomainCreateWithFlags failed: internal error: process exited while connecting to monitor: qemu-system-x86_64: -realtime mlock=off: warning: '-realtime mlock=...' is deprecated, please use '-overcommit mem-lock=...' instead 00:00:32.943 2020-10-20T17:47:47.443254Z qemu-system-x86_64: unable to map backing store for guest RAM: Cannot allocate memory 00:00:32.963 + cd fedora31-libvirt 00:00:32.963 + vagrant destroy -f 00:00:35.969 ==> default: Removing domain... 00:00:36.032 Build step 'Execute shell' marked build as failure 00:00:36.102 [PostBuildScript] - [INFO] Executing post build scripts. 00:00:36.108 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins9062853704202237667.sh 00:00:36.112 + [[ -n fedora31 ]] 00:00:36.112 + mv /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/fedora31-libvirt/output /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/output 00:00:36.129 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins7126983447848921706.sh
1.0
Vagrant VM startup failure (libvirt) - Failure log shows: 00:00:31.690 ==> default: -> value=-drive, 00:00:31.690 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-zero.img,if=none,id=nvme-2, 00:00:31.690 ==> default: -> value=-device, 00:00:31.690 ==> default: -> value=nvme,drive=nvme-2,serial=12342,namespaces=0, 00:00:31.691 ==> default: -> value=-drive, 00:00:31.691 ==> default: -> value=format=raw,file=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img,if=none,id=ocssd-3, 00:00:31.692 ==> default: -> value=-device, 00:00:31.692 ==> default: -> value=nvme,drive=ocssd-3,serial=deadbeef,oacs=0,namespaces=1,lver=2,lba_index=4,mdts=10,lnum_lun=8,lnum_pln=2,lpgs_per_blk=1536,lsecs_per_pg=2,lblks_per_pln=45,metadata=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_md,nsdatafile=/var/lib/libvirt/images/backends/ex0-nvme-ocssd.img_ocssd_blknvme.ns,laer_thread_sleep=3000,stride=4, 00:00:32.501 ==> default: Creating shared folders metadata... 00:00:32.536 ==> default: Starting domain. 00:00:32.943 There was an error talking to Libvirt. The error message is shown 00:00:32.943 below: 00:00:32.943 00:00:32.943 Call to virDomainCreateWithFlags failed: internal error: process exited while connecting to monitor: qemu-system-x86_64: -realtime mlock=off: warning: '-realtime mlock=...' is deprecated, please use '-overcommit mem-lock=...' instead 00:00:32.943 2020-10-20T17:47:47.443254Z qemu-system-x86_64: unable to map backing store for guest RAM: Cannot allocate memory 00:00:32.963 + cd fedora31-libvirt 00:00:32.963 + vagrant destroy -f 00:00:35.969 ==> default: Removing domain... 00:00:36.032 Build step 'Execute shell' marked build as failure 00:00:36.102 [PostBuildScript] - [INFO] Executing post build scripts. 00:00:36.108 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins9062853704202237667.sh 00:00:36.112 + [[ -n fedora31 ]] 00:00:36.112 + mv /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/fedora31-libvirt/output /var/jenkins/workspace/nvmf-tcp-uring-vg-autotest/output 00:00:36.129 [nvmf-tcp-uring-vg-autotest] $ /usr/bin/env bash /tmp/jenkins7126983447848921706.sh
non_priority
vagrant vm startup failure libvirt failure log shows default value drive default value format raw file var lib libvirt images backends nvme zero img if none id nvme default value device default value nvme drive nvme serial namespaces default value drive default value format raw file var lib libvirt images backends nvme ocssd img if none id ocssd default value device default value nvme drive ocssd serial deadbeef oacs namespaces lver lba index mdts lnum lun lnum pln lpgs per blk lsecs per pg lblks per pln metadata var lib libvirt images backends nvme ocssd img ocssd md nsdatafile var lib libvirt images backends nvme ocssd img ocssd blknvme ns laer thread sleep stride default creating shared folders metadata default starting domain there was an error talking to libvirt the error message is shown below call to virdomaincreatewithflags failed internal error process exited while connecting to monitor qemu system realtime mlock off warning realtime mlock is deprecated please use overcommit mem lock instead qemu system unable to map backing store for guest ram cannot allocate memory cd libvirt vagrant destroy f default removing domain build step execute shell marked build as failure executing post build scripts usr bin env bash tmp sh mv var jenkins workspace nvmf tcp uring vg autotest libvirt output var jenkins workspace nvmf tcp uring vg autotest output usr bin env bash tmp sh
0
34,140
6,299,691,103
IssuesEvent
2017-07-21 00:09:35
dgraph-io/dgraph
https://api.github.com/repos/dgraph-io/dgraph
closed
Update Go client documentation
documentation in progress
The client has a new API. We should update the documentation at docs.dgraph.io and in the Godoc.
1.0
Update Go client documentation - The client has a new API. We should update the documentation at docs.dgraph.io and in the Godoc.
non_priority
update go client documentation the client has a new api we should update the documentation at docs dgraph io and in the godoc
0
517,703
15,018,708,058
IssuesEvent
2021-02-01 12:36:14
buidl-labs/crypto-code-school-inside-tezos
https://api.github.com/repos/buidl-labs/crypto-code-school-inside-tezos
opened
Auth flow
Priority: Medium Type: Enhancement
Edge case that needs to be handled: 1. If the wallet is Uninstall after log in. Don't allow users to login. 2. Update copy for creating account in case where If no account is found.
1.0
Auth flow - Edge case that needs to be handled: 1. If the wallet is Uninstall after log in. Don't allow users to login. 2. Update copy for creating account in case where If no account is found.
priority
auth flow edge case that needs to be handled if the wallet is uninstall after log in don t allow users to login update copy for creating account in case where if no account is found
1
397,285
27,158,460,109
IssuesEvent
2023-02-17 09:53:21
appsmithorg/appsmith-docs
https://api.github.com/repos/appsmithorg/appsmith-docs
closed
[Docs]: Button Widget
Documentation Doc Rehaul User Education Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Documentation Link https://docs.appsmith.com/reference/widgets/button ### Discord/slack/intercom Link _No response_ ### Describe the problem Restructuring ### Describe the improvement Rehaul
1.0
[Docs]: Button Widget - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Documentation Link https://docs.appsmith.com/reference/widgets/button ### Discord/slack/intercom Link _No response_ ### Describe the problem Restructuring ### Describe the improvement Rehaul
non_priority
button widget is there an existing issue for this i have searched the existing issues documentation link discord slack intercom link no response describe the problem restructuring describe the improvement rehaul
0
161,824
20,156,384,692
IssuesEvent
2022-02-09 16:49:49
ioana-nicolae/renovate2
https://api.github.com/repos/ioana-nicolae/renovate2
opened
CVE-2017-16137 (Medium) detected in debug-2.2.0.tgz
security vulnerability
## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/debug/package.json</p> <p> Dependency Hierarchy: - :x: **debug-2.2.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/renovate2/commit/9d0f2692d355dfff0d0ba18825cd40f2074a31f9">9d0f2692d355dfff0d0ba18825cd40f2074a31f9</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 2.6.9</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-16137 (Medium) detected in debug-2.2.0.tgz - ## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/debug/package.json</p> <p> Dependency Hierarchy: - :x: **debug-2.2.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/renovate2/commit/9d0f2692d355dfff0d0ba18825cd40f2074a31f9">9d0f2692d355dfff0d0ba18825cd40f2074a31f9</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 2.6.9</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in debug tgz cve medium severity vulnerability vulnerable library debug tgz small debugging utility library home page a href path to dependency file package json path to vulnerable library node modules debug package json dependency hierarchy x debug tgz vulnerable library found in head commit a href found in base branch main vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree debug isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue vulnerabilityurl
0
250,292
7,974,832,021
IssuesEvent
2018-07-17 07:23:42
poanetwork/token-wizard
https://api.github.com/repos/poanetwork/token-wizard
reopened
(Bug) Contribution page:countdown timer has incorrect status after finalization
Step6-Contribute UI bug high priority migration to auth-os
Steps: 1. Create crowdsale with any parameter, ex. :Minted, 1 tier, duration 2 min, minCap 2. Wait until crowdsale ended 3. As an owner open manage page and finalize 4. Open contribution page and observe countdown timer. Expected result: -timer's status is "CROWDSALE HAS BEEN FINALIZED"" Actual result: -timer's status is "CROWDSALE HAS ENDED" If you are reporting a problem with Token Wizard, please include the following information: ### Which network did you use? (Mainnet, Kovan, Rinkeby, etc.) *localhost* ### If you were able to create it, what is the URL of your crowdsale? *n/a* ### Do you have screenshots showing the problem? <img width="1440" alt="screen shot 2018-07-11 at 00 08 37" src="https://user-images.githubusercontent.com/14067096/42556205-b78fd908-849f-11e8-9538-5f621802fc91.png"> ### Do you see errors in the dev console? If yes, please include a screenshot No ---
1.0
(Bug) Contribution page:countdown timer has incorrect status after finalization - Steps: 1. Create crowdsale with any parameter, ex. :Minted, 1 tier, duration 2 min, minCap 2. Wait until crowdsale ended 3. As an owner open manage page and finalize 4. Open contribution page and observe countdown timer. Expected result: -timer's status is "CROWDSALE HAS BEEN FINALIZED"" Actual result: -timer's status is "CROWDSALE HAS ENDED" If you are reporting a problem with Token Wizard, please include the following information: ### Which network did you use? (Mainnet, Kovan, Rinkeby, etc.) *localhost* ### If you were able to create it, what is the URL of your crowdsale? *n/a* ### Do you have screenshots showing the problem? <img width="1440" alt="screen shot 2018-07-11 at 00 08 37" src="https://user-images.githubusercontent.com/14067096/42556205-b78fd908-849f-11e8-9538-5f621802fc91.png"> ### Do you see errors in the dev console? If yes, please include a screenshot No ---
priority
bug contribution page countdown timer has incorrect status after finalization steps create crowdsale with any parameter ex minted tier duration min mincap wait until crowdsale ended as an owner open manage page and finalize open contribution page and observe countdown timer expected result timer s status is crowdsale has been finalized actual result timer s status is crowdsale has ended if you are reporting a problem with token wizard please include the following information which network did you use mainnet kovan rinkeby etc localhost if you were able to create it what is the url of your crowdsale n a do you have screenshots showing the problem img width alt screen shot at src do you see errors in the dev console if yes please include a screenshot no
1
3,176
9,160,856,330
IssuesEvent
2019-03-01 08:52:24
akvo/akvo-platform
https://api.github.com/repos/akvo/akvo-platform
closed
Revisit Authz
architecture in progress to do
* RSR Authz Research the possibility of outsourcing the authorization of request to Keycloak when moving the granularity of authorization from organization to project. * Lumen Authz Work with Lumen team in implementing the fine-grained authorization in the application * Flow Authz Revisit the document created by Dan (1 year ago) on the feasibility of modeling Flow authorization in Keycloak Authz services
1.0
Revisit Authz - * RSR Authz Research the possibility of outsourcing the authorization of request to Keycloak when moving the granularity of authorization from organization to project. * Lumen Authz Work with Lumen team in implementing the fine-grained authorization in the application * Flow Authz Revisit the document created by Dan (1 year ago) on the feasibility of modeling Flow authorization in Keycloak Authz services
non_priority
revisit authz rsr authz research the possibility of outsourcing the authorization of request to keycloak when moving the granularity of authorization from organization to project lumen authz work with lumen team in implementing the fine grained authorization in the application flow authz revisit the document created by dan year ago on the feasibility of modeling flow authorization in keycloak authz services
0