Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
73,684
| 19,761,170,496
|
IssuesEvent
|
2022-01-16 12:47:06
|
chaotic-aur/packages
|
https://api.github.com/repos/chaotic-aur/packages
|
closed
|
[Outdated] python-manimpango
|
request:rebuild-pkg priority:high
|
### If available, link to the latest build
Build log site timed out multiple times.
### Package name
`python-manimpango`
### Latest build
`0.3.1-1`
### Latest version available
`0.4.0.post0-1`
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### More information
In my case, it's being used as a dependency for `emote`.
|
1.0
|
[Outdated] python-manimpango - ### If available, link to the latest build
Build log site timed out multiple times.
### Package name
`python-manimpango`
### Latest build
`0.3.1-1`
### Latest version available
`0.4.0.post0-1`
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### More information
In my case, it's being used as a dependency for `emote`.
|
non_defect
|
python manimpango if available link to the latest build build log site timed out multiple times package name python manimpango latest build latest version available have you tested if the package builds in a clean chroot yes more information in my case it s being used as a dependency for emote
| 0
|
20,307
| 3,332,446,509
|
IssuesEvent
|
2015-11-11 20:08:51
|
lixun910/geoda
|
https://api.github.com/repos/lixun910/geoda
|
closed
|
Average Chart: enable right click menu
|
fixed: verify! Milestone-Release1.8 Module-AveragesTool Priority-Medium Type-Defect
|
```
In versions 1.7.24 ~ 1.7.33:
To init an average chart, there is a prompt "Please use Options->Add/Remove
Variable..".
Here the right click doesn't work, and this message seems duplicated since the
"Add/Remove Variable" dialog is by default showed up.
```
Original issue reported on code.google.com by `lixun...@gmail.com` on 1 Jul 2015 at 6:38
|
1.0
|
Average Chart: enable right click menu - ```
In versions 1.7.24 ~ 1.7.33:
To init an average chart, there is a prompt "Please use Options->Add/Remove
Variable..".
Here the right click doesn't work, and this message seems duplicated since the
"Add/Remove Variable" dialog is by default showed up.
```
Original issue reported on code.google.com by `lixun...@gmail.com` on 1 Jul 2015 at 6:38
|
defect
|
average chart enable right click menu in versions to init an average chart there is a prompt please use options add remove variable here the right click doesn t work and this message seems duplicated since the add remove variable dialog is by default showed up original issue reported on code google com by lixun gmail com on jul at
| 1
|
62,113
| 17,023,854,024
|
IssuesEvent
|
2021-07-03 04:11:33
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
user menu on user page includes empty changesets in count
|
Component: website Priority: trivial Resolution: wontfix Type: defect
|
**[Submitted to the original trac issue database at 4.22pm, Sunday, 17th February 2013]**
The user menu on http://www.openstreetmap.org/user/aseerel4c26/ does count empty changesets (the counter in "Eigene Bearbeitungen 786" increases for each empty changeset) but http://www.openstreetmap.org/user/aseerel4c26/edits does not list empty changesets.
The counting of empty changesets in the website GUI should be consistent - similar bug: https://trac.openstreetmap.org/ticket/2114
|
1.0
|
user menu on user page includes empty changesets in count - **[Submitted to the original trac issue database at 4.22pm, Sunday, 17th February 2013]**
The user menu on http://www.openstreetmap.org/user/aseerel4c26/ does count empty changesets (the counter in "Eigene Bearbeitungen 786" increases for each empty changeset) but http://www.openstreetmap.org/user/aseerel4c26/edits does not list empty changesets.
The counting of empty changesets in the website GUI should be consistent - similar bug: https://trac.openstreetmap.org/ticket/2114
|
defect
|
user menu on user page includes empty changesets in count the user menu on does count empty changesets the counter in eigene bearbeitungen increases for each empty changeset but does not list empty changesets the counting of empty changesets in the website gui should be consistent similar bug
| 1
|
143,272
| 11,540,545,276
|
IssuesEvent
|
2020-02-18 00:29:57
|
davidskalinder/mpeds-coder
|
https://api.github.com/repos/davidskalinder/mpeds-coder
|
reopened
|
Handle uncertainty for coding fields
|
user testing
|
Could be for every item, for particular items, or for groups of items. Probably could be implemented in current DB structure (as separate and arbitrary "variable" types), linked to corresponding items elsewhere (or by name, though that's dubious). As to whether it *should* be stored like that... needs further thought.
|
1.0
|
Handle uncertainty for coding fields - Could be for every item, for particular items, or for groups of items. Probably could be implemented in current DB structure (as separate and arbitrary "variable" types), linked to corresponding items elsewhere (or by name, though that's dubious). As to whether it *should* be stored like that... needs further thought.
|
non_defect
|
handle uncertainty for coding fields could be for every item for particular items or for groups of items probably could be implemented in current db structure as separate and arbitrary variable types linked to corresponding items elsewhere or by name though that s dubious as to whether it should be stored like that needs further thought
| 0
|
77,838
| 27,190,002,041
|
IssuesEvent
|
2023-02-19 17:31:56
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
Data corruption after TRIM
|
Type: Defect
|
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Void Linux
Distribution Version | musl, up to date (rolling)
Kernel Version | 6.1.8_1
Architecture | x86_64
OpenZFS Version | 2.1.7-1
### Describe the problem you're observing
Data corruption in some files and lot of checksum errors.
### Describe how to reproduce the problem
I ran `zpool scrub` on a simple zpool, consisting of a single partition on a 120 GiB SSD. It finished without any errors (neither data or checksum).
The I ran `zpool trim` on the pool. Then a few minutes later I saw 2 checksum errors by the report. Then a few more.
(I'm not sure if the first error appeared before or after the TRIM.)
I ran `zfs trim` again, because the first time it finished quite fast despite there were much space to trim. Few more errors. (Probably not related to running TRIM again, because the errors has been constantly increasing since then.)
Then I was thinking maybe the first scrub before the trim somehow did not notice the errors, so I ran `zfs scrub` again. The errors grown to 100-300 in the first few seconds, so I stopped it.
At this point I couldn't use `zfs send` to save snapshot, so I used `rsync` to backup the data. It reported I/O error for a few files.
It's an SSD, which report 512 blocks size, so I used `ashift=9`. Later I learned that most SSDs actually use larger blocks, but report 512. I don't know if it affected the trim.
The `sda1` BTRFS partition contains ~2 GiB data (kernels, initramfs, and boot related stuff). I run a TRIM and a `btrfs scrub` on it. No errors were found.
A SMART extended self test was running when I first ran the `zfs scrub` (and possibly when I run `zfs trim` the first time, but it could have finished then).
Partition scheme:
```
NAME FSTYPE FSVER LABEL
sda
├─sda1 btrfs VoidBoot
├─sda2 zfs_member 5000 rpool
└─sda3 swap 1 VoidSwap
```
Output of `zpool status` before submitting issue:
```
pool: rpool
state: DEGRADED
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A
scan: scrub canceled on Sun Feb 19 16:29:42 2023
config:
NAME STATE READ WRITE CKSUM
rpool DEGRADED 0 0 0
sda2 DEGRADED 0 0 4.28K too many errors
errors: 481 data errors, use '-v' for a list
```
### Include any warning/errors/backtraces from the system logs
No errors in dmesg.
|
1.0
|
Data corruption after TRIM - ### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Void Linux
Distribution Version | musl, up to date (rolling)
Kernel Version | 6.1.8_1
Architecture | x86_64
OpenZFS Version | 2.1.7-1
### Describe the problem you're observing
Data corruption in some files and lot of checksum errors.
### Describe how to reproduce the problem
I ran `zpool scrub` on a simple zpool, consisting of a single partition on a 120 GiB SSD. It finished without any errors (neither data or checksum).
The I ran `zpool trim` on the pool. Then a few minutes later I saw 2 checksum errors by the report. Then a few more.
(I'm not sure if the first error appeared before or after the TRIM.)
I ran `zfs trim` again, because the first time it finished quite fast despite there were much space to trim. Few more errors. (Probably not related to running TRIM again, because the errors has been constantly increasing since then.)
Then I was thinking maybe the first scrub before the trim somehow did not notice the errors, so I ran `zfs scrub` again. The errors grown to 100-300 in the first few seconds, so I stopped it.
At this point I couldn't use `zfs send` to save snapshot, so I used `rsync` to backup the data. It reported I/O error for a few files.
It's an SSD, which report 512 blocks size, so I used `ashift=9`. Later I learned that most SSDs actually use larger blocks, but report 512. I don't know if it affected the trim.
The `sda1` BTRFS partition contains ~2 GiB data (kernels, initramfs, and boot related stuff). I run a TRIM and a `btrfs scrub` on it. No errors were found.
A SMART extended self test was running when I first ran the `zfs scrub` (and possibly when I run `zfs trim` the first time, but it could have finished then).
Partition scheme:
```
NAME FSTYPE FSVER LABEL
sda
├─sda1 btrfs VoidBoot
├─sda2 zfs_member 5000 rpool
└─sda3 swap 1 VoidSwap
```
Output of `zpool status` before submitting issue:
```
pool: rpool
state: DEGRADED
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-8A
scan: scrub canceled on Sun Feb 19 16:29:42 2023
config:
NAME STATE READ WRITE CKSUM
rpool DEGRADED 0 0 0
sda2 DEGRADED 0 0 4.28K too many errors
errors: 481 data errors, use '-v' for a list
```
### Include any warning/errors/backtraces from the system logs
No errors in dmesg.
|
defect
|
data corruption after trim system information type version name distribution name void linux distribution version musl up to date rolling kernel version architecture openzfs version describe the problem you re observing data corruption in some files and lot of checksum errors describe how to reproduce the problem i ran zpool scrub on a simple zpool consisting of a single partition on a gib ssd it finished without any errors neither data or checksum the i ran zpool trim on the pool then a few minutes later i saw checksum errors by the report then a few more i m not sure if the first error appeared before or after the trim i ran zfs trim again because the first time it finished quite fast despite there were much space to trim few more errors probably not related to running trim again because the errors has been constantly increasing since then then i was thinking maybe the first scrub before the trim somehow did not notice the errors so i ran zfs scrub again the errors grown to in the first few seconds so i stopped it at this point i couldn t use zfs send to save snapshot so i used rsync to backup the data it reported i o error for a few files it s an ssd which report blocks size so i used ashift later i learned that most ssds actually use larger blocks but report i don t know if it affected the trim the btrfs partition contains gib data kernels initramfs and boot related stuff i run a trim and a btrfs scrub on it no errors were found a smart extended self test was running when i first ran the zfs scrub and possibly when i run zfs trim the first time but it could have finished then partition scheme name fstype fsver label sda ├─ btrfs voidboot ├─ zfs member rpool └─ swap voidswap output of zpool status before submitting issue pool rpool state degraded status one or more devices has experienced an error resulting in data corruption applications may be affected action restore the file in question if possible otherwise restore the entire pool from backup see scan scrub canceled on sun feb config name state read write cksum rpool degraded degraded too many errors errors data errors use v for a list include any warning errors backtraces from the system logs no errors in dmesg
| 1
|
46,422
| 13,055,910,490
|
IssuesEvent
|
2020-07-30 03:05:40
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
Improved line fit resources directory cleanup (Trac #1181)
|
Incomplete Migration Migrated from Trac combo reconstruction defect
|
Migrated from https://code.icecube.wisc.edu/ticket/1181
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "There are no links to improved linefit documentation in the meta-project documentation. there is a decent index.rst, but it is in resources/ not resources/docs/ it needs the name the maintainer and have links to the CHANGELOG and the doxygen documentation.\n\nIn addition the example script example.py needs to be updated, it tries to load something called libNFE. and moved to resources/examples directory",
"reporter": "kjmeagher",
"cc": "",
"resolution": "wontfix",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "Improved line fit resources directory cleanup",
"priority": "blocker",
"keywords": "",
"time": "2015-08-19T11:40:01",
"milestone": "",
"owner": "gmaggi",
"type": "defect"
}
```
|
1.0
|
Improved line fit resources directory cleanup (Trac #1181) - Migrated from https://code.icecube.wisc.edu/ticket/1181
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "There are no links to improved linefit documentation in the meta-project documentation. there is a decent index.rst, but it is in resources/ not resources/docs/ it needs the name the maintainer and have links to the CHANGELOG and the doxygen documentation.\n\nIn addition the example script example.py needs to be updated, it tries to load something called libNFE. and moved to resources/examples directory",
"reporter": "kjmeagher",
"cc": "",
"resolution": "wontfix",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "Improved line fit resources directory cleanup",
"priority": "blocker",
"keywords": "",
"time": "2015-08-19T11:40:01",
"milestone": "",
"owner": "gmaggi",
"type": "defect"
}
```
|
defect
|
improved line fit resources directory cleanup trac migrated from json status closed changetime description there are no links to improved linefit documentation in the meta project documentation there is a decent index rst but it is in resources not resources docs it needs the name the maintainer and have links to the changelog and the doxygen documentation n nin addition the example script example py needs to be updated it tries to load something called libnfe and moved to resources examples directory reporter kjmeagher cc resolution wontfix ts component combo reconstruction summary improved line fit resources directory cleanup priority blocker keywords time milestone owner gmaggi type defect
| 1
|
513,072
| 14,915,550,648
|
IssuesEvent
|
2021-01-22 16:53:01
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
Hulu.com unable to get location for live tv.
|
OS/Desktop priority/P3 webcompat
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
Hulu.com (live TV) unable to get location when requested,
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Login to `Hulu.com` and livetv subscription, and open livetv option.
2. Hulu asks for me to share my location, and after I select “allow” it cannot load. Instead it chews on the request either indefinitely, or errors out with “unable to find your location - get help.”
## Actual result:
Unable to get location data for hulu.com livetv.
## Expected result:
Expect `hulu.com` to get location details for livetv option.
## Reproduces how often:
Need a US hulu account, with livetv enabled.
## Brave version (brave://version info)
`Version 1.15.72 Chromium: 86.0.4240.75 (Official Build) (64-bit)`
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? Yes
- Can you reproduce this issue with the beta channel? Unsure
- Can you reproduce this issue with the nightly channel? Unsure.
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
Reported here: https://community.brave.com/t/hulu-unable-to-get-location-and-load/165895/6
Have tested `hulu.com` which playback worked fine. Just unable to test the livetv option.
|
1.0
|
Hulu.com unable to get location for live tv. - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
Hulu.com (live TV) unable to get location when requested,
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Login to `Hulu.com` and livetv subscription, and open livetv option.
2. Hulu asks for me to share my location, and after I select “allow” it cannot load. Instead it chews on the request either indefinitely, or errors out with “unable to find your location - get help.”
## Actual result:
Unable to get location data for hulu.com livetv.
## Expected result:
Expect `hulu.com` to get location details for livetv option.
## Reproduces how often:
Need a US hulu account, with livetv enabled.
## Brave version (brave://version info)
`Version 1.15.72 Chromium: 86.0.4240.75 (Official Build) (64-bit)`
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? Yes
- Can you reproduce this issue with the beta channel? Unsure
- Can you reproduce this issue with the nightly channel? Unsure.
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
## Miscellaneous Information:
Reported here: https://community.brave.com/t/hulu-unable-to-get-location-and-load/165895/6
Have tested `hulu.com` which playback worked fine. Just unable to test the livetv option.
|
non_defect
|
hulu com unable to get location for live tv have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description hulu com live tv unable to get location when requested steps to reproduce login to hulu com and livetv subscription and open livetv option hulu asks for me to share my location and after i select “allow” it cannot load instead it chews on the request either indefinitely or errors out with “unable to find your location get help ” actual result unable to get location data for hulu com livetv expected result expect hulu com to get location details for livetv option reproduces how often need a us hulu account with livetv enabled brave version brave version info version chromium official build bit version channel information can you reproduce this issue with the current release yes can you reproduce this issue with the beta channel unsure can you reproduce this issue with the nightly channel unsure other additional information does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome miscellaneous information reported here have tested hulu com which playback worked fine just unable to test the livetv option
| 0
|
74,848
| 25,358,920,984
|
IssuesEvent
|
2022-11-20 17:15:23
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Video calls freeze after a few seconds at __strlen_avx2_rtm
|
T-Defect
|
### Steps to reproduce
1. Where are you starting? What can you see?
I ran element-desktop 1.11.14 (from NixOS unstable, but it crashes with 1.11.10 from the stable branch too).
(It worked reliably for months when I was using nixos-22.05, but fails since switching to nixos-unstable.)
On my tablet, I run element in the chrome with a test account. However, it also happens from a laptop running Ubuntu; the tablet is just for reproducing it locally.
2. What do you click?
On the tablet, I click to start a video call, and then accept it on element-desktop.
For testing, I turn off the mic, but it also happens without that.
### Outcome
#### What did you expect?
Video to continue working.
#### What happened instead?
After around 10 seconds, some of the video streams freeze.
e.g. the video in element-desktop freezes (but it is still sending to the tablet, which doesn't show anything wrong).
Turning off the video on the tablet and then turning it on again got it going again.
Sometimes it fails differently. e.g. both sides can see the other side's camera, but their local previews both freeze.
This happens whether using Wayland or X11 (via `NIXOS_OZONE_WL`).
The journal contains e.g.
```
[🡕] Process 513053 (.electron-wrapp) of user 1000 dumped core.
Module /nix/store/4nlgxhb09sdr51nc9hdm8az5b08vzkgx-glibc-2.35-163/lib/ld-linux-x86-64.so.2 with build-id db50353a26600bb848b9a5541b1506e0a24cb34b
Module linux-vdso.so.1 with build-id 97640497af8bdd9493208b8ce8243c9f775e9fc6
Module libspa-audioconvert.so without build-id.
Module libpipewire-module-session-manager.so without build-id.
Module libpipewire-module-metadata.so without build-id.
Module libpipewire-module-adapter.so without build-id.
Module libpipewire-module-client-device.so without build-id.
Module libpipewire-module-client-node.so without build-id.
Module libpipewire-module-protocol-native.so without build-id.
Module libpipewire-module-rt.so without build-id.
Module libspa-dbus.so without build-id.
Module libspa-journal.so without build-id.
Module libspa-support.so without build-id.
Module libpipewire-0.3.so.0 without build-id.
Module libasound_module_pcm_pipewire.so without build-id.
Module libstdc++.so.6 without build-id.
Module libicudata.so.72 without build-id.
Module libGLX.so.0 without build-id.
Module libGLdispatch.so.0 without build-id.
Module libdatrie.so.1 without build-id.
Module libsqlite3.so.0 with build-id 377f9d7f0fb8f5896be673d87eb739eb7866db92
Module libxml2.so.2 without build-id.
Module libjson-glib-1.0.so.0 without build-id.
Module libicui18n.so.72 without build-id.
Module libicuuc.so.72 without build-id.
Module libjpeg.so.62 without build-id.
Module libbz2.so.1 without build-id.
Module libgraphite2.so.3 without build-id.
Module libXinerama.so.1 without build-id.
Module libXcursor.so.1 without build-id.
Module libwayland-egl.so.1 with build-id d6b466ff99696870068c564a8ebded6a81cae225
Module libwayland-cursor.so.0 with build-id 5552be47a749c6825465c9a93d43f69f9701c9de
Module libwayland-client.so.0 with build-id c3616c06165ba2231464dfdde480b9ae48f92ec7
Module libcap.so.2 without build-id.
Module libgmp.so.10 without build-id.
Module libhogweed.so.6 without build-id.
Module libnettle.so.8 without build-id.
Module libtasn1.so.6 without build-id.
Module libunistring.so.2 without build-id.
Module libidn2.so.0 without build-id.
Module libp11-kit.so.0 without build-id.
Module libssp.so.0 without build-id.
Module libpcre.so.1 without build-id.
Module libblkid.so.1 with build-id f4f9ebcbcca3e44b7094b20e2ee71709825f36bb
Module libXdmcp.so.6 without build-id.
Module libXau.so.6 without build-id.
Module libwayland-server.so.0 with build-id 91f99799f5cca19715660617ace0a8444d822efe
Module libGL.so.1 without build-id.
Module libXrender.so.1 without build-id.
Module libxcb-render.so.0 without build-id.
Module libxcb-shm.so.0 without build-id.
Module libpng16.so.16 without build-id.
Module libEGL.so.1 without build-id.
Module libfreetype.so.6 without build-id.
Module libpixman-1.so.0 with build-id a4a8d46c6b2f698ceaf6661c79d06700493d31ab
Module libthai.so.0 without build-id.
Module libtracker-sparql-3.0.so.0 without build-id.
Module libXi.so.6 without build-id.
Module libepoxy.so.0 without build-id.
Module libgdk_pixbuf-2.0.so.0 with build-id 8f15f562c170a916afff71fb3e34401dd4e20d9e
Module libcairo-gobject.so.2 with build-id df83675e5c22873ce86626535519d51ff3cc513a
Module libfribidi.so.0 without build-id.
Module libfontconfig.so.1 without build-id.
Module libpangoft2-1.0.so.0 without build-id.
Module libharfbuzz.so.0 without build-id.
Module libpangocairo-1.0.so.0 without build-id.
Module libgdk-3.so.0 with build-id 2cfb8d0f3702bc5c753dbfac0002a0b9c18b5d83
Module libsystemd.so.0 without build-id.
Module libgnutls.so.30 without build-id.
Module libavahi-client.so.3 without build-id.
Module libavahi-common.so.3 without build-id.
Module librt.so.1 with build-id 7c9aae26f0646a27bf0f7c49c914b3258c5fa43e
Module libplc4.so without build-id.
Module libplds4.so without build-id.
Module libselinux.so.1 without build-id.
Module libmount.so.1 with build-id ed8fa2ae9881fc31bd8f5963397b42c8644f162d
Module libgmodule-2.0.so.0 with build-id 5ea22aa96ea6851566bb6ab070a1621683ae4e88
Module libpcre2-8.so.0 without build-id.
Module libffi.so.8 without build-id.
Module libcrypto.so.3 with build-id 5ba9c3862d2fed33339255247444fc34d53cb4cc
Module libz.so.1 without build-id.
Module libc.so.6 with build-id 2bb226bc600b443958c7566207d0d02f8345e6ea
Module libgcc_s.so.1 without build-id.
Module libatspi.so.0 without build-id.
Module libasound.so.2 without build-id.
Module libxkbcommon.so.0 without build-id.
Module libxcb.so.1 without build-id.
Module libexpat.so.1 without build-id.
Module libgbm.so.1 without build-id.
Module libXrandr.so.2 without build-id.
Module libXfixes.so.3 without build-id.
Module libXext.so.6 without build-id.
Module libXdamage.so.1 without build-id.
Module libXcomposite.so.1 without build-id.
Module libX11.so.6 without build-id.
Module libm.so.6 with build-id b8454b40db819599169f3a948939aed4b3fc7f82
Module libcairo.so.2 with build-id 21e308ba73f784934d4eb8cb2efd507151a8d65e
Module libpango-1.0.so.0 without build-id.
Module libgtk-3.so.0 with build-id 43ad91b494d9bf2e052ade1c14ed724c2c3030b2
Module libdrm.so.2 without build-id.
Module libdbus-1.so.3 without build-id.
Module libcups.so.2 without build-id.
Module libatk-bridge-2.0.so.0 without build-id.
Module libatk-1.0.so.0 without build-id.
Module libnspr4.so without build-id.
Module libsmime3.so without build-id.
Module libnssutil3.so without build-id.
Module libnss3.so without build-id.
Module libgio-2.0.so.0 with build-id 9c3d32e1d5dbf7d39ea67d2bb7045fbf51126a85
Module libglib-2.0.so.0 with build-id b13cd968ce6f5320e45dde1446f3066371403d7c
Module libgobject-2.0.so.0 with build-id cc0205109407a5b4ace0874f64aee611878a482d
Module libpthread.so.0 with build-id 85431f01160c3de171d3baeb3f8cf1c9578dc441
Module libdl.so.2 with build-id 67c430223def0be24c4ae1a4c3985f26566b8831
Module libffmpeg.so with build-id da7bfd439eb2866765067ecab210ebcb6184bb50
Module libsqlcipher.so without build-id.
Module .electron-wrapped with build-id be7e0a8182dc5bdd72ab8f92cc743fb0cf4ff95f
Stack trace of thread 513053:
#0 0x00007fd12aa534bd __strlen_avx2_rtm (libc.so.6 + 0x1664bd)
#1 0x000056046ff1ab86 n/a (.electron-wrapped + 0x2c2bb86)
#2 0x000056046ff19b11 n/a (.electron-wrapped + 0x2c2ab11)
#3 0x000056046ff11798 n/a (.electron-wrapped + 0x2c22798)
#4 0x000056046ff11722 n/a (.electron-wrapped + 0x2c22722)
#5 0x000056046ff130d1 n/a (.electron-wrapped + 0x2c240d1)
#6 0x000056046ff12664 n/a (.electron-wrapped + 0x2c23664)
#7 0x0000560471be23a3 n/a (.electron-wrapped + 0x48f33a3)
#8 0x0000560471bef4df n/a (.electron-wrapped + 0x49004df)
#9 0x0000560471bf10fd n/a (.electron-wrapped + 0x49020fd)
#10 0x0000560471bf6e18 n/a (.electron-wrapped + 0x4907e18)
#11 0x0000560471bf4fcd n/a (.electron-wrapped + 0x4905fcd)
#12 0x0000560471bf1bcf n/a (.electron-wrapped + 0x4902bcf)
#13 0x000056047234b616 n/a (.electron-wrapped + 0x505c616)
#14 0x0000560472368b85 n/a (.electron-wrapped + 0x5079b85)
#15 0x0000560472311d8e n/a (.electron-wrapped + 0x5022d8e)
#16 0x00005604723694e1 n/a (.electron-wrapped + 0x507a4e1)
#17 0x0000560472330fb2 n/a (.electron-wrapped + 0x5041fb2)
#18 0x0000560471cd79d1 n/a (.electron-wrapped + 0x49e89d1)
#19 0x000056046f71efa9 n/a (.electron-wrapped + 0x242ffa9)
#20 0x000056046f71fc4b n/a (.electron-wrapped + 0x2430c4b)
#21 0x000056046f71d18d n/a (.electron-wrapped + 0x242e18d)
#22 0x000056046f71d974 n/a (.electron-wrapped + 0x242e974)
#23 0x000056046f48491b n/a (.electron-wrapped + 0x219591b)
#24 0x00007fd12a91624e __libc_start_call_main (libc.so.6 + 0x2924e)
#25 0x00007fd12a916309 __libc_start_main@@GLIBC_2.34 (libc.so.6 + 0x29309)
#26 0x000056046f0fe02a _start (.electron-wrapped + 0x1e0f02a)
Stack trace of thread 513057:
#0 0x00007fd12a9727d5 __futex_abstimed_wait_common (libc.so.6 + 0x857d5)
#1 0x00007fd12a975524 pthread_cond_timedwait@@GLIBC_2.3.2 (libc.so.6 + 0x88524)
#2 0x00005604723a3206 n/a (.electron-wrapped + 0x50b4206)
#3 0x00005604723a3850 n/a (.electron-wrapped + 0x50b4850)
#4 0x000056047237cc98 n/a (.electron-wrapped + 0x508dc98)
#5 0x000056047237d526 n/a (.electron-wrapped + 0x508e526)
#6 0x000056047237d39d n/a (.electron-wrapped + 0x508e39d)
#7 0x000056047237d2b1 n/a (.electron-wrapped + 0x508e2b1)
#8 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#9 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#10 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513072:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513061:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513067:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513077:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513059:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513069:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513079:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513074:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513055:
#0 0x00007fd12a9727d5 __futex_abstimed_wait_common (libc.so.6 + 0x857d5)
#1 0x00007fd12a975524 pthread_cond_timedwait@@GLIBC_2.3.2 (libc.so.6 + 0x88524)
#2 0x00005604723a3206 n/a (.electron-wrapped + 0x50b4206)
#3 0x00005604723a3850 n/a (.electron-wrapped + 0x50b4850)
#4 0x000056047237cc98 n/a (.electron-wrapped + 0x508dc98)
#5 0x000056047237d752 n/a (.electron-wrapped + 0x508e752)
#6 0x000056047237d39d n/a (.electron-wrapped + 0x508e39d)
#7 0x000056047237d2b1 n/a (.electron-wrapped + 0x508e2b1)
#8 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#9 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#10 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513054:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00005604724bcf2a n/a (.electron-wrapped + 0x51cdf2a)
#2 0x00005604724baa6b n/a (.electron-wrapped + 0x51cba6b)
#3 0x00005604723b46d2 n/a (.electron-wrapped + 0x50c56d2)
#4 0x00005604723694e1 n/a (.electron-wrapped + 0x507a4e1)
#5 0x0000560472330fb2 n/a (.electron-wrapped + 0x5041fb2)
#6 0x0000560472383818 n/a (.electron-wrapped + 0x5094818)
#7 0x000056047237043d n/a (.electron-wrapped + 0x508143d)
#8 0x00005604723839a7 n/a (.electron-wrapped + 0x50949a7)
#9 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#10 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#11 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513064:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513066:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513056:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00005604724bcf2a n/a (.electron-wrapped + 0x51cdf2a)
#2 0x00005604724baa6b n/a (.electron-wrapped + 0x51cba6b)
#3 0x00005604723b4634 n/a (.electron-wrapped + 0x50c5634)
#4 0x00005604723694e1 n/a (.electron-wrapped + 0x507a4e1)
#5 0x0000560472330fb2 n/a (.electron-wrapped + 0x5041fb2)
#6 0x0000560472383818 n/a (.electron-wrapped + 0x5094818)
#7 0x0000560473b807df n/a (.electron-wrapped + 0x68917df)
#8 0x00005604723839a7 n/a (.electron-wrapped + 0x50949a7)
#9 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#10 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#11 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
ELF object binary architecture: AMD x86-64
```
### Operating system
nixos-unstable (nixpkgs rev 52b2ac8ae1)
### Application version
Element version: 1.11.14, Olm version: 3.2.12
### How did you install the app?
https://github.com/NixOS/nixpkgs/commit/52b2ac8ae18bbad4374ff0dd5aeee0fdf1aea739
### Homeserver
Unsure. I have a matrix_synapse-1.68.0.dist-info file.
### Will you send logs?
No
|
1.0
|
Video calls freeze after a few seconds at __strlen_avx2_rtm - ### Steps to reproduce
1. Where are you starting? What can you see?
I ran element-desktop 1.11.14 (from NixOS unstable, but it crashes with 1.11.10 from the stable branch too).
(It worked reliably for months when I was using nixos-22.05, but fails since switching to nixos-unstable.)
On my tablet, I run element in the chrome with a test account. However, it also happens from a laptop running Ubuntu; the tablet is just for reproducing it locally.
2. What do you click?
On the tablet, I click to start a video call, and then accept it on element-desktop.
For testing, I turn off the mic, but it also happens without that.
### Outcome
#### What did you expect?
Video to continue working.
#### What happened instead?
After around 10 seconds, some of the video streams freeze.
e.g. the video in element-desktop freezes (but it is still sending to the tablet, which doesn't show anything wrong).
Turning off the video on the tablet and then turning it on again got it going again.
Sometimes it fails differently. e.g. both sides can see the other side's camera, but their local previews both freeze.
This happens whether using Wayland or X11 (via `NIXOS_OZONE_WL`).
The journal contains e.g.
```
[🡕] Process 513053 (.electron-wrapp) of user 1000 dumped core.
Module /nix/store/4nlgxhb09sdr51nc9hdm8az5b08vzkgx-glibc-2.35-163/lib/ld-linux-x86-64.so.2 with build-id db50353a26600bb848b9a5541b1506e0a24cb34b
Module linux-vdso.so.1 with build-id 97640497af8bdd9493208b8ce8243c9f775e9fc6
Module libspa-audioconvert.so without build-id.
Module libpipewire-module-session-manager.so without build-id.
Module libpipewire-module-metadata.so without build-id.
Module libpipewire-module-adapter.so without build-id.
Module libpipewire-module-client-device.so without build-id.
Module libpipewire-module-client-node.so without build-id.
Module libpipewire-module-protocol-native.so without build-id.
Module libpipewire-module-rt.so without build-id.
Module libspa-dbus.so without build-id.
Module libspa-journal.so without build-id.
Module libspa-support.so without build-id.
Module libpipewire-0.3.so.0 without build-id.
Module libasound_module_pcm_pipewire.so without build-id.
Module libstdc++.so.6 without build-id.
Module libicudata.so.72 without build-id.
Module libGLX.so.0 without build-id.
Module libGLdispatch.so.0 without build-id.
Module libdatrie.so.1 without build-id.
Module libsqlite3.so.0 with build-id 377f9d7f0fb8f5896be673d87eb739eb7866db92
Module libxml2.so.2 without build-id.
Module libjson-glib-1.0.so.0 without build-id.
Module libicui18n.so.72 without build-id.
Module libicuuc.so.72 without build-id.
Module libjpeg.so.62 without build-id.
Module libbz2.so.1 without build-id.
Module libgraphite2.so.3 without build-id.
Module libXinerama.so.1 without build-id.
Module libXcursor.so.1 without build-id.
Module libwayland-egl.so.1 with build-id d6b466ff99696870068c564a8ebded6a81cae225
Module libwayland-cursor.so.0 with build-id 5552be47a749c6825465c9a93d43f69f9701c9de
Module libwayland-client.so.0 with build-id c3616c06165ba2231464dfdde480b9ae48f92ec7
Module libcap.so.2 without build-id.
Module libgmp.so.10 without build-id.
Module libhogweed.so.6 without build-id.
Module libnettle.so.8 without build-id.
Module libtasn1.so.6 without build-id.
Module libunistring.so.2 without build-id.
Module libidn2.so.0 without build-id.
Module libp11-kit.so.0 without build-id.
Module libssp.so.0 without build-id.
Module libpcre.so.1 without build-id.
Module libblkid.so.1 with build-id f4f9ebcbcca3e44b7094b20e2ee71709825f36bb
Module libXdmcp.so.6 without build-id.
Module libXau.so.6 without build-id.
Module libwayland-server.so.0 with build-id 91f99799f5cca19715660617ace0a8444d822efe
Module libGL.so.1 without build-id.
Module libXrender.so.1 without build-id.
Module libxcb-render.so.0 without build-id.
Module libxcb-shm.so.0 without build-id.
Module libpng16.so.16 without build-id.
Module libEGL.so.1 without build-id.
Module libfreetype.so.6 without build-id.
Module libpixman-1.so.0 with build-id a4a8d46c6b2f698ceaf6661c79d06700493d31ab
Module libthai.so.0 without build-id.
Module libtracker-sparql-3.0.so.0 without build-id.
Module libXi.so.6 without build-id.
Module libepoxy.so.0 without build-id.
Module libgdk_pixbuf-2.0.so.0 with build-id 8f15f562c170a916afff71fb3e34401dd4e20d9e
Module libcairo-gobject.so.2 with build-id df83675e5c22873ce86626535519d51ff3cc513a
Module libfribidi.so.0 without build-id.
Module libfontconfig.so.1 without build-id.
Module libpangoft2-1.0.so.0 without build-id.
Module libharfbuzz.so.0 without build-id.
Module libpangocairo-1.0.so.0 without build-id.
Module libgdk-3.so.0 with build-id 2cfb8d0f3702bc5c753dbfac0002a0b9c18b5d83
Module libsystemd.so.0 without build-id.
Module libgnutls.so.30 without build-id.
Module libavahi-client.so.3 without build-id.
Module libavahi-common.so.3 without build-id.
Module librt.so.1 with build-id 7c9aae26f0646a27bf0f7c49c914b3258c5fa43e
Module libplc4.so without build-id.
Module libplds4.so without build-id.
Module libselinux.so.1 without build-id.
Module libmount.so.1 with build-id ed8fa2ae9881fc31bd8f5963397b42c8644f162d
Module libgmodule-2.0.so.0 with build-id 5ea22aa96ea6851566bb6ab070a1621683ae4e88
Module libpcre2-8.so.0 without build-id.
Module libffi.so.8 without build-id.
Module libcrypto.so.3 with build-id 5ba9c3862d2fed33339255247444fc34d53cb4cc
Module libz.so.1 without build-id.
Module libc.so.6 with build-id 2bb226bc600b443958c7566207d0d02f8345e6ea
Module libgcc_s.so.1 without build-id.
Module libatspi.so.0 without build-id.
Module libasound.so.2 without build-id.
Module libxkbcommon.so.0 without build-id.
Module libxcb.so.1 without build-id.
Module libexpat.so.1 without build-id.
Module libgbm.so.1 without build-id.
Module libXrandr.so.2 without build-id.
Module libXfixes.so.3 without build-id.
Module libXext.so.6 without build-id.
Module libXdamage.so.1 without build-id.
Module libXcomposite.so.1 without build-id.
Module libX11.so.6 without build-id.
Module libm.so.6 with build-id b8454b40db819599169f3a948939aed4b3fc7f82
Module libcairo.so.2 with build-id 21e308ba73f784934d4eb8cb2efd507151a8d65e
Module libpango-1.0.so.0 without build-id.
Module libgtk-3.so.0 with build-id 43ad91b494d9bf2e052ade1c14ed724c2c3030b2
Module libdrm.so.2 without build-id.
Module libdbus-1.so.3 without build-id.
Module libcups.so.2 without build-id.
Module libatk-bridge-2.0.so.0 without build-id.
Module libatk-1.0.so.0 without build-id.
Module libnspr4.so without build-id.
Module libsmime3.so without build-id.
Module libnssutil3.so without build-id.
Module libnss3.so without build-id.
Module libgio-2.0.so.0 with build-id 9c3d32e1d5dbf7d39ea67d2bb7045fbf51126a85
Module libglib-2.0.so.0 with build-id b13cd968ce6f5320e45dde1446f3066371403d7c
Module libgobject-2.0.so.0 with build-id cc0205109407a5b4ace0874f64aee611878a482d
Module libpthread.so.0 with build-id 85431f01160c3de171d3baeb3f8cf1c9578dc441
Module libdl.so.2 with build-id 67c430223def0be24c4ae1a4c3985f26566b8831
Module libffmpeg.so with build-id da7bfd439eb2866765067ecab210ebcb6184bb50
Module libsqlcipher.so without build-id.
Module .electron-wrapped with build-id be7e0a8182dc5bdd72ab8f92cc743fb0cf4ff95f
Stack trace of thread 513053:
#0 0x00007fd12aa534bd __strlen_avx2_rtm (libc.so.6 + 0x1664bd)
#1 0x000056046ff1ab86 n/a (.electron-wrapped + 0x2c2bb86)
#2 0x000056046ff19b11 n/a (.electron-wrapped + 0x2c2ab11)
#3 0x000056046ff11798 n/a (.electron-wrapped + 0x2c22798)
#4 0x000056046ff11722 n/a (.electron-wrapped + 0x2c22722)
#5 0x000056046ff130d1 n/a (.electron-wrapped + 0x2c240d1)
#6 0x000056046ff12664 n/a (.electron-wrapped + 0x2c23664)
#7 0x0000560471be23a3 n/a (.electron-wrapped + 0x48f33a3)
#8 0x0000560471bef4df n/a (.electron-wrapped + 0x49004df)
#9 0x0000560471bf10fd n/a (.electron-wrapped + 0x49020fd)
#10 0x0000560471bf6e18 n/a (.electron-wrapped + 0x4907e18)
#11 0x0000560471bf4fcd n/a (.electron-wrapped + 0x4905fcd)
#12 0x0000560471bf1bcf n/a (.electron-wrapped + 0x4902bcf)
#13 0x000056047234b616 n/a (.electron-wrapped + 0x505c616)
#14 0x0000560472368b85 n/a (.electron-wrapped + 0x5079b85)
#15 0x0000560472311d8e n/a (.electron-wrapped + 0x5022d8e)
#16 0x00005604723694e1 n/a (.electron-wrapped + 0x507a4e1)
#17 0x0000560472330fb2 n/a (.electron-wrapped + 0x5041fb2)
#18 0x0000560471cd79d1 n/a (.electron-wrapped + 0x49e89d1)
#19 0x000056046f71efa9 n/a (.electron-wrapped + 0x242ffa9)
#20 0x000056046f71fc4b n/a (.electron-wrapped + 0x2430c4b)
#21 0x000056046f71d18d n/a (.electron-wrapped + 0x242e18d)
#22 0x000056046f71d974 n/a (.electron-wrapped + 0x242e974)
#23 0x000056046f48491b n/a (.electron-wrapped + 0x219591b)
#24 0x00007fd12a91624e __libc_start_call_main (libc.so.6 + 0x2924e)
#25 0x00007fd12a916309 __libc_start_main@@GLIBC_2.34 (libc.so.6 + 0x29309)
#26 0x000056046f0fe02a _start (.electron-wrapped + 0x1e0f02a)
Stack trace of thread 513057:
#0 0x00007fd12a9727d5 __futex_abstimed_wait_common (libc.so.6 + 0x857d5)
#1 0x00007fd12a975524 pthread_cond_timedwait@@GLIBC_2.3.2 (libc.so.6 + 0x88524)
#2 0x00005604723a3206 n/a (.electron-wrapped + 0x50b4206)
#3 0x00005604723a3850 n/a (.electron-wrapped + 0x50b4850)
#4 0x000056047237cc98 n/a (.electron-wrapped + 0x508dc98)
#5 0x000056047237d526 n/a (.electron-wrapped + 0x508e526)
#6 0x000056047237d39d n/a (.electron-wrapped + 0x508e39d)
#7 0x000056047237d2b1 n/a (.electron-wrapped + 0x508e2b1)
#8 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#9 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#10 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513072:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513061:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513067:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513077:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513059:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513069:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513079:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513074:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513055:
#0 0x00007fd12a9727d5 __futex_abstimed_wait_common (libc.so.6 + 0x857d5)
#1 0x00007fd12a975524 pthread_cond_timedwait@@GLIBC_2.3.2 (libc.so.6 + 0x88524)
#2 0x00005604723a3206 n/a (.electron-wrapped + 0x50b4206)
#3 0x00005604723a3850 n/a (.electron-wrapped + 0x50b4850)
#4 0x000056047237cc98 n/a (.electron-wrapped + 0x508dc98)
#5 0x000056047237d752 n/a (.electron-wrapped + 0x508e752)
#6 0x000056047237d39d n/a (.electron-wrapped + 0x508e39d)
#7 0x000056047237d2b1 n/a (.electron-wrapped + 0x508e2b1)
#8 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#9 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#10 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513054:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00005604724bcf2a n/a (.electron-wrapped + 0x51cdf2a)
#2 0x00005604724baa6b n/a (.electron-wrapped + 0x51cba6b)
#3 0x00005604723b46d2 n/a (.electron-wrapped + 0x50c56d2)
#4 0x00005604723694e1 n/a (.electron-wrapped + 0x507a4e1)
#5 0x0000560472330fb2 n/a (.electron-wrapped + 0x5041fb2)
#6 0x0000560472383818 n/a (.electron-wrapped + 0x5094818)
#7 0x000056047237043d n/a (.electron-wrapped + 0x508143d)
#8 0x00005604723839a7 n/a (.electron-wrapped + 0x50949a7)
#9 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#10 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#11 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513064:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116100df4 do_loop (libpipewire-0.3.so.0 + 0x46df4)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513066:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00007fd1160b0810 impl_pollfd_wait (libspa-support.so + 0x15810)
#2 0x00007fd1160a3cbb loop_iterate (libspa-support.so + 0x8cbb)
#3 0x00007fd116155822 do_loop (libpipewire-0.3.so.0 + 0x9b822)
#4 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#5 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
Stack trace of thread 513056:
#0 0x00007fd12a9fc237 epoll_wait (libc.so.6 + 0x10f237)
#1 0x00005604724bcf2a n/a (.electron-wrapped + 0x51cdf2a)
#2 0x00005604724baa6b n/a (.electron-wrapped + 0x51cba6b)
#3 0x00005604723b4634 n/a (.electron-wrapped + 0x50c5634)
#4 0x00005604723694e1 n/a (.electron-wrapped + 0x507a4e1)
#5 0x0000560472330fb2 n/a (.electron-wrapped + 0x5041fb2)
#6 0x0000560472383818 n/a (.electron-wrapped + 0x5094818)
#7 0x0000560473b807df n/a (.electron-wrapped + 0x68917df)
#8 0x00005604723839a7 n/a (.electron-wrapped + 0x50949a7)
#9 0x00005604723a70bf n/a (.electron-wrapped + 0x50b80bf)
#10 0x00007fd12a975e86 start_thread (libc.so.6 + 0x88e86)
#11 0x00007fd12a9fcc60 __clone3 (libc.so.6 + 0x10fc60)
ELF object binary architecture: AMD x86-64
```
### Operating system
nixos-unstable (nixpkgs rev 52b2ac8ae1)
### Application version
Element version: 1.11.14, Olm version: 3.2.12
### How did you install the app?
https://github.com/NixOS/nixpkgs/commit/52b2ac8ae18bbad4374ff0dd5aeee0fdf1aea739
### Homeserver
Unsure. I have a matrix_synapse-1.68.0.dist-info file.
### Will you send logs?
No
|
defect
|
video calls freeze after a few seconds at strlen rtm steps to reproduce where are you starting what can you see i ran element desktop from nixos unstable but it crashes with from the stable branch too it worked reliably for months when i was using nixos but fails since switching to nixos unstable on my tablet i run element in the chrome with a test account however it also happens from a laptop running ubuntu the tablet is just for reproducing it locally what do you click on the tablet i click to start a video call and then accept it on element desktop for testing i turn off the mic but it also happens without that outcome what did you expect video to continue working what happened instead after around seconds some of the video streams freeze e g the video in element desktop freezes but it is still sending to the tablet which doesn t show anything wrong turning off the video on the tablet and then turning it on again got it going again sometimes it fails differently e g both sides can see the other side s camera but their local previews both freeze this happens whether using wayland or via nixos ozone wl the journal contains e g process electron wrapp of user dumped core module nix store glibc lib ld linux so with build id module linux vdso so with build id module libspa audioconvert so without build id module libpipewire module session manager so without build id module libpipewire module metadata so without build id module libpipewire module adapter so without build id module libpipewire module client device so without build id module libpipewire module client node so without build id module libpipewire module protocol native so without build id module libpipewire module rt so without build id module libspa dbus so without build id module libspa journal so without build id module libspa support so without build id module libpipewire so without build id module libasound module pcm pipewire so without build id module libstdc so without build id module libicudata so without build id module libglx so without build id module libgldispatch so without build id module libdatrie so without build id module so with build id module so without build id module libjson glib so without build id module so without build id module libicuuc so without build id module libjpeg so without build id module so without build id module so without build id module libxinerama so without build id module libxcursor so without build id module libwayland egl so with build id module libwayland cursor so with build id module libwayland client so with build id module libcap so without build id module libgmp so without build id module libhogweed so without build id module libnettle so without build id module so without build id module libunistring so without build id module so without build id module kit so without build id module libssp so without build id module libpcre so without build id module libblkid so with build id module libxdmcp so without build id module libxau so without build id module libwayland server so with build id module libgl so without build id module libxrender so without build id module libxcb render so without build id module libxcb shm so without build id module so without build id module libegl so without build id module libfreetype so without build id module libpixman so with build id module libthai so without build id module libtracker sparql so without build id module libxi so without build id module libepoxy so without build id module libgdk pixbuf so with build id module libcairo gobject so with build id module libfribidi so without build id module libfontconfig so without build id module so without build id module libharfbuzz so without build id module libpangocairo so without build id module libgdk so with build id module libsystemd so without build id module libgnutls so without build id module libavahi client so without build id module libavahi common so without build id module librt so with build id module so without build id module so without build id module libselinux so without build id module libmount so with build id module libgmodule so with build id module so without build id module libffi so without build id module libcrypto so with build id module libz so without build id module libc so with build id module libgcc s so without build id module libatspi so without build id module libasound so without build id module libxkbcommon so without build id module libxcb so without build id module libexpat so without build id module libgbm so without build id module libxrandr so without build id module libxfixes so without build id module libxext so without build id module libxdamage so without build id module libxcomposite so without build id module so without build id module libm so with build id module libcairo so with build id module libpango so without build id module libgtk so with build id module libdrm so without build id module libdbus so without build id module libcups so without build id module libatk bridge so without build id module libatk so without build id module so without build id module so without build id module so without build id module so without build id module libgio so with build id module libglib so with build id module libgobject so with build id module libpthread so with build id module libdl so with build id module libffmpeg so with build id module libsqlcipher so without build id module electron wrapped with build id stack trace of thread strlen rtm libc so n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped libc start call main libc so libc start main glibc libc so start electron wrapped stack trace of thread futex abstimed wait common libc so pthread cond timedwait glibc libc so n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread futex abstimed wait common libc so pthread cond timedwait glibc libc so n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped start thread libc so libc so stack trace of thread epoll wait libc so n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so impl pollfd wait libspa support so loop iterate libspa support so do loop libpipewire so start thread libc so libc so stack trace of thread epoll wait libc so n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped n a electron wrapped start thread libc so libc so elf object binary architecture amd operating system nixos unstable nixpkgs rev application version element version olm version how did you install the app homeserver unsure i have a matrix synapse dist info file will you send logs no
| 1
|
51,671
| 13,211,279,425
|
IssuesEvent
|
2020-08-15 22:00:46
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
[photoflash] Add back to the trunk (Trac #808)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/808">https://code.icecube.wisc.edu/projects/icecube/ticket/808</a>, reported by olivasand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"_ts": "1458335643235016",
"description": "This was removed at some point for reasons that aren't clear. Start including it again.",
"reporter": "olivas",
"cc": "",
"resolution": "wontfix",
"time": "2014-11-12T03:10:19",
"component": "combo simulation",
"summary": "[photoflash] Add back to the trunk",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[photoflash] Add back to the trunk (Trac #808) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/808">https://code.icecube.wisc.edu/projects/icecube/ticket/808</a>, reported by olivasand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"_ts": "1458335643235016",
"description": "This was removed at some point for reasons that aren't clear. Start including it again.",
"reporter": "olivas",
"cc": "",
"resolution": "wontfix",
"time": "2014-11-12T03:10:19",
"component": "combo simulation",
"summary": "[photoflash] Add back to the trunk",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
add back to the trunk trac migrated from json status closed changetime ts description this was removed at some point for reasons that aren t clear start including it again reporter olivas cc resolution wontfix time component combo simulation summary add back to the trunk priority normal keywords milestone owner olivas type defect
| 1
|
25,033
| 4,178,504,161
|
IssuesEvent
|
2016-06-22 07:08:08
|
netty/netty
|
https://api.github.com/repos/netty/netty
|
closed
|
io.netty.resolver.dns.DnsNameResolver does not resolve localhost on Windows
|
defect
|
On Windows localhost is not in hosts file and the DNS server does not resolve this address either, i.e it is handled by the Windows API. So using a `Bootstrap` (among others) with the resolver based on `DnsNameResolver` will not resolve localhost.
http://serverfault.com/questions/4689/windows-7-localhost-name-resolution-is-handled-within-dns-itself-why
|
1.0
|
io.netty.resolver.dns.DnsNameResolver does not resolve localhost on Windows - On Windows localhost is not in hosts file and the DNS server does not resolve this address either, i.e it is handled by the Windows API. So using a `Bootstrap` (among others) with the resolver based on `DnsNameResolver` will not resolve localhost.
http://serverfault.com/questions/4689/windows-7-localhost-name-resolution-is-handled-within-dns-itself-why
|
defect
|
io netty resolver dns dnsnameresolver does not resolve localhost on windows on windows localhost is not in hosts file and the dns server does not resolve this address either i e it is handled by the windows api so using a bootstrap among others with the resolver based on dnsnameresolver will not resolve localhost
| 1
|
97,286
| 12,224,717,592
|
IssuesEvent
|
2020-05-03 00:20:45
|
CERT-Polska/mquery
|
https://api.github.com/repos/CERT-Polska/mquery
|
closed
|
Support case insenstive strings in yara rules
|
level:hard priority:medium status:needs more design zone:flask backend zone:ursadb
|
Right now, we just ignore strings with the nocase flag:
```
rule CaseInsensitiveTextExample
{
strings:
$text_string = "foobar" nocase
condition:
$text_string
}
```
Supporting them correctly is... harder than it looks like. This can match:
```
foobar
foobaR
foobAr
foobAR
fooBar
... and 59 strings more
```
and ursadb query language is not expressive enough to support this.
We can't hack around this by chopping the query in the backend to something like:
```
( "foo" AND (
"oob" AND (
"oba" AND (
...
) OR
"obA" AND (
)
) OR
"ooB" AND (
"oBa" AND (
...
) OR
"oBA" AND (
...
)
)
) OR "foO AND (
...
) OR "fOo" AND (
...
) OR "fOO" AND (
...
) ...
```
Because of exponential growth.
OTOH I feel like like this can solved with a C++ method (needs investigation). In this case we need to introduce `nocase` strings to ursadb.
Needs investigation (if this results in too many false positives, we may as well give up).
|
1.0
|
Support case insenstive strings in yara rules - Right now, we just ignore strings with the nocase flag:
```
rule CaseInsensitiveTextExample
{
strings:
$text_string = "foobar" nocase
condition:
$text_string
}
```
Supporting them correctly is... harder than it looks like. This can match:
```
foobar
foobaR
foobAr
foobAR
fooBar
... and 59 strings more
```
and ursadb query language is not expressive enough to support this.
We can't hack around this by chopping the query in the backend to something like:
```
( "foo" AND (
"oob" AND (
"oba" AND (
...
) OR
"obA" AND (
)
) OR
"ooB" AND (
"oBa" AND (
...
) OR
"oBA" AND (
...
)
)
) OR "foO AND (
...
) OR "fOo" AND (
...
) OR "fOO" AND (
...
) ...
```
Because of exponential growth.
OTOH I feel like like this can solved with a C++ method (needs investigation). In this case we need to introduce `nocase` strings to ursadb.
Needs investigation (if this results in too many false positives, we may as well give up).
|
non_defect
|
support case insenstive strings in yara rules right now we just ignore strings with the nocase flag rule caseinsensitivetextexample strings text string foobar nocase condition text string supporting them correctly is harder than it looks like this can match foobar foobar foobar foobar foobar and strings more and ursadb query language is not expressive enough to support this we can t hack around this by chopping the query in the backend to something like foo and oob and oba and or oba and or oob and oba and or oba and or foo and or foo and or foo and because of exponential growth otoh i feel like like this can solved with a c method needs investigation in this case we need to introduce nocase strings to ursadb needs investigation if this results in too many false positives we may as well give up
| 0
|
732,080
| 25,244,021,778
|
IssuesEvent
|
2022-11-15 09:46:13
|
marcusolsson/obsidian-projects
|
https://api.github.com/repos/marcusolsson/obsidian-projects
|
closed
|
Duplicate a view
|
kind/feature triage/accepted priority/backlog area/core
|
### What would you like to be added?
The ability to duplicate a view and keep all the configuration in the newly created view.
### Why is this needed?
It would make it much easier to add new views if you want to change only something minor, and don't want to configure the view all over again. This will be especially helpful when filtering will be available.
|
1.0
|
Duplicate a view - ### What would you like to be added?
The ability to duplicate a view and keep all the configuration in the newly created view.
### Why is this needed?
It would make it much easier to add new views if you want to change only something minor, and don't want to configure the view all over again. This will be especially helpful when filtering will be available.
|
non_defect
|
duplicate a view what would you like to be added the ability to duplicate a view and keep all the configuration in the newly created view why is this needed it would make it much easier to add new views if you want to change only something minor and don t want to configure the view all over again this will be especially helpful when filtering will be available
| 0
|
11,050
| 2,622,956,292
|
IssuesEvent
|
2015-03-04 09:03:40
|
folded/carve
|
https://api.github.com/repos/folded/carve
|
opened
|
Another case with error message 'UNEXPECTED face loop with size!=1'
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. load obj1 and obj2
2. run obj = csg.compute(obj1,obj2,csg.INTERSECTION);
3. get the error message and the output obj is not correct
What is the expected output? What do you see instead?
get a correct intersection.
What version of the product are you using? On what operating system?
latest
Please provide any additional information below.
Please see attachements obj1 and obj2.
```
Original issue reported on code.google.com by `tanglaoy...@gmail.com` on 1 Aug 2014 at 3:32
Attachments:
* [obj1.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj1.ply)
* [obj1_bin.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj1_bin.ply)
* [obj2.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj2.ply)
* [obj2_bin.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj2_bin.ply)
|
1.0
|
Another case with error message 'UNEXPECTED face loop with size!=1' - ```
What steps will reproduce the problem?
1. load obj1 and obj2
2. run obj = csg.compute(obj1,obj2,csg.INTERSECTION);
3. get the error message and the output obj is not correct
What is the expected output? What do you see instead?
get a correct intersection.
What version of the product are you using? On what operating system?
latest
Please provide any additional information below.
Please see attachements obj1 and obj2.
```
Original issue reported on code.google.com by `tanglaoy...@gmail.com` on 1 Aug 2014 at 3:32
Attachments:
* [obj1.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj1.ply)
* [obj1_bin.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj1_bin.ply)
* [obj2.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj2.ply)
* [obj2_bin.ply](https://storage.googleapis.com/google-code-attachments/carve/issue-67/comment-0/obj2_bin.ply)
|
defect
|
another case with error message unexpected face loop with size what steps will reproduce the problem load and run obj csg compute csg intersection get the error message and the output obj is not correct what is the expected output what do you see instead get a correct intersection what version of the product are you using on what operating system latest please provide any additional information below please see attachements and original issue reported on code google com by tanglaoy gmail com on aug at attachments
| 1
|
19,130
| 3,144,819,096
|
IssuesEvent
|
2015-09-14 15:08:29
|
ox-it/ords
|
https://api.github.com/repos/ox-it/ords
|
closed
|
Can't edit fields with the Time data type
|
auto-migrated Priority-Critical Type-Defect
|
```
What steps will reproduce the problem?
1. Open a table containing a field with the Time data type
2. Edit this field, and try to save the contents
3. Now try adding a new record to the same table, with content in the Time field
What is the expected output? What do you see instead?
In both cases, ORDS refuses to save the record, claiming that the data isn't
consistent with the schema.
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `meriel.p...@gmail.com` on 3 Aug 2015 at 1:00
|
1.0
|
Can't edit fields with the Time data type - ```
What steps will reproduce the problem?
1. Open a table containing a field with the Time data type
2. Edit this field, and try to save the contents
3. Now try adding a new record to the same table, with content in the Time field
What is the expected output? What do you see instead?
In both cases, ORDS refuses to save the record, claiming that the data isn't
consistent with the schema.
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `meriel.p...@gmail.com` on 3 Aug 2015 at 1:00
|
defect
|
can t edit fields with the time data type what steps will reproduce the problem open a table containing a field with the time data type edit this field and try to save the contents now try adding a new record to the same table with content in the time field what is the expected output what do you see instead in both cases ords refuses to save the record claiming that the data isn t consistent with the schema please use labels and text to provide additional information original issue reported on code google com by meriel p gmail com on aug at
| 1
|
9,344
| 2,615,145,288
|
IssuesEvent
|
2015-03-01 06:21:06
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Podcast player - eric - 90%, needs to be committed
|
auto-migrated Milestone-2 Priority-Medium Studio Type-Defect
|
```
mostly done, need to commit
```
Original issue reported on code.google.com by `v...@google.com` on 29 Jul 2010 at 4:30
* Merged into: #122
|
1.0
|
Podcast player - eric - 90%, needs to be committed - ```
mostly done, need to commit
```
Original issue reported on code.google.com by `v...@google.com` on 29 Jul 2010 at 4:30
* Merged into: #122
|
defect
|
podcast player eric needs to be committed mostly done need to commit original issue reported on code google com by v google com on jul at merged into
| 1
|
400,686
| 11,779,173,751
|
IssuesEvent
|
2020-03-16 17:31:52
|
wc-work/crumby
|
https://api.github.com/repos/wc-work/crumby
|
closed
|
Add Nat's footer changes
|
high priority
|
dynamic fixes branch doesn't share history with dev or master, and I can't automatically merge. I need to manually add.
|
1.0
|
Add Nat's footer changes - dynamic fixes branch doesn't share history with dev or master, and I can't automatically merge. I need to manually add.
|
non_defect
|
add nat s footer changes dynamic fixes branch doesn t share history with dev or master and i can t automatically merge i need to manually add
| 0
|
55,319
| 7,975,923,136
|
IssuesEvent
|
2018-07-17 10:54:39
|
pods-framework/pods-beaver-builder-themer-add-on
|
https://api.github.com/repos/pods-framework/pods-beaver-builder-themer-add-on
|
closed
|
Linking with multiple of a CPT and using their featured images in a slider
|
bug documentation
|
## Issue Overview
Trying to use featured image of multiple CPT posts to create a row background slideshow with beaver builder. Made the relationship and selected the custom posts to link and set up the row background slideshow. The issue is that the images aren't showing up.
## Expected Behavior
Featured images of CPT's should be showing as a slideshow in page header.
## Current Behavior
The images are not showing in my page header.
If I create a new image field in my CPT and link it with "rel_gallery_featured_images.gallery_slider_image" then it works fine.
## Possible Solution
## Steps to Reproduce (for bugs)
1. Relationship set up between page and multiple of the CPT "gallery"
2. Page header set up as "Slideshow" and linked to "gallery" posts with advanced > "rel_gallery_featured_images.post_thumbnail"
3. Images not showing
## Pods Package Export (helpful!)
{"meta":{"version":"2.7.6","build":1531321591},"pods":{"2044":{"id":2044,"name":"gallery","label":"Galleries","description":"","type":"post_type","storage":"meta","object":"","alias":"","fields":{"image_gallery":{"id":2045,"name":"image_gallery","label":"Image Gallery","description":"","help":"","class":"","type":"file","weight":0,"pick_object":"custom-simple","pick_val":"","sister_id":"","required":"1","file_format_type":"multi","file_uploader":"attachment","file_attachment_tab":"upload","file_edit_title":"1","file_show_edit_link":"0","file_linked":"0","file_limit":"0","file_restrict_filesize":"10MB","file_type":"images","file_field_template":"rows","file_add_button":"Add File","file_modal_title":"Attach a file","file_modal_add_button":"Add File","file_wp_gallery_output":"0","file_wp_gallery_link":"post","file_wp_gallery_columns":"1","file_wp_gallery_random_sort":"0","file_wp_gallery_size":"thumbnail","pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","unique":"0","file_allowed_extensions":""},"gallery_slider_image":{"id":2342,"name":"gallery_slider_image","label":"Gallery slider image","description":"","help":"","class":"","type":"file","weight":1,"pick_object":"custom-simple","pick_val":"","sister_id":"","required":"1","file_format_type":"single","file_uploader":"attachment","file_attachment_tab":"upload","file_edit_title":"1","file_show_edit_link":"0","file_linked":"0","file_limit":"0","file_restrict_filesize":"10MB","file_type":"images","file_field_template":"rows","file_add_button":"Add File","file_modal_title":"Attach a file","file_modal_add_button":"Add File","file_wp_gallery_output":"0","file_wp_gallery_link":"post","file_wp_gallery_columns":"1","file_wp_gallery_random_sort":"0","file_wp_gallery_size":"thumbnail","pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","unique":"0","file_allowed_extensions":""}},"show_in_menu":"1","label_singular":"Gallery","public":"1","show_ui":"1","supports_title":"1","supports_editor":"1","publicly_queryable":"1","exclude_from_search":"0","capability_type":"post","capability_type_custom":"gallery","capability_type_extra":"1","has_archive":"0","hierarchical":"0","rewrite":"1","rewrite_with_front":"1","rewrite_feeds":"0","rewrite_pages":"1","query_var":"1","can_export":"1","default_status":"draft","supports_author":"0","supports_thumbnail":"1","supports_excerpt":"0","supports_trackbacks":"0","supports_custom_fields":"0","supports_comments":"0","supports_revisions":"0","supports_page_attributes":"0","supports_post_formats":"0","built_in_taxonomies_attachment_category":"0","built_in_taxonomies_attachment_tag":"0","built_in_taxonomies_category":"0","built_in_taxonomies_fl-builder-template-category":"0","built_in_taxonomies_fl-builder-template-type":"0","built_in_taxonomies_link_category":"0","built_in_taxonomies_post_tag":"0","show_in_nav_menus":"1","show_in_admin_bar":"1","pfat_enable":"0","pfat_run_outside_loop":"0","pfat_append_single":"append","pfat_filter_single":"the_content","pfat_append_archive":"append","pfat_filter_archive":"the_content","rest_enable":"0","read_all":"0","write_all":"0","menu_icon":"dashicons-images-alt2","menu_position":"4","supports_jetpack_publicize":"0","supports_jetpack_markdown":"0","built_in_taxonomies_product_cat":"0","built_in_taxonomies_product_shipping_class":"0","built_in_taxonomies_product_tag":"0","built_in_taxonomies_product_type":"0","built_in_taxonomies_product_visibility":"0"},"2331":{"id":2331,"name":"page","label":"Page","description":"","type":"post_type","storage":"meta","object":"page","alias":"","fields":{"rel_gallery_featured_images":{"id":2332,"name":"rel_gallery_featured_images","label":"REL gallery Featured Images","description":"","help":"","class":"","type":"pick","weight":0,"pick_object":"post_type","pick_val":"gallery","sister_id":"","required":"0","pick_format_type":"multi","pick_format_single":"dropdown","pick_format_multi":"checkbox","pick_allow_add_new":"1","pick_taggable":"0","pick_show_icon":"1","pick_show_edit_link":"1","pick_show_view_link":"1","pick_limit":"0","pick_user_role":[],"pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","unique":"0","pick_select_text":"","pick_table_id":"","pick_table_index":"","pick_display":"","pick_where":"","pick_orderby":"","pick_groupby":""},"rel_category_featured_images_copy":{"id":2349,"name":"rel_category_featured_images_copy","label":"REL category Featured Images (Copy)","description":"","help":"","class":"","type":"pick","weight":1,"pick_object":"taxonomy","pick_val":"product_cat","sister_id":"","required":"0","unique":"0","pick_format_type":"multi","pick_format_single":"dropdown","pick_format_multi":"checkbox","pick_allow_add_new":"1","pick_taggable":"0","pick_show_icon":"1","pick_show_edit_link":"1","pick_show_view_link":"1","pick_limit":"0","pick_user_role":[],"pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","pick_select_text":"","pick_table_id":"","pick_table_index":"","pick_display":"","pick_where":"","pick_orderby":"","pick_groupby":""}},"show_in_menu":1,"pfat_enable":"0","pfat_run_outside_loop":"0","pfat_append_single":"append","pfat_filter_single":"the_content","pfat_append_archive":"append","pfat_filter_archive":"the_content","rest_enable":"0","read_all":"0","write_all":"0"}}}
```
## Related Issues and/or PRs
<!-- List related issues or PRs against other branches: -->
## Todos
- [ ] Tests
- [ ] Documentation
## WordPress Environment
<details>
```
Copy and paste your System Details from **Pods Admin > Settings > Debug Information** in WordPress admin here.
```
</details>
|
1.0
|
Linking with multiple of a CPT and using their featured images in a slider - ## Issue Overview
Trying to use featured image of multiple CPT posts to create a row background slideshow with beaver builder. Made the relationship and selected the custom posts to link and set up the row background slideshow. The issue is that the images aren't showing up.
## Expected Behavior
Featured images of CPT's should be showing as a slideshow in page header.
## Current Behavior
The images are not showing in my page header.
If I create a new image field in my CPT and link it with "rel_gallery_featured_images.gallery_slider_image" then it works fine.
## Possible Solution
## Steps to Reproduce (for bugs)
1. Relationship set up between page and multiple of the CPT "gallery"
2. Page header set up as "Slideshow" and linked to "gallery" posts with advanced > "rel_gallery_featured_images.post_thumbnail"
3. Images not showing
## Pods Package Export (helpful!)
{"meta":{"version":"2.7.6","build":1531321591},"pods":{"2044":{"id":2044,"name":"gallery","label":"Galleries","description":"","type":"post_type","storage":"meta","object":"","alias":"","fields":{"image_gallery":{"id":2045,"name":"image_gallery","label":"Image Gallery","description":"","help":"","class":"","type":"file","weight":0,"pick_object":"custom-simple","pick_val":"","sister_id":"","required":"1","file_format_type":"multi","file_uploader":"attachment","file_attachment_tab":"upload","file_edit_title":"1","file_show_edit_link":"0","file_linked":"0","file_limit":"0","file_restrict_filesize":"10MB","file_type":"images","file_field_template":"rows","file_add_button":"Add File","file_modal_title":"Attach a file","file_modal_add_button":"Add File","file_wp_gallery_output":"0","file_wp_gallery_link":"post","file_wp_gallery_columns":"1","file_wp_gallery_random_sort":"0","file_wp_gallery_size":"thumbnail","pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","unique":"0","file_allowed_extensions":""},"gallery_slider_image":{"id":2342,"name":"gallery_slider_image","label":"Gallery slider image","description":"","help":"","class":"","type":"file","weight":1,"pick_object":"custom-simple","pick_val":"","sister_id":"","required":"1","file_format_type":"single","file_uploader":"attachment","file_attachment_tab":"upload","file_edit_title":"1","file_show_edit_link":"0","file_linked":"0","file_limit":"0","file_restrict_filesize":"10MB","file_type":"images","file_field_template":"rows","file_add_button":"Add File","file_modal_title":"Attach a file","file_modal_add_button":"Add File","file_wp_gallery_output":"0","file_wp_gallery_link":"post","file_wp_gallery_columns":"1","file_wp_gallery_random_sort":"0","file_wp_gallery_size":"thumbnail","pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","unique":"0","file_allowed_extensions":""}},"show_in_menu":"1","label_singular":"Gallery","public":"1","show_ui":"1","supports_title":"1","supports_editor":"1","publicly_queryable":"1","exclude_from_search":"0","capability_type":"post","capability_type_custom":"gallery","capability_type_extra":"1","has_archive":"0","hierarchical":"0","rewrite":"1","rewrite_with_front":"1","rewrite_feeds":"0","rewrite_pages":"1","query_var":"1","can_export":"1","default_status":"draft","supports_author":"0","supports_thumbnail":"1","supports_excerpt":"0","supports_trackbacks":"0","supports_custom_fields":"0","supports_comments":"0","supports_revisions":"0","supports_page_attributes":"0","supports_post_formats":"0","built_in_taxonomies_attachment_category":"0","built_in_taxonomies_attachment_tag":"0","built_in_taxonomies_category":"0","built_in_taxonomies_fl-builder-template-category":"0","built_in_taxonomies_fl-builder-template-type":"0","built_in_taxonomies_link_category":"0","built_in_taxonomies_post_tag":"0","show_in_nav_menus":"1","show_in_admin_bar":"1","pfat_enable":"0","pfat_run_outside_loop":"0","pfat_append_single":"append","pfat_filter_single":"the_content","pfat_append_archive":"append","pfat_filter_archive":"the_content","rest_enable":"0","read_all":"0","write_all":"0","menu_icon":"dashicons-images-alt2","menu_position":"4","supports_jetpack_publicize":"0","supports_jetpack_markdown":"0","built_in_taxonomies_product_cat":"0","built_in_taxonomies_product_shipping_class":"0","built_in_taxonomies_product_tag":"0","built_in_taxonomies_product_type":"0","built_in_taxonomies_product_visibility":"0"},"2331":{"id":2331,"name":"page","label":"Page","description":"","type":"post_type","storage":"meta","object":"page","alias":"","fields":{"rel_gallery_featured_images":{"id":2332,"name":"rel_gallery_featured_images","label":"REL gallery Featured Images","description":"","help":"","class":"","type":"pick","weight":0,"pick_object":"post_type","pick_val":"gallery","sister_id":"","required":"0","pick_format_type":"multi","pick_format_single":"dropdown","pick_format_multi":"checkbox","pick_allow_add_new":"1","pick_taggable":"0","pick_show_icon":"1","pick_show_edit_link":"1","pick_show_view_link":"1","pick_limit":"0","pick_user_role":[],"pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","unique":"0","pick_select_text":"","pick_table_id":"","pick_table_index":"","pick_display":"","pick_where":"","pick_orderby":"","pick_groupby":""},"rel_category_featured_images_copy":{"id":2349,"name":"rel_category_featured_images_copy","label":"REL category Featured Images (Copy)","description":"","help":"","class":"","type":"pick","weight":1,"pick_object":"taxonomy","pick_val":"product_cat","sister_id":"","required":"0","unique":"0","pick_format_type":"multi","pick_format_single":"dropdown","pick_format_multi":"checkbox","pick_allow_add_new":"1","pick_taggable":"0","pick_show_icon":"1","pick_show_edit_link":"1","pick_show_view_link":"1","pick_limit":"0","pick_user_role":[],"pick_post_status":["publish"],"admin_only":"0","restrict_role":"0","restrict_capability":"0","hidden":"0","read_only":"0","roles_allowed":["administrator"],"rest_read":"0","rest_write":"0","rest_pick_response":"array","rest_pick_depth":"2","pick_select_text":"","pick_table_id":"","pick_table_index":"","pick_display":"","pick_where":"","pick_orderby":"","pick_groupby":""}},"show_in_menu":1,"pfat_enable":"0","pfat_run_outside_loop":"0","pfat_append_single":"append","pfat_filter_single":"the_content","pfat_append_archive":"append","pfat_filter_archive":"the_content","rest_enable":"0","read_all":"0","write_all":"0"}}}
```
## Related Issues and/or PRs
<!-- List related issues or PRs against other branches: -->
## Todos
- [ ] Tests
- [ ] Documentation
## WordPress Environment
<details>
```
Copy and paste your System Details from **Pods Admin > Settings > Debug Information** in WordPress admin here.
```
</details>
|
non_defect
|
linking with multiple of a cpt and using their featured images in a slider issue overview trying to use featured image of multiple cpt posts to create a row background slideshow with beaver builder made the relationship and selected the custom posts to link and set up the row background slideshow the issue is that the images aren t showing up expected behavior featured images of cpt s should be showing as a slideshow in page header current behavior the images are not showing in my page header if i create a new image field in my cpt and link it with rel gallery featured images gallery slider image then it works fine possible solution steps to reproduce for bugs relationship set up between page and multiple of the cpt gallery page header set up as slideshow and linked to gallery posts with advanced rel gallery featured images post thumbnail images not showing pods package export helpful meta version build pods id name gallery label galleries description type post type storage meta object alias fields image gallery id name image gallery label image gallery description help class type file weight pick object custom simple pick val sister id required file format type multi file uploader attachment file attachment tab upload file edit title file show edit link file linked file limit file restrict filesize file type images file field template rows file add button add file file modal title attach a file file modal add button add file file wp gallery output file wp gallery link post file wp gallery columns file wp gallery random sort file wp gallery size thumbnail pick post status admin only restrict role restrict capability hidden read only roles allowed rest read rest write rest pick response array rest pick depth unique file allowed extensions gallery slider image id name gallery slider image label gallery slider image description help class type file weight pick object custom simple pick val sister id required file format type single file uploader attachment file attachment tab upload file edit title file show edit link file linked file limit file restrict filesize file type images file field template rows file add button add file file modal title attach a file file modal add button add file file wp gallery output file wp gallery link post file wp gallery columns file wp gallery random sort file wp gallery size thumbnail pick post status admin only restrict role restrict capability hidden read only roles allowed rest read rest write rest pick response array rest pick depth unique file allowed extensions show in menu label singular gallery public show ui supports title supports editor publicly queryable exclude from search capability type post capability type custom gallery capability type extra has archive hierarchical rewrite rewrite with front rewrite feeds rewrite pages query var can export default status draft supports author supports thumbnail supports excerpt supports trackbacks supports custom fields supports comments supports revisions supports page attributes supports post formats built in taxonomies attachment category built in taxonomies attachment tag built in taxonomies category built in taxonomies fl builder template category built in taxonomies fl builder template type built in taxonomies link category built in taxonomies post tag show in nav menus show in admin bar pfat enable pfat run outside loop pfat append single append pfat filter single the content pfat append archive append pfat filter archive the content rest enable read all write all menu icon dashicons images menu position supports jetpack publicize supports jetpack markdown built in taxonomies product cat built in taxonomies product shipping class built in taxonomies product tag built in taxonomies product type built in taxonomies product visibility id name page label page description type post type storage meta object page alias fields rel gallery featured images id name rel gallery featured images label rel gallery featured images description help class type pick weight pick object post type pick val gallery sister id required pick format type multi pick format single dropdown pick format multi checkbox pick allow add new pick taggable pick show icon pick show edit link pick show view link pick limit pick user role pick post status admin only restrict role restrict capability hidden read only roles allowed rest read rest write rest pick response array rest pick depth unique pick select text pick table id pick table index pick display pick where pick orderby pick groupby rel category featured images copy id name rel category featured images copy label rel category featured images copy description help class type pick weight pick object taxonomy pick val product cat sister id required unique pick format type multi pick format single dropdown pick format multi checkbox pick allow add new pick taggable pick show icon pick show edit link pick show view link pick limit pick user role pick post status admin only restrict role restrict capability hidden read only roles allowed rest read rest write rest pick response array rest pick depth pick select text pick table id pick table index pick display pick where pick orderby pick groupby show in menu pfat enable pfat run outside loop pfat append single append pfat filter single the content pfat append archive append pfat filter archive the content rest enable read all write all related issues and or prs todos tests documentation wordpress environment copy and paste your system details from pods admin settings debug information in wordpress admin here
| 0
|
58,972
| 16,984,270,802
|
IssuesEvent
|
2021-06-30 12:45:45
|
galasa-dev/projectmanagement
|
https://api.github.com/repos/galasa-dev/projectmanagement
|
closed
|
zOS Manager Documentation Links Not Appearing Correctly
|
defect documentation
|
There are some sections on the[ zOS Manager documentation](https://galasa.dev/docs/managers/zos-manager) that look like they should be formatting as a link, but are not. e.g. `{@link IZosFileHandler#newDataset(String, dev.galasa.zos.IZosImage)}`
I believe it should possibly, be formatting like so: `@IZosFileHandler#newDataset(String, dev.galasa.zos.IZosImage)` so I am unsure of why the `@link` may be there.
This can be seen when navigating (on the page linked to above) to the "Notes" field of any accordion item in the "Annotations" section.
|
1.0
|
zOS Manager Documentation Links Not Appearing Correctly - There are some sections on the[ zOS Manager documentation](https://galasa.dev/docs/managers/zos-manager) that look like they should be formatting as a link, but are not. e.g. `{@link IZosFileHandler#newDataset(String, dev.galasa.zos.IZosImage)}`
I believe it should possibly, be formatting like so: `@IZosFileHandler#newDataset(String, dev.galasa.zos.IZosImage)` so I am unsure of why the `@link` may be there.
This can be seen when navigating (on the page linked to above) to the "Notes" field of any accordion item in the "Annotations" section.
|
defect
|
zos manager documentation links not appearing correctly there are some sections on the that look like they should be formatting as a link but are not e g link izosfilehandler newdataset string dev galasa zos izosimage i believe it should possibly be formatting like so izosfilehandler newdataset string dev galasa zos izosimage so i am unsure of why the link may be there this can be seen when navigating on the page linked to above to the notes field of any accordion item in the annotations section
| 1
|
291,481
| 8,925,657,267
|
IssuesEvent
|
2019-01-21 23:55:47
|
Lustrel/money-management
|
https://api.github.com/repos/Lustrel/money-management
|
closed
|
Update Instalments' status when there is a late payment
|
high-priority
|
Whenever a Instalment is not paid in time, we should update its status to `in_arrears` (`Entity\Installment::$ID_IN_ARREARS`).
Since the customer's server doesn't support workers (also known as cron jobs), we need to implement an alternative.
As discussed before, we can start by performing a check on all instalments whenever a user access the "Parcelas" screen. To avoid performing this check on every access (since we only need this once a day), we can store in an auxiliar db table the date of the last update... and do this everyday, for the first user that access the page.
Oh, and one thing that must be remembered (perhaps it is worth to be implemented on this same task): instalments with late payments should be updated by X% (fee). We still need to check with the client which is the correct percentage value.
|
1.0
|
Update Instalments' status when there is a late payment - Whenever a Instalment is not paid in time, we should update its status to `in_arrears` (`Entity\Installment::$ID_IN_ARREARS`).
Since the customer's server doesn't support workers (also known as cron jobs), we need to implement an alternative.
As discussed before, we can start by performing a check on all instalments whenever a user access the "Parcelas" screen. To avoid performing this check on every access (since we only need this once a day), we can store in an auxiliar db table the date of the last update... and do this everyday, for the first user that access the page.
Oh, and one thing that must be remembered (perhaps it is worth to be implemented on this same task): instalments with late payments should be updated by X% (fee). We still need to check with the client which is the correct percentage value.
|
non_defect
|
update instalments status when there is a late payment whenever a instalment is not paid in time we should update its status to in arrears entity installment id in arrears since the customer s server doesn t support workers also known as cron jobs we need to implement an alternative as discussed before we can start by performing a check on all instalments whenever a user access the parcelas screen to avoid performing this check on every access since we only need this once a day we can store in an auxiliar db table the date of the last update and do this everyday for the first user that access the page oh and one thing that must be remembered perhaps it is worth to be implemented on this same task instalments with late payments should be updated by x fee we still need to check with the client which is the correct percentage value
| 0
|
78,878
| 27,803,229,537
|
IssuesEvent
|
2023-03-17 17:27:51
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG: pdf(.) of vonmises distribution returns nan when kappa=0
|
defect
|
### Describe your issue.
The probability density function a `scipy.stats.vonmises` returns `nan` when called with `kappa = 0`, whereas it should return ~0.159.
The `_pdf` method of the `vonmises_gen` class in `_continuous_distns.py` should return the correct results. My guess is that things go wrong in the `pdf` method of the `rv_continous` in `_distn_infrastructure.py`, where `scale = 0` is treated as a special case.
I tried to write a patch that fixes this, but don't have enough insight into the details of distribution handling in `scipy` to not potentially screw up some other distributions.
### Reproducing Code Example
```python
import scipy.stats
print(scipy.stats.vonmises.pdf(x=0, kappa=1))
print(scipy.stats.vonmises.pdf(x=0, kappa=0))
```
### Error message
```shell
0.3417104886234632
nan
```
### SciPy/NumPy/Python version and system information
```shell
1.9.1 1.23.3 sys.version_info(major=3, minor=10, micro=10, releaselevel='final', serial=0)
lapack_mkl_info:
NOT AVAILABLE
openblas_lapack_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
lapack_opt_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
blas_mkl_info:
NOT AVAILABLE
blis_info:
NOT AVAILABLE
openblas_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
blas_opt_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
```
|
1.0
|
BUG: pdf(.) of vonmises distribution returns nan when kappa=0 - ### Describe your issue.
The probability density function a `scipy.stats.vonmises` returns `nan` when called with `kappa = 0`, whereas it should return ~0.159.
The `_pdf` method of the `vonmises_gen` class in `_continuous_distns.py` should return the correct results. My guess is that things go wrong in the `pdf` method of the `rv_continous` in `_distn_infrastructure.py`, where `scale = 0` is treated as a special case.
I tried to write a patch that fixes this, but don't have enough insight into the details of distribution handling in `scipy` to not potentially screw up some other distributions.
### Reproducing Code Example
```python
import scipy.stats
print(scipy.stats.vonmises.pdf(x=0, kappa=1))
print(scipy.stats.vonmises.pdf(x=0, kappa=0))
```
### Error message
```shell
0.3417104886234632
nan
```
### SciPy/NumPy/Python version and system information
```shell
1.9.1 1.23.3 sys.version_info(major=3, minor=10, micro=10, releaselevel='final', serial=0)
lapack_mkl_info:
NOT AVAILABLE
openblas_lapack_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
lapack_opt_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
blas_mkl_info:
NOT AVAILABLE
blis_info:
NOT AVAILABLE
openblas_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
blas_opt_info:
libraries = ['openblas', 'openblas']
library_dirs = ['/opt/arm64-builds/lib']
language = c
define_macros = [('HAVE_CBLAS', None)]
runtime_library_dirs = ['/opt/arm64-builds/lib']
```
|
defect
|
bug pdf of vonmises distribution returns nan when kappa describe your issue the probability density function a scipy stats vonmises returns nan when called with kappa whereas it should return the pdf method of the vonmises gen class in continuous distns py should return the correct results my guess is that things go wrong in the pdf method of the rv continous in distn infrastructure py where scale is treated as a special case i tried to write a patch that fixes this but don t have enough insight into the details of distribution handling in scipy to not potentially screw up some other distributions reproducing code example python import scipy stats print scipy stats vonmises pdf x kappa print scipy stats vonmises pdf x kappa error message shell nan scipy numpy python version and system information shell sys version info major minor micro releaselevel final serial lapack mkl info not available openblas lapack info libraries library dirs language c define macros runtime library dirs lapack opt info libraries library dirs language c define macros runtime library dirs blas mkl info not available blis info not available openblas info libraries library dirs language c define macros runtime library dirs blas opt info libraries library dirs language c define macros runtime library dirs
| 1
|
17,364
| 10,021,375,304
|
IssuesEvent
|
2019-07-16 14:31:25
|
MicrosoftDocs/windows-itpro-docs
|
https://api.github.com/repos/MicrosoftDocs/windows-itpro-docs
|
closed
|
AMD and MBEC
|
security
|
This page states that HVCI works on both Intel and AMD platforms, however it also states that MBEC is required, which is not present on any AMD CPU ("HVCI works with [...] AMD. CPU new feature is required Mode based execution control (MBE) Virtualization.")
Does that means that it uses the software emulation mentioned later (does it have an impact on performance?) or there's more magic CPU features on AMD that allows it to work? In this case, what would be the AMD equivalent to a 7th generation Intel Core CPU?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 1526c4cb-4584-7fc6-e633-a48e13dad7e8
* Version Independent ID: 13a88ed8-f43d-f658-cbe4-d148d5ffaa12
* Content: [Enable virtualization-based protection of code integrity](https://docs.microsoft.com/en-us/windows/security/threat-protection/windows-defender-exploit-guard/enable-virtualization-based-protection-of-code-integrity#feedback)
* Content Source: [windows/security/threat-protection/windows-defender-exploit-guard/enable-virtualization-based-protection-of-code-integrity.md](https://github.com/MicrosoftDocs/windows-itpro-docs/blob/master/windows/security/threat-protection/windows-defender-exploit-guard/enable-virtualization-based-protection-of-code-integrity.md)
* Product: **w10**
* Technology: **windows**
* GitHub Login: @levinec
* Microsoft Alias: **ellevin**
|
True
|
AMD and MBEC - This page states that HVCI works on both Intel and AMD platforms, however it also states that MBEC is required, which is not present on any AMD CPU ("HVCI works with [...] AMD. CPU new feature is required Mode based execution control (MBE) Virtualization.")
Does that means that it uses the software emulation mentioned later (does it have an impact on performance?) or there's more magic CPU features on AMD that allows it to work? In this case, what would be the AMD equivalent to a 7th generation Intel Core CPU?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 1526c4cb-4584-7fc6-e633-a48e13dad7e8
* Version Independent ID: 13a88ed8-f43d-f658-cbe4-d148d5ffaa12
* Content: [Enable virtualization-based protection of code integrity](https://docs.microsoft.com/en-us/windows/security/threat-protection/windows-defender-exploit-guard/enable-virtualization-based-protection-of-code-integrity#feedback)
* Content Source: [windows/security/threat-protection/windows-defender-exploit-guard/enable-virtualization-based-protection-of-code-integrity.md](https://github.com/MicrosoftDocs/windows-itpro-docs/blob/master/windows/security/threat-protection/windows-defender-exploit-guard/enable-virtualization-based-protection-of-code-integrity.md)
* Product: **w10**
* Technology: **windows**
* GitHub Login: @levinec
* Microsoft Alias: **ellevin**
|
non_defect
|
amd and mbec this page states that hvci works on both intel and amd platforms however it also states that mbec is required which is not present on any amd cpu hvci works with amd cpu new feature is required mode based execution control mbe virtualization does that means that it uses the software emulation mentioned later does it have an impact on performance or there s more magic cpu features on amd that allows it to work in this case what would be the amd equivalent to a generation intel core cpu document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product technology windows github login levinec microsoft alias ellevin
| 0
|
71,951
| 9,545,347,019
|
IssuesEvent
|
2019-05-01 16:47:01
|
adobe/aem-core-wcm-components
|
https://api.github.com/repos/adobe/aem-core-wcm-components
|
closed
|
[ContentFragmentList] Add Documentation
|
documentation done
|
The newly-created Content Fragment List component requires documentation:
* Creation of a component README.md at [0].
* Clarification over the component icon to use for this component.
[0] - https://github.com/adobe/aem-core-wcm-components/tree/development/content/src/content/jcr_root/apps/core/wcm/components/contentfragmentlist/v1/contentfragmentlist
CQ-4266573
|
1.0
|
[ContentFragmentList] Add Documentation - The newly-created Content Fragment List component requires documentation:
* Creation of a component README.md at [0].
* Clarification over the component icon to use for this component.
[0] - https://github.com/adobe/aem-core-wcm-components/tree/development/content/src/content/jcr_root/apps/core/wcm/components/contentfragmentlist/v1/contentfragmentlist
CQ-4266573
|
non_defect
|
add documentation the newly created content fragment list component requires documentation creation of a component readme md at clarification over the component icon to use for this component cq
| 0
|
58,815
| 16,794,584,707
|
IssuesEvent
|
2021-06-16 00:30:27
|
ascott18/TellMeWhen
|
https://api.github.com/repos/ascott18/TellMeWhen
|
closed
|
[Bug] Unit range does not work in some dungeons
|
S: more-info-needed S: needs-reproduction T: defect V: retail
|
Im using version 9.0.7
The unit range condition does not work in some dungeons.
Always appears as imprecise.
|
1.0
|
[Bug] Unit range does not work in some dungeons - Im using version 9.0.7
The unit range condition does not work in some dungeons.
Always appears as imprecise.
|
defect
|
unit range does not work in some dungeons im using version the unit range condition does not work in some dungeons always appears as imprecise
| 1
|
18,022
| 12,493,285,259
|
IssuesEvent
|
2020-06-01 08:58:27
|
ClickHouse/ClickHouse
|
https://api.github.com/repos/ClickHouse/ClickHouse
|
opened
|
Better diagnostics on mutation failure.
|
feature usability
|
Mutations and alters are completely asynchronous operations. If we start some incorrect `ALTER` query from `clickhouse-client ` it will hang forever (if `mutations_sync=2` or `alter_repliction_partitions_sync=1 or 2` specified). We need to implement the check of `MutationStatus` and if some exception happened, we should rethrow it to the client.
Related issues: #11166.
|
True
|
Better diagnostics on mutation failure. - Mutations and alters are completely asynchronous operations. If we start some incorrect `ALTER` query from `clickhouse-client ` it will hang forever (if `mutations_sync=2` or `alter_repliction_partitions_sync=1 or 2` specified). We need to implement the check of `MutationStatus` and if some exception happened, we should rethrow it to the client.
Related issues: #11166.
|
non_defect
|
better diagnostics on mutation failure mutations and alters are completely asynchronous operations if we start some incorrect alter query from clickhouse client it will hang forever if mutations sync or alter repliction partitions sync or specified we need to implement the check of mutationstatus and if some exception happened we should rethrow it to the client related issues
| 0
|
169,674
| 20,841,859,792
|
IssuesEvent
|
2022-03-21 01:42:14
|
benlazarine/datahog
|
https://api.github.com/repos/benlazarine/datahog
|
opened
|
CVE-2022-24772 (High) detected in node-forge-0.7.6.tgz
|
security vulnerability
|
## CVE-2022-24772 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.7.6.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz</a></p>
<p>Path to dependency file: /react/package.json</p>
<p>Path to vulnerable library: /react/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- parcel-bundler-1.12.4.tgz (Root Library)
- :x: **node-forge-0.7.6.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24772>CVE-2022-24772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-24772 (High) detected in node-forge-0.7.6.tgz - ## CVE-2022-24772 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.7.6.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.6.tgz</a></p>
<p>Path to dependency file: /react/package.json</p>
<p>Path to vulnerable library: /react/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- parcel-bundler-1.12.4.tgz (Root Library)
- :x: **node-forge-0.7.6.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24772>CVE-2022-24772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in node forge tgz cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file react package json path to vulnerable library react node modules node forge package json dependency hierarchy parcel bundler tgz root library x node forge tgz vulnerable library vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not check for tailing garbage bytes after decoding a digestinfo asn structure this can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource
| 0
|
28,326
| 5,241,273,309
|
IssuesEvent
|
2017-01-31 15:20:39
|
google/gtm-oauth2
|
https://api.github.com/repos/google/gtm-oauth2
|
closed
|
Unable to customize navigationBar/navigationItem properties if GTMOAuth2ViewController is the rootViewController
|
auto-migrated Priority-Medium Type-Defect
|
```
As the title I'm unable to customize these properties.
I'm looking to edit the navigationBar color, navigationBar
button/attributedString attributes, navigationBar title/attributedString
attributes.
I've already tried overriding setUpNavigation and subclassing, but the result
is shitty; with shitty I means:
1. If I set the attributedString attributes the text will disappear
2. Can't even set the tintColor
This is happening only with this class.
```
Original issue reported on code.google.com by `playerst...@gmail.com` on 13 Dec 2013 at 6:24
|
1.0
|
Unable to customize navigationBar/navigationItem properties if GTMOAuth2ViewController is the rootViewController - ```
As the title I'm unable to customize these properties.
I'm looking to edit the navigationBar color, navigationBar
button/attributedString attributes, navigationBar title/attributedString
attributes.
I've already tried overriding setUpNavigation and subclassing, but the result
is shitty; with shitty I means:
1. If I set the attributedString attributes the text will disappear
2. Can't even set the tintColor
This is happening only with this class.
```
Original issue reported on code.google.com by `playerst...@gmail.com` on 13 Dec 2013 at 6:24
|
defect
|
unable to customize navigationbar navigationitem properties if is the rootviewcontroller as the title i m unable to customize these properties i m looking to edit the navigationbar color navigationbar button attributedstring attributes navigationbar title attributedstring attributes i ve already tried overriding setupnavigation and subclassing but the result is shitty with shitty i means if i set the attributedstring attributes the text will disappear can t even set the tintcolor this is happening only with this class original issue reported on code google com by playerst gmail com on dec at
| 1
|
43,610
| 11,777,135,538
|
IssuesEvent
|
2020-03-16 14:22:29
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
closed
|
[DEFECT] Windows plugin tests failing
|
defect priority_normal
|
--------
Defect Description
--------
##### What did you expect to see happen?
The windows plugin tests are failing, for example:
https://civet.inl.gov/job/480411/
##### What did you see instead?
The tests pass.
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
1.0
|
[DEFECT] Windows plugin tests failing - --------
Defect Description
--------
##### What did you expect to see happen?
The windows plugin tests are failing, for example:
https://civet.inl.gov/job/480411/
##### What did you see instead?
The tests pass.
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
|
defect
|
windows plugin tests failing defect description what did you expect to see happen the windows plugin tests are failing for example what did you see instead the tests pass do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
| 1
|
19,889
| 3,273,628,510
|
IssuesEvent
|
2015-10-26 04:24:42
|
npgall/cqengine
|
https://api.github.com/repos/npgall/cqengine
|
closed
|
Set remove() and removeAll() doesn't follow substitutability principle.
|
auto-migrated Priority-Medium Type-Defect
|
```
Set.remove(o) takes on Object where as Set.add(E) takes generic, this is
because Set.remove() can not damage internals of Set representation even if the
type doesn't match the element. Because only if the element is present it gets
removed.
However in IndexedCollection implementation we pass the object after (an
unchecked cast), to the indices. Now the indices are called even the object is
removed or not. This causes Class Cast Exception. Instead we should check first
if the element is a member only then call index update steps.
More here:
http://smallwig.blogspot.in/2007/12/why-does-setcontains-take-object-not-e.html
```
Original issue reported on code.google.com by `anita.v...@inmobi.com` on 9 Nov 2013 at 9:40
Attachments:
* [uniform_removal.patch](https://storage.googleapis.com/google-code-attachments/cqengine/issue-27/comment-0/uniform_removal.patch)
|
1.0
|
Set remove() and removeAll() doesn't follow substitutability principle. - ```
Set.remove(o) takes on Object where as Set.add(E) takes generic, this is
because Set.remove() can not damage internals of Set representation even if the
type doesn't match the element. Because only if the element is present it gets
removed.
However in IndexedCollection implementation we pass the object after (an
unchecked cast), to the indices. Now the indices are called even the object is
removed or not. This causes Class Cast Exception. Instead we should check first
if the element is a member only then call index update steps.
More here:
http://smallwig.blogspot.in/2007/12/why-does-setcontains-take-object-not-e.html
```
Original issue reported on code.google.com by `anita.v...@inmobi.com` on 9 Nov 2013 at 9:40
Attachments:
* [uniform_removal.patch](https://storage.googleapis.com/google-code-attachments/cqengine/issue-27/comment-0/uniform_removal.patch)
|
defect
|
set remove and removeall doesn t follow substitutability principle set remove o takes on object where as set add e takes generic this is because set remove can not damage internals of set representation even if the type doesn t match the element because only if the element is present it gets removed however in indexedcollection implementation we pass the object after an unchecked cast to the indices now the indices are called even the object is removed or not this causes class cast exception instead we should check first if the element is a member only then call index update steps more here original issue reported on code google com by anita v inmobi com on nov at attachments
| 1
|
25,784
| 4,445,398,271
|
IssuesEvent
|
2016-08-20 01:57:50
|
FoldingAtHome/fah-client-pub
|
https://api.github.com/repos/FoldingAtHome/fah-client-pub
|
closed
|
Don't delete .nfs files
|
defect wontfix
|
Trac | Data
---: | :---
Ticket | 577
Reported by | @sandstar
Status | accepted
Component | FAHClient
Priority | 7 (minor)
Version | 7.1
When the client is run from NFS shares there's a minor problem when work directories are being cleaned of files.
The client tries to delete every file in a specific work directory-slot but can't remove the .nfs file. These files are created by the NFS-server and the NFS-server will delete them once there are no active files in that directory.
It would be nice if the client would not try to delete these .nfs files.
(I've just learned about .nfs files so I could be wrong a little bit here and there)
|
1.0
|
Don't delete .nfs files - Trac | Data
---: | :---
Ticket | 577
Reported by | @sandstar
Status | accepted
Component | FAHClient
Priority | 7 (minor)
Version | 7.1
When the client is run from NFS shares there's a minor problem when work directories are being cleaned of files.
The client tries to delete every file in a specific work directory-slot but can't remove the .nfs file. These files are created by the NFS-server and the NFS-server will delete them once there are no active files in that directory.
It would be nice if the client would not try to delete these .nfs files.
(I've just learned about .nfs files so I could be wrong a little bit here and there)
|
defect
|
don t delete nfs files trac data ticket reported by sandstar status accepted component fahclient priority minor version when the client is run from nfs shares there s a minor problem when work directories are being cleaned of files the client tries to delete every file in a specific work directory slot but can t remove the nfs file these files are created by the nfs server and the nfs server will delete them once there are no active files in that directory it would be nice if the client would not try to delete these nfs files i ve just learned about nfs files so i could be wrong a little bit here and there
| 1
|
34,026
| 7,327,469,956
|
IssuesEvent
|
2018-03-04 10:37:12
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
width and height fails for Charts
|
defect
|
**I'm submitting a ...**
```
[X] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Current behavior**
<!-- Describe how the bug manifests. -->
p-chart attribute height = "50%" don't work in PRIME NG 5.2.0, work in PRIME NG 5.0.2
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
* **Angular version:** 5.2.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 5.2.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ]
<!-- All browsers where this could be reproduced -->
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
* **Node (for AoT issues):** `node --version` =
|
1.0
|
width and height fails for Charts - **I'm submitting a ...**
```
[X] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Current behavior**
<!-- Describe how the bug manifests. -->
p-chart attribute height = "50%" don't work in PRIME NG 5.2.0, work in PRIME NG 5.0.2
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
**Please tell us about your environment:**
<!-- Operating system, IDE, package manager, HTTP server, ... -->
* **Angular version:** 5.2.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 5.2.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ]
<!-- All browsers where this could be reproduced -->
* **Language:** [all | TypeScript X.X | ES6/7 | ES5]
* **Node (for AoT issues):** `node --version` =
|
defect
|
width and height fails for charts i m submitting a bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see current behavior p chart attribute height don t work in prime ng work in prime ng what is the motivation use case for changing the behavior please tell us about your environment angular version x primeng version x browser language node for aot issues node version
| 1
|
58,878
| 16,851,754,860
|
IssuesEvent
|
2021-06-20 16:53:39
|
mozilla/jetstream
|
https://api.github.com/repos/mozilla/jetstream
|
closed
|
Max retries exceeded in ensure-enrollments
|
Defect
|
I'm seeing that yesterday's analysis run of `reminder-set-firefox-as-default` failed due to `Exception: Too many retries for https://experimenter.services.mozilla.com/api/v1/experiments/` in the `ensure-enrollments` step. I only noticed because I looked at the Argo dashboard.
We should make sure that the step get retried and additionally see if it makes sense to log to BigQuery if experiment analysis does not complete because the number of retries has been exceeded.
┆Issue is synchronized with this [Jira Task](https://jira.mozilla.com/browse/CIRRUS-90)
┆Issue Number: CIRRUS-90
|
1.0
|
Max retries exceeded in ensure-enrollments - I'm seeing that yesterday's analysis run of `reminder-set-firefox-as-default` failed due to `Exception: Too many retries for https://experimenter.services.mozilla.com/api/v1/experiments/` in the `ensure-enrollments` step. I only noticed because I looked at the Argo dashboard.
We should make sure that the step get retried and additionally see if it makes sense to log to BigQuery if experiment analysis does not complete because the number of retries has been exceeded.
┆Issue is synchronized with this [Jira Task](https://jira.mozilla.com/browse/CIRRUS-90)
┆Issue Number: CIRRUS-90
|
defect
|
max retries exceeded in ensure enrollments i m seeing that yesterday s analysis run of reminder set firefox as default failed due to exception too many retries for in the ensure enrollments step i only noticed because i looked at the argo dashboard we should make sure that the step get retried and additionally see if it makes sense to log to bigquery if experiment analysis does not complete because the number of retries has been exceeded ┆issue is synchronized with this ┆issue number cirrus
| 1
|
53,960
| 29,190,844,959
|
IssuesEvent
|
2023-05-19 19:50:02
|
MaterializeInc/materialize
|
https://api.github.com/repos/MaterializeInc/materialize
|
closed
|
replica isolation: multiple replicas cause degradation in performance of SELECT statements due to conflicts in Postgres
|
C-bug T-performance P3
|
### What version of Materialize are you using?
v0.26.1-dev (df970e97b)
### How did you install Materialize?
Built from source
### What is the issue?
The presence of multiple replicas causes SELECT statements to take ~5 seconds, likely due to contention on the Postgres side. The error log has this:
```
materialized: 2022-06-08T10:36:01.503433Z INFO mz_persist_client::r#impl::machine: external operation apply_unbatched_cmd::cas failed, retrying in 32ms: determinate: db error: ERROR: could not serialize access due to read/write dependencies among transactions
DETAIL: Reason code: Canceled on commit attempt with conflict in from prepared pivot.
HINT: The transaction might succeed if retried.: ERROR: could not serialize access due to read/write dependencies among transactions
DETAIL: Reason code: Canceled on commit attempt with conflict in from prepared pivot.
HINT: The transaction might succeed if retried.
```
To reproduce:
```
\timing on
drop cluster default cascade;
create cluster default replicas();
create table t1 (f1 integer);
insert into t1 values (5);
create cluster replica default.replica1 SIZE '1';
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
create cluster replica default.replica2 SIZE '1';
create cluster replica default.replica3 SIZE '1';
create cluster replica default.replica4 SIZE '1';
create cluster replica default.replica5 SIZE '1';
create cluster replica default.replica6 SIZE '1';
create cluster replica default.replica7 SIZE '1';
create cluster replica default.replica8 SIZE '1';
create cluster replica default.replica9 SIZE '1';
create cluster replica default.replica10 SIZE '1';
create cluster replica default.replica11 SIZE '1';
create cluster replica default.replica12 SIZE '1';
create cluster replica default.replica13 SIZE '1';
create cluster replica default.replica14 SIZE '1';
create cluster replica default.replica15 SIZE '1';
create cluster replica default.replica16 SIZE '1';
create cluster replica default.replica17 SIZE '1';
create cluster replica default.replica18 SIZE '1';
create cluster replica default.replica19 SIZE '1';
create cluster replica default.replica20 SIZE '1';
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
```
### Relevant log output
_No response_
|
True
|
replica isolation: multiple replicas cause degradation in performance of SELECT statements due to conflicts in Postgres - ### What version of Materialize are you using?
v0.26.1-dev (df970e97b)
### How did you install Materialize?
Built from source
### What is the issue?
The presence of multiple replicas causes SELECT statements to take ~5 seconds, likely due to contention on the Postgres side. The error log has this:
```
materialized: 2022-06-08T10:36:01.503433Z INFO mz_persist_client::r#impl::machine: external operation apply_unbatched_cmd::cas failed, retrying in 32ms: determinate: db error: ERROR: could not serialize access due to read/write dependencies among transactions
DETAIL: Reason code: Canceled on commit attempt with conflict in from prepared pivot.
HINT: The transaction might succeed if retried.: ERROR: could not serialize access due to read/write dependencies among transactions
DETAIL: Reason code: Canceled on commit attempt with conflict in from prepared pivot.
HINT: The transaction might succeed if retried.
```
To reproduce:
```
\timing on
drop cluster default cascade;
create cluster default replicas();
create table t1 (f1 integer);
insert into t1 values (5);
create cluster replica default.replica1 SIZE '1';
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
create cluster replica default.replica2 SIZE '1';
create cluster replica default.replica3 SIZE '1';
create cluster replica default.replica4 SIZE '1';
create cluster replica default.replica5 SIZE '1';
create cluster replica default.replica6 SIZE '1';
create cluster replica default.replica7 SIZE '1';
create cluster replica default.replica8 SIZE '1';
create cluster replica default.replica9 SIZE '1';
create cluster replica default.replica10 SIZE '1';
create cluster replica default.replica11 SIZE '1';
create cluster replica default.replica12 SIZE '1';
create cluster replica default.replica13 SIZE '1';
create cluster replica default.replica14 SIZE '1';
create cluster replica default.replica15 SIZE '1';
create cluster replica default.replica16 SIZE '1';
create cluster replica default.replica17 SIZE '1';
create cluster replica default.replica18 SIZE '1';
create cluster replica default.replica19 SIZE '1';
create cluster replica default.replica20 SIZE '1';
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
select * from t1;
```
### Relevant log output
_No response_
|
non_defect
|
replica isolation multiple replicas cause degradation in performance of select statements due to conflicts in postgres what version of materialize are you using dev how did you install materialize built from source what is the issue the presence of multiple replicas causes select statements to take seconds likely due to contention on the postgres side the error log has this materialized info mz persist client r impl machine external operation apply unbatched cmd cas failed retrying in determinate db error error could not serialize access due to read write dependencies among transactions detail reason code canceled on commit attempt with conflict in from prepared pivot hint the transaction might succeed if retried error could not serialize access due to read write dependencies among transactions detail reason code canceled on commit attempt with conflict in from prepared pivot hint the transaction might succeed if retried to reproduce timing on drop cluster default cascade create cluster default replicas create table integer insert into values create cluster replica default size select from select from select from select from select from select from select from select from select from select from create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size create cluster replica default size select from select from select from select from select from select from select from select from select from select from relevant log output no response
| 0
|
47,877
| 13,066,318,138
|
IssuesEvent
|
2020-07-30 21:26:37
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
[dataio] dataio-pyshovel frame mixing (Trac #1338)
|
Migrated from Trac combo core defect
|
When examining a file like:
GCDG
The last G frame will have C and D objects mixed into it. This is obviously wrong, so something isn't right.
Migrated from https://code.icecube.wisc.edu/ticket/1338
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:10",
"description": "When examining a file like:\n\nGCDG\n\nThe last G frame will have C and D objects mixed into it. This is obviously wrong, so something isn't right.",
"reporter": "david.schultz",
"cc": "",
"resolution": "invalid",
"_ts": "1458335650323600",
"component": "combo core",
"summary": "[dataio] dataio-pyshovel frame mixing",
"priority": "critical",
"keywords": "",
"time": "2015-09-10T16:30:42",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
1.0
|
[dataio] dataio-pyshovel frame mixing (Trac #1338) - When examining a file like:
GCDG
The last G frame will have C and D objects mixed into it. This is obviously wrong, so something isn't right.
Migrated from https://code.icecube.wisc.edu/ticket/1338
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:10",
"description": "When examining a file like:\n\nGCDG\n\nThe last G frame will have C and D objects mixed into it. This is obviously wrong, so something isn't right.",
"reporter": "david.schultz",
"cc": "",
"resolution": "invalid",
"_ts": "1458335650323600",
"component": "combo core",
"summary": "[dataio] dataio-pyshovel frame mixing",
"priority": "critical",
"keywords": "",
"time": "2015-09-10T16:30:42",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
|
defect
|
dataio pyshovel frame mixing trac when examining a file like gcdg the last g frame will have c and d objects mixed into it this is obviously wrong so something isn t right migrated from json status closed changetime description when examining a file like n ngcdg n nthe last g frame will have c and d objects mixed into it this is obviously wrong so something isn t right reporter david schultz cc resolution invalid ts component combo core summary dataio pyshovel frame mixing priority critical keywords time milestone owner david schultz type defect
| 1
|
806,136
| 29,802,745,193
|
IssuesEvent
|
2023-06-16 09:19:49
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.twitch.tv - video or audio doesn't play
|
browser-firefox priority-important engine-gecko os-win11
|
<!-- @browser: Firefox 116.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/116.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/123486 -->
**URL**: https://www.twitch.tv/
**Browser / Version**: Firefox 116.0
**Operating System**: Windows 11
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Muting tab allows playback of video, when tab is unmuted, video stops playing. Does not occur on other browsers, including Firefox 113.0.2
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/c2c30c91-859b-4cb9-aa16-e7598ebf15df.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230610094613</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/6/ba8eb9a7-9759-4467-b3e0-26e0e262bec8)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.twitch.tv - video or audio doesn't play - <!-- @browser: Firefox 116.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/116.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/123486 -->
**URL**: https://www.twitch.tv/
**Browser / Version**: Firefox 116.0
**Operating System**: Windows 11
**Tested Another Browser**: Yes Chrome
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Muting tab allows playback of video, when tab is unmuted, video stops playing. Does not occur on other browsers, including Firefox 113.0.2
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/6/c2c30c91-859b-4cb9-aa16-e7598ebf15df.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20230610094613</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/6/ba8eb9a7-9759-4467-b3e0-26e0e262bec8)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
video or audio doesn t play url browser version firefox operating system windows tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce muting tab allows playback of video when tab is unmuted video stops playing does not occur on other browsers including firefox view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
108,085
| 4,327,056,593
|
IssuesEvent
|
2016-07-26 09:07:23
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
[k8s.io] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] {Kubernetes e2e suite}
|
kind/flake priority/P2
|
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke/11648/
Failed: [k8s.io] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:209
Expected error:
<*errors.errorString | 0xc820652e70>: {
s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.30.15 --kubeconfig=/workspace/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-5z9m9] [] <nil> failed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.\n [] <nil> 0xc820c8ba40 exit status 1 <nil> true [0xc820542440 0xc820542458 0xc820542470] [0xc820542440 0xc820542458 0xc820542470] [0xc820542450 0xc820542468] [0xab1c70 0xab1c70] 0xc820be0cc0}:\nCommand stdout:\n\nstderr:\nfailed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.\n\nerror:\nexit status 1\n",
}
Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.30.15 --kubeconfig=/workspace/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-5z9m9] [] <nil> failed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.
[] <nil> 0xc820c8ba40 exit status 1 <nil> true [0xc820542440 0xc820542458 0xc820542470] [0xc820542440 0xc820542458 0xc820542470] [0xc820542450 0xc820542468] [0xab1c70 0xab1c70] 0xc820be0cc0}:
Command stdout:
stderr:
failed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.
error:
exit status 1
not to have occurred
```
Previous issues for this test: #28565 #29072
|
1.0
|
[k8s.io] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] {Kubernetes e2e suite} - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke/11648/
Failed: [k8s.io] Kubectl client [k8s.io] Update Demo should create and stop a replication controller [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:209
Expected error:
<*errors.errorString | 0xc820652e70>: {
s: "Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.30.15 --kubeconfig=/workspace/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-5z9m9] [] <nil> failed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.\n [] <nil> 0xc820c8ba40 exit status 1 <nil> true [0xc820542440 0xc820542458 0xc820542470] [0xc820542440 0xc820542458 0xc820542470] [0xc820542450 0xc820542468] [0xab1c70 0xab1c70] 0xc820be0cc0}:\nCommand stdout:\n\nstderr:\nfailed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.\n\nerror:\nexit status 1\n",
}
Error running &{/workspace/kubernetes/platforms/linux/amd64/kubectl [kubectl --server=https://104.198.30.15 --kubeconfig=/workspace/.kube/config get rc,svc -l name=update-demo --no-headers --namespace=e2e-tests-kubectl-5z9m9] [] <nil> failed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.
[] <nil> 0xc820c8ba40 exit status 1 <nil> true [0xc820542440 0xc820542458 0xc820542470] [0xc820542440 0xc820542458 0xc820542470] [0xc820542450 0xc820542468] [0xab1c70 0xab1c70] 0xc820be0cc0}:
Command stdout:
stderr:
failed to find client for version v1: error: google: could not find default credentials. See https://developers.google.com/accounts/docs/application-default-credentials for more information.
error:
exit status 1
not to have occurred
```
Previous issues for this test: #28565 #29072
|
non_defect
|
kubectl client update demo should create and stop a replication controller kubernetes suite failed kubectl client update demo should create and stop a replication controller kubernetes suite go src io kubernetes output dockerized go src io kubernetes test kubectl go expected error s error running workspace kubernetes platforms linux kubectl failed to find client for version error google could not find default credentials see for more information n exit status true ncommand stdout n nstderr nfailed to find client for version error google could not find default credentials see for more information n nerror nexit status n error running workspace kubernetes platforms linux kubectl failed to find client for version error google could not find default credentials see for more information exit status true command stdout stderr failed to find client for version error google could not find default credentials see for more information error exit status not to have occurred previous issues for this test
| 0
|
69,510
| 22,394,557,161
|
IssuesEvent
|
2022-06-17 11:06:48
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Missing parts in the timeline
|
T-Defect
|
### Steps to reproduce
Not entirely clear how this happens yet, but I think this is a pretty severe issue.
1. Use Element Desktop for a day
2. Next day, use Element again, possibly from a different session like phone or another computer
### Outcome
#### What did you expect?
Timeline is complete
#### What happened instead?
Some messages are missing from the timeline, including some I sent myself from the same session.
Can be worked around be restarting Element.
### Operating system
arch, win 10
### Application version
Element Nightly version: 2022061501 Olm version: 3.2.8 (arch), stable (win 10)
### How did you install the app?
aur, element.io
### Homeserver
synapse 1.60.0
### Will you send logs?
Yes
|
1.0
|
Missing parts in the timeline - ### Steps to reproduce
Not entirely clear how this happens yet, but I think this is a pretty severe issue.
1. Use Element Desktop for a day
2. Next day, use Element again, possibly from a different session like phone or another computer
### Outcome
#### What did you expect?
Timeline is complete
#### What happened instead?
Some messages are missing from the timeline, including some I sent myself from the same session.
Can be worked around be restarting Element.
### Operating system
arch, win 10
### Application version
Element Nightly version: 2022061501 Olm version: 3.2.8 (arch), stable (win 10)
### How did you install the app?
aur, element.io
### Homeserver
synapse 1.60.0
### Will you send logs?
Yes
|
defect
|
missing parts in the timeline steps to reproduce not entirely clear how this happens yet but i think this is a pretty severe issue use element desktop for a day next day use element again possibly from a different session like phone or another computer outcome what did you expect timeline is complete what happened instead some messages are missing from the timeline including some i sent myself from the same session can be worked around be restarting element operating system arch win application version element nightly version olm version arch stable win how did you install the app aur element io homeserver synapse will you send logs yes
| 1
|
16,117
| 11,841,767,745
|
IssuesEvent
|
2020-03-23 21:27:55
|
microsoft/react-native-windows
|
https://api.github.com/repos/microsoft/react-native-windows
|
closed
|
RNTester: Crash loading <Image> page
|
Area: Image Area: Infrastructure bug
|
The <Image> page in RNTester is having issues and crashes when you try to load it.
Callstack to the crash:
React.UWP.dll!winrt::to_hresult() Line 4298 C++
React.UWP.dll!winrt::terminate() Line 4372 C++
React.UWP.dll!std::experimental::coroutine_traits<winrt::fire_and_forget,react::uwp::ReactImage *,react::uwp::ImageSource>::promise_type::unhandled_exception() Line 7942 C++
> React.UWP.dll!react::uwp::ReactImage::Source$_ResumeCoro$2() Line 112 C++
React.UWP.dll!react::uwp::ReactImage::Source$_InitCoro$1() Line 112 C++
React.UWP.dll!react::uwp::ReactImage::Source(react::uwp::ImageSource source) Line 112 C++
React.UWP.dll!react::uwp::ImageViewManager::setSource(winrt::Windows::UI::Xaml::Controls::Canvas canvas, const folly::dynamic & data) Line 162 C++
React.UWP.dll!react::uwp::ImageViewManager::UpdateProperties(react::uwp::ShadowNodeBase * nodeToUpdate, const folly::dynamic & reactDiffMap) Line 124 C++
React.UWP.dll!react::uwp::ShadowNodeBase::updateProperties(const folly::dynamic && props) Line 28 C++
React.UWP.dll!facebook::react::UIManager::createView(__int64 tag, std::string && className, __int64 __formal, folly::dynamic && props) Line 260 C++
React.UWP.dll!facebook::react::UIManagerModule::getMethods::__l2::<lambda>(folly::dynamic args) Line 451 C++
Debug spew says:
Exception thrown at 0x764A35D2 (KernelBase.dll) in Playground.exe: WinRT originate error - 0x80070057 : 'legacy_image is not a valid absolute URI.'.
Exception thrown at 0x764A35D2 in Playground.exe: Microsoft C++ exception: winrt::hresult_invalid_argument at memory location 0x04AFC88C.
Exception thrown at 0x764A35D2 in Playground.exe: Microsoft C++ exception: [rethrow] at memory location 0x00000000.
Debugging back to the property setter, we have this:
| Name | Value | Type
-- | -- | -- | --
◢ | sources[0] | {uri="legacy_image" method="" bundleRootPath="ms-appx:///Bundle/" ...} | react::uwp::ImageSource
| ▶ uri | "legacy_image" | std::string
| ▶ method | "" | std::string
| ▶ bundleRootPath | "ms-appx:///Bundle/" | std::string
So, we have an <Image> with a uri of "legacy_image" and that's failing to resolve to a valid resource in our appx package. It looks like we either aren't packaging image resources for RNTester properly, or aren't resolving the URI propertly.
|
1.0
|
RNTester: Crash loading <Image> page - The <Image> page in RNTester is having issues and crashes when you try to load it.
Callstack to the crash:
React.UWP.dll!winrt::to_hresult() Line 4298 C++
React.UWP.dll!winrt::terminate() Line 4372 C++
React.UWP.dll!std::experimental::coroutine_traits<winrt::fire_and_forget,react::uwp::ReactImage *,react::uwp::ImageSource>::promise_type::unhandled_exception() Line 7942 C++
> React.UWP.dll!react::uwp::ReactImage::Source$_ResumeCoro$2() Line 112 C++
React.UWP.dll!react::uwp::ReactImage::Source$_InitCoro$1() Line 112 C++
React.UWP.dll!react::uwp::ReactImage::Source(react::uwp::ImageSource source) Line 112 C++
React.UWP.dll!react::uwp::ImageViewManager::setSource(winrt::Windows::UI::Xaml::Controls::Canvas canvas, const folly::dynamic & data) Line 162 C++
React.UWP.dll!react::uwp::ImageViewManager::UpdateProperties(react::uwp::ShadowNodeBase * nodeToUpdate, const folly::dynamic & reactDiffMap) Line 124 C++
React.UWP.dll!react::uwp::ShadowNodeBase::updateProperties(const folly::dynamic && props) Line 28 C++
React.UWP.dll!facebook::react::UIManager::createView(__int64 tag, std::string && className, __int64 __formal, folly::dynamic && props) Line 260 C++
React.UWP.dll!facebook::react::UIManagerModule::getMethods::__l2::<lambda>(folly::dynamic args) Line 451 C++
Debug spew says:
Exception thrown at 0x764A35D2 (KernelBase.dll) in Playground.exe: WinRT originate error - 0x80070057 : 'legacy_image is not a valid absolute URI.'.
Exception thrown at 0x764A35D2 in Playground.exe: Microsoft C++ exception: winrt::hresult_invalid_argument at memory location 0x04AFC88C.
Exception thrown at 0x764A35D2 in Playground.exe: Microsoft C++ exception: [rethrow] at memory location 0x00000000.
Debugging back to the property setter, we have this:
| Name | Value | Type
-- | -- | -- | --
◢ | sources[0] | {uri="legacy_image" method="" bundleRootPath="ms-appx:///Bundle/" ...} | react::uwp::ImageSource
| ▶ uri | "legacy_image" | std::string
| ▶ method | "" | std::string
| ▶ bundleRootPath | "ms-appx:///Bundle/" | std::string
So, we have an <Image> with a uri of "legacy_image" and that's failing to resolve to a valid resource in our appx package. It looks like we either aren't packaging image resources for RNTester properly, or aren't resolving the URI propertly.
|
non_defect
|
rntester crash loading page the page in rntester is having issues and crashes when you try to load it callstack to the crash react uwp dll winrt to hresult line c react uwp dll winrt terminate line c react uwp dll std experimental coroutine traits promise type unhandled exception line c react uwp dll react uwp reactimage source resumecoro line c react uwp dll react uwp reactimage source initcoro line c react uwp dll react uwp reactimage source react uwp imagesource source line c react uwp dll react uwp imageviewmanager setsource winrt windows ui xaml controls canvas canvas const folly dynamic data line c react uwp dll react uwp imageviewmanager updateproperties react uwp shadownodebase nodetoupdate const folly dynamic reactdiffmap line c react uwp dll react uwp shadownodebase updateproperties const folly dynamic props line c react uwp dll facebook react uimanager createview tag std string classname formal folly dynamic props line c react uwp dll facebook react uimanagermodule getmethods folly dynamic args line c debug spew says exception thrown at kernelbase dll in playground exe winrt originate error legacy image is not a valid absolute uri exception thrown at in playground exe microsoft c exception winrt hresult invalid argument at memory location exception thrown at in playground exe microsoft c exception at memory location debugging back to the property setter we have this name value type ◢ sources uri legacy image method bundlerootpath ms appx bundle react uwp imagesource ▶ uri legacy image std string ▶ method std string ▶ bundlerootpath ms appx bundle std string so we have an with a uri of legacy image and that s failing to resolve to a valid resource in our appx package it looks like we either aren t packaging image resources for rntester properly or aren t resolving the uri propertly
| 0
|
351,853
| 10,524,869,654
|
IssuesEvent
|
2019-09-30 14:05:32
|
kubeapps/kubeapps
|
https://api.github.com/repos/kubeapps/kubeapps
|
closed
|
Improve the base64 decoding command in the documentation
|
good first issue help wanted kind/docs priority/low
|
Very minor tweak. in the docs, rather then this:
-o jsonpath='{.data.token}' | base64 --decode
you can do:
-o go-template='{{.data.token | base64decode}}'
It should work on all platforms as its built into kubectl now.
|
1.0
|
Improve the base64 decoding command in the documentation - Very minor tweak. in the docs, rather then this:
-o jsonpath='{.data.token}' | base64 --decode
you can do:
-o go-template='{{.data.token | base64decode}}'
It should work on all platforms as its built into kubectl now.
|
non_defect
|
improve the decoding command in the documentation very minor tweak in the docs rather then this o jsonpath data token decode you can do o go template data token it should work on all platforms as its built into kubectl now
| 0
|
27,178
| 4,912,212,194
|
IssuesEvent
|
2016-11-23 08:05:00
|
TNGSB/eWallet
|
https://api.github.com/repos/TNGSB/eWallet
|
closed
|
eWallet_MobileApp_The Android version of the App is not available on Playstore for User to download # 47
|
Defect - High (Sev-2)
|
For a new User who wishes to download the app on Play store from an Android phone, the Touch N Go application is not available on Play store to initiate the download
|
1.0
|
eWallet_MobileApp_The Android version of the App is not available on Playstore for User to download # 47 - For a new User who wishes to download the app on Play store from an Android phone, the Touch N Go application is not available on Play store to initiate the download
|
defect
|
ewallet mobileapp the android version of the app is not available on playstore for user to download for a new user who wishes to download the app on play store from an android phone the touch n go application is not available on play store to initiate the download
| 1
|
336,873
| 10,198,499,573
|
IssuesEvent
|
2019-08-13 05:35:22
|
siddhi-io/distribution
|
https://api.github.com/repos/siddhi-io/distribution
|
closed
|
wso2.carbon.server >> port.offset is not user friendly
|
priority/highest severity/major type/improvement
|
**Description:**
$subject as it is only supported with databridge transports. Either MSF4J needs to support this configuration or this configuration needs to be depreciated
**Affected Product Version:**
v5.1.0-m2
|
1.0
|
wso2.carbon.server >> port.offset is not user friendly - **Description:**
$subject as it is only supported with databridge transports. Either MSF4J needs to support this configuration or this configuration needs to be depreciated
**Affected Product Version:**
v5.1.0-m2
|
non_defect
|
carbon server port offset is not user friendly description subject as it is only supported with databridge transports either needs to support this configuration or this configuration needs to be depreciated affected product version
| 0
|
805,886
| 29,735,619,302
|
IssuesEvent
|
2023-06-14 00:26:36
|
space-wizards/space-station-14
|
https://api.github.com/repos/space-wizards/space-station-14
|
closed
|
Remove FTL proximity requirement
|
Priority: 2-Before Release Issue: Feature Request Difficulty: 1-Easy
|
We need a bypass for planets and tbh it's more annoying anyway. The only benefit to it atm is it adds some extra gameplay to cargo but eh.
|
1.0
|
Remove FTL proximity requirement - We need a bypass for planets and tbh it's more annoying anyway. The only benefit to it atm is it adds some extra gameplay to cargo but eh.
|
non_defect
|
remove ftl proximity requirement we need a bypass for planets and tbh it s more annoying anyway the only benefit to it atm is it adds some extra gameplay to cargo but eh
| 0
|
349,769
| 31,828,461,559
|
IssuesEvent
|
2023-09-14 09:03:22
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
App crashes shortly after launch on fresh profile [Android 8.1.0]
|
crash release/blocking priority/P3 QA/Yes release-notes/include regression OS/Android Android 8 - Test/Verification
|
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
App crashes shortly after launch on fresh profile, happening in 1.58.x. Regression as this does not crash in 1.57.x using same steps.
Not reproducible on `Samsung GS 21` | `Android 13`
```
2023-09-13 16:44:33.900 1922-1922/? W/ResourceType: Failure getting entry for 0x7f010d5e (t=0 e=3422) (error -75)
2023-09-13 16:44:33.978 1922-1982/? D/OpenGLRenderer: HWUI GL Pipeline
2023-09-13 16:44:33.980 1922-1922/? D/AndroidRuntime: Shutting down VM
2023-09-13 16:44:33.980 1922-1922/? E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.brave.browser, PID: 1922
java.lang.RuntimeException: Unable to instantiate service com.google.android.datatransport.runtime.scheduling.jobscheduling.JobInfoSchedulerService: java.lang.ClassNotFoundException: Didn't find class "com.google.android.datatransport.runtime.scheduling.jobscheduling.JobInfoSchedulerService" on path: DexPathList[[zip file "/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk"],nativeLibraryDirectories=[/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/lib/arm64, /data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk!/lib/arm64-v8a, /system/lib64]]
at android.app.ActivityThread.handleCreateService(ActivityThread.java:3324)
at android.app.ActivityThread.-wrap4(Unknown Source:0)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1677)
at android.os.Handler.dispatchMessage(Handler.java:106)
at android.os.Looper.loop(Looper.java:164)
at android.app.ActivityThread.main(ActivityThread.java:6494)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:438)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:807)
Caused by: java.lang.ClassNotFoundException: Didn't find class "com.google.android.datatransport.runtime.scheduling.jobscheduling.JobInfoSchedulerService" on path: DexPathList[[zip file "/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk"],nativeLibraryDirectories=[/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/lib/arm64, /data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk!/lib/arm64-v8a, /system/lib64]]
at dalvik.system.BaseDexClassLoader.findClass(BaseDexClassLoader.java:125)
at java.lang.ClassLoader.loadClass(ClassLoader.java:379)
at java.lang.ClassLoader.loadClass(ClassLoader.java:312)
at android.app.ActivityThread.handleCreateService(ActivityThread.java:3321)
at android.app.ActivityThread.-wrap4(Unknown Source:0)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1677)
at android.os.Handler.dispatchMessage(Handler.java:106)
at android.os.Looper.loop(Looper.java:164)
at android.app.ActivityThread.main(ActivityThread.java:6494)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:438)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:807)
```
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. Fresh install
2. Complete onboarding
3. Wait 30 seconds or so
4. Does not matter what user does in this time.
## Actual result <!-- Please add screenshots if needed -->
App crashes after launch
https://github.com/brave/brave-browser/assets/17885425/8e46e890-630a-4345-abcc-084bfc93b96a
## Expected result
Should not crash
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
Easily
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version? No
- Can you reproduce this issue with the current Play Store Beta version? TBD
- Can you reproduce this issue with the current Play Store Nightly version? TBD
## Device details
- Install type (ARM, x86): ARM
- Device type (Phone, Tablet, Phablet): Pixel 2 XL
- Android version: 8.1.0
## Brave version
1.58.122
### Website problems only
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
### Additional information
<!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
|
1.0
|
App crashes shortly after launch on fresh profile [Android 8.1.0] - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description <!-- Provide a brief description of the issue -->
App crashes shortly after launch on fresh profile, happening in 1.58.x. Regression as this does not crash in 1.57.x using same steps.
Not reproducible on `Samsung GS 21` | `Android 13`
```
2023-09-13 16:44:33.900 1922-1922/? W/ResourceType: Failure getting entry for 0x7f010d5e (t=0 e=3422) (error -75)
2023-09-13 16:44:33.978 1922-1982/? D/OpenGLRenderer: HWUI GL Pipeline
2023-09-13 16:44:33.980 1922-1922/? D/AndroidRuntime: Shutting down VM
2023-09-13 16:44:33.980 1922-1922/? E/AndroidRuntime: FATAL EXCEPTION: main
Process: com.brave.browser, PID: 1922
java.lang.RuntimeException: Unable to instantiate service com.google.android.datatransport.runtime.scheduling.jobscheduling.JobInfoSchedulerService: java.lang.ClassNotFoundException: Didn't find class "com.google.android.datatransport.runtime.scheduling.jobscheduling.JobInfoSchedulerService" on path: DexPathList[[zip file "/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk"],nativeLibraryDirectories=[/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/lib/arm64, /data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk!/lib/arm64-v8a, /system/lib64]]
at android.app.ActivityThread.handleCreateService(ActivityThread.java:3324)
at android.app.ActivityThread.-wrap4(Unknown Source:0)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1677)
at android.os.Handler.dispatchMessage(Handler.java:106)
at android.os.Looper.loop(Looper.java:164)
at android.app.ActivityThread.main(ActivityThread.java:6494)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:438)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:807)
Caused by: java.lang.ClassNotFoundException: Didn't find class "com.google.android.datatransport.runtime.scheduling.jobscheduling.JobInfoSchedulerService" on path: DexPathList[[zip file "/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk"],nativeLibraryDirectories=[/data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/lib/arm64, /data/app/com.brave.browser-eVC6uZo8OfRnoW9o2rm_9Q==/base.apk!/lib/arm64-v8a, /system/lib64]]
at dalvik.system.BaseDexClassLoader.findClass(BaseDexClassLoader.java:125)
at java.lang.ClassLoader.loadClass(ClassLoader.java:379)
at java.lang.ClassLoader.loadClass(ClassLoader.java:312)
at android.app.ActivityThread.handleCreateService(ActivityThread.java:3321)
at android.app.ActivityThread.-wrap4(Unknown Source:0)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1677)
at android.os.Handler.dispatchMessage(Handler.java:106)
at android.os.Looper.loop(Looper.java:164)
at android.app.ActivityThread.main(ActivityThread.java:6494)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:438)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:807)
```
## Steps to reproduce <!-- Please add a series of steps to reproduce the issue -->
1. Fresh install
2. Complete onboarding
3. Wait 30 seconds or so
4. Does not matter what user does in this time.
## Actual result <!-- Please add screenshots if needed -->
App crashes after launch
https://github.com/brave/brave-browser/assets/17885425/8e46e890-630a-4345-abcc-084bfc93b96a
## Expected result
Should not crash
## Issue reproduces how often <!-- [Easily reproduced/Intermittent issue/No steps to reproduce] -->
Easily
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current Play Store version? No
- Can you reproduce this issue with the current Play Store Beta version? TBD
- Can you reproduce this issue with the current Play Store Nightly version? TBD
## Device details
- Install type (ARM, x86): ARM
- Device type (Phone, Tablet, Phablet): Pixel 2 XL
- Android version: 8.1.0
## Brave version
1.58.122
### Website problems only
- Does the issue resolve itself when disabling Brave Shields?
- Does the issue resolve itself when disabling Brave Rewards?
- Is the issue reproducible on the latest version of Chrome?
### Additional information
<!-- Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue -->
|
non_defect
|
app crashes shortly after launch on fresh profile have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description app crashes shortly after launch on fresh profile happening in x regression as this does not crash in x using same steps not reproducible on samsung gs android w resourcetype failure getting entry for t e error d openglrenderer hwui gl pipeline d androidruntime shutting down vm e androidruntime fatal exception main process com brave browser pid java lang runtimeexception unable to instantiate service com google android datatransport runtime scheduling jobscheduling jobinfoschedulerservice java lang classnotfoundexception didn t find class com google android datatransport runtime scheduling jobscheduling jobinfoschedulerservice on path dexpathlist nativelibrarydirectories at android app activitythread handlecreateservice activitythread java at android app activitythread unknown source at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java caused by java lang classnotfoundexception didn t find class com google android datatransport runtime scheduling jobscheduling jobinfoschedulerservice on path dexpathlist nativelibrarydirectories at dalvik system basedexclassloader findclass basedexclassloader java at java lang classloader loadclass classloader java at java lang classloader loadclass classloader java at android app activitythread handlecreateservice activitythread java at android app activitythread unknown source at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java steps to reproduce fresh install complete onboarding wait seconds or so does not matter what user does in this time actual result app crashes after launch expected result should not crash issue reproduces how often easily version channel information can you reproduce this issue with the current play store version no can you reproduce this issue with the current play store beta version tbd can you reproduce this issue with the current play store nightly version tbd device details install type arm arm device type phone tablet phablet pixel xl android version brave version website problems only does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome additional information
| 0
|
101,869
| 16,529,728,742
|
IssuesEvent
|
2021-05-27 03:11:24
|
alpersonalwebsite/postME
|
https://api.github.com/repos/alpersonalwebsite/postME
|
opened
|
CVE-2020-7788 (High) detected in ini-1.3.5.tgz
|
security vulnerability
|
## CVE-2020-7788 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ini-1.3.5.tgz</b></p></summary>
<p>An ini encoder/decoder for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.5.tgz">https://registry.npmjs.org/ini/-/ini-1.3.5.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/ini</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- global-modules-2.0.0.tgz
- global-prefix-3.0.0.tgz
- :x: **ini-1.3.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788>CVE-2020-7788</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: v1.3.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7788 (High) detected in ini-1.3.5.tgz - ## CVE-2020-7788 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ini-1.3.5.tgz</b></p></summary>
<p>An ini encoder/decoder for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/ini/-/ini-1.3.5.tgz">https://registry.npmjs.org/ini/-/ini-1.3.5.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/ini</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- global-modules-2.0.0.tgz
- global-prefix-3.0.0.tgz
- :x: **ini-1.3.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package ini before 1.3.6. If an attacker submits a malicious INI file to an application that parses it with ini.parse, they will pollute the prototype on the application. This can be exploited further depending on the context.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788>CVE-2020-7788</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7788</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: v1.3.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in ini tgz cve high severity vulnerability vulnerable library ini tgz an ini encoder decoder for node library home page a href path to dependency file postme frontend package json path to vulnerable library postme frontend node modules ini dependency hierarchy react scripts tgz root library react dev utils tgz global modules tgz global prefix tgz x ini tgz vulnerable library found in head commit a href vulnerability details this affects the package ini before if an attacker submits a malicious ini file to an application that parses it with ini parse they will pollute the prototype on the application this can be exploited further depending on the context publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
28,568
| 5,292,734,184
|
IssuesEvent
|
2017-02-09 03:36:29
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
Cache.viewPrefix problem - always refreshes cache when you use Cache.viewPrefix in controller
|
Defect
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 2.9.4, 2.9.5
* Platform and Target: WIN10, Linux Ubuntu 16LTS
### What you did
I am trying to switch on cache on views (I have two themes: default and mobile and want to switch between them).
### What happened
In controller after I set: Configure::write('Cache.viewPrefix', 'mobile'); I see that the view cache is always regenerated (see timestamp and cachetime inside the file). If I remove Configure::write('Cache.viewPrefix', 'mobile'); then the cache is stable.
```php
<?php
App::uses('AppController', 'Controller');
class RomansController extends AppController {
public $helpers = array('Cache');
public $cacheAction = array(
'index' => 86400*365
);
public function index() {
$this->layout = 'ajax';
// Configure::write('Cache.viewPrefix', 'mobile');
}
}
```
View:
```php
<?php
echo date('Y-m-d H:i:s');
?>
```
### What you expected to happen
the cache will not be regenerated if I set Cache.viewPrefix dynamically.
|
1.0
|
Cache.viewPrefix problem - always refreshes cache when you use Cache.viewPrefix in controller - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 2.9.4, 2.9.5
* Platform and Target: WIN10, Linux Ubuntu 16LTS
### What you did
I am trying to switch on cache on views (I have two themes: default and mobile and want to switch between them).
### What happened
In controller after I set: Configure::write('Cache.viewPrefix', 'mobile'); I see that the view cache is always regenerated (see timestamp and cachetime inside the file). If I remove Configure::write('Cache.viewPrefix', 'mobile'); then the cache is stable.
```php
<?php
App::uses('AppController', 'Controller');
class RomansController extends AppController {
public $helpers = array('Cache');
public $cacheAction = array(
'index' => 86400*365
);
public function index() {
$this->layout = 'ajax';
// Configure::write('Cache.viewPrefix', 'mobile');
}
}
```
View:
```php
<?php
echo date('Y-m-d H:i:s');
?>
```
### What you expected to happen
the cache will not be regenerated if I set Cache.viewPrefix dynamically.
|
defect
|
cache viewprefix problem always refreshes cache when you use cache viewprefix in controller this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target linux ubuntu what you did i am trying to switch on cache on views i have two themes default and mobile and want to switch between them what happened in controller after i set configure write cache viewprefix mobile i see that the view cache is always regenerated see timestamp and cachetime inside the file if i remove configure write cache viewprefix mobile then the cache is stable php php app uses appcontroller controller class romanscontroller extends appcontroller public helpers array cache public cacheaction array index public function index this layout ajax configure write cache viewprefix mobile view php php echo date y m d h i s what you expected to happen the cache will not be regenerated if i set cache viewprefix dynamically
| 1
|
30,965
| 6,378,708,843
|
IssuesEvent
|
2017-08-02 13:20:46
|
sous-chefs/ruby_rbenv
|
https://api.github.com/repos/sous-chefs/ruby_rbenv
|
closed
|
Unable to find 'gem' binary in path
|
Defect Pending Feedback
|
### Cookbook version
Latest I believe, not 100% sure how to check.
### Chef-client version
13.0.118
### Platform Details
Vagrant 1.9.3, Windows 10, Hyper V
### Scenario:
Trying to run Vagrant Up
### Steps to Reproduce:
Running this vagrant file:
```
Vagrant.configure("2") do |config|
# The most common configuration options are documented and commented below.
# For a complete reference, please see the online documentation at
# https://docs.vagrantup.com.
# Set the provider order.
config.vm.provider "hyperv"
config.vm.provider "virtualbox"
# Choose Box
config.vm.box = "remyjette/wily64"
config.vm.boot_timeout = 2000
config.vm.network :private_network, ip: "192.168.2.45"
# Configure ports.
config.vm.network :forwarded_port, guest: 5000, host: 5000 # Rails Web Server
config.vm.network :forwarded_port, guest: 5432, host: 5432 # Postgres
# config.vm.synced_folder "../data", "/vagrant_data"
config.vm.synced_folder ".", "/vagrant", type: "smb", mount_options: ["mfsymlinks"]
# Setup box details.
config.vm.provider :hyperv do |hv|
hv.memory = 2048
hv.cpus = 2
end
# Run chef to configure software.
config.vm.provision :chef_solo do |chef|
chef.cookbooks_path = ["cookbooks"]
# Ruby
chef.add_recipe "ruby_build"
chef.add_recipe "ruby_rbenv::user"
# Install Ruby 2.3.3 and Bundler
chef.json = {
# Configure Ruby.
rbenv: {
user_installs: [{
user: 'vagrant',
rubies: ["2.3.3"],
global: "2.3.3",
gems: {
'2.3.3' => [
{ name: 'bundler'},
{ name: 'rails' },
{ name: 'rake' },
{ name: 'taps' },
{ name: 'puma' },
],
}
}]
},
}
end
end
```
### Expected Result:
Installation of ruby 2.3.3
### Actual Result:
System errors with:
```
==> default: - change owner from '' to 'vagrant'
==> default: * bash[Initialize rbenv (vagrant)] action run
==> default:
==> default: [execute]
==> default: export PATH="/home/vagrant/.rbenv/shims:${PATH}"
==> default: source "/home/vagrant/.rbenv/libexec/../completions/rbenv.bash"
==> default: rbenv rehash 2>/dev/null
==> default: rbenv() {
==> default: typeset command
==> default: command="$1"
==> default: if [ "$#" -gt 0 ]; then
==> default: shift
==> default: fi
==> default:
==> default: case "$command" in
==> default: rehash|shell)
==> default: eval `rbenv "sh-$command" "$@"`;;
==> default: *)
==> default: command rbenv "$command" "$@";;
==> default: esac
==> default: }
==> default: [2017-05-04T04:08:36-04:00] INFO: bash[Initialize rbenv (vagrant)] ran successfully
==> default: - execute "bash" "/tmp/chef-script20170504-3292-4pz37q"
==> default: * log[rbenv-post-init-vagrant] action write
==> default: [2017-05-04T04:08:36-04:00] INFO: rbenv-post-init-vagrant
==> default:
==> default: * ruby_block[Add rbenv to PATH] action run
==> default: [2017-05-04T04:08:36-04:00] INFO: ruby_block[Add rbenv to PATH] called
==> default:
==> default: - execute the ruby block Add rbenv to PATH
==> default: Recipe: ruby_rbenv::user
==> default: * ruby_rbenv_ruby[2.3.3 (vagrant)] (vagrant) action install
==> default:
==> default: * apt_package[autoconf] action install
==> default: [2017-05-04T04:08:45-04:00] INFO: apt_package[autoconf] installed autoconf at 2.69-8
==> default: - install version 2.69-8 of package autoconf
==> default: * apt_package[bison] action install
==> default: [2017-05-04T04:08:51-04:00] INFO: apt_package[bison] installed bison at 2:3.0.2.dfsg-2
==> default:
==> default: - install version 2:3.0.2.dfsg-2 of package bison
==> default: * apt_package[build-essential] action install
==> default: [2017-05-04T04:11:23-04:00] INFO: apt_package[build-essential] installed build-essential at 12.1ubuntu2
==> default:
==> default: - install version 12.1ubuntu2 of package build-essential
==> default: * apt_package[libssl-dev] action install
==> default: [2017-05-04T04:11:44-04:00] INFO: apt_package[libssl-dev] installed libssl-dev at 1.0.2d-0ubuntu1.5
==> default:
==> default: - install version 1.0.2d-0ubuntu1.5 of package libssl-dev
==> default: * apt_package[libyaml-dev] action install
==> default: [2017-05-04T04:11:49-04:00] INFO: apt_package[libyaml-dev] installed libyaml-dev at 0.1.6-3
==> default:
==> default: - install version 0.1.6-3 of package libyaml-dev
==> default: * apt_package[libreadline6-dev] action install
==> default: [2017-05-04T04:11:53-04:00] INFO: apt_package[libreadline6-dev] installed libreadline6-dev at 6.3-8ubuntu1
==> default: - install version 6.3-8ubuntu1 of package libreadline6-dev
==> default: * apt_package[zlib1g-dev] action install
==> default: (up to date)
==> default: * apt_package[libsqlite3-dev] action install
==> default: [2017-05-04T04:11:57-04:00] INFO: apt_package[libsqlite3-dev] installed libsqlite3-dev at 3.8.11.1-1
==> default:
==> default: - install version 3.8.11.1-1 of package libsqlite3-dev
==> default: * apt_package[libxml2-dev] action install
==> default: [2017-05-04T04:12:28-04:00] INFO: apt_package[libxml2-dev] installed libxml2-dev at 2.9.2+zdfsg1-4ubuntu0.4
==> default:
==> default:
==> default: - install version 2.9.2+zdfsg1-4ubuntu0.4 of package libxml2-dev
==> default: * apt_package[libxslt1-dev] action install
==> default: [2017-05-04T04:12:34-04:00] INFO: apt_package[libxslt1-dev] installed libxslt1-dev at 1.1.28-2build2
==> default:
==> default: - install version 1.1.28-2build2 of package libxslt1-dev
==> default: * apt_package[libc6-dev] action install
==> default: (up to date)
==> default: * apt_package[libffi-dev] action install
==> default: [2017-05-04T04:12:38-04:00] INFO: apt_package[libffi-dev] installed libffi-dev at 3.2.1-3
==> default:
==> default: - install version 3.2.1-3 of package libffi-dev
==> default: * apt_package[libgdbm3] action install
==> default: (up to date)
==> default: * apt_package[libgdbm-dev] action install
==> default: [2017-05-04T04:12:41-04:00] INFO: apt_package[libgdbm-dev] installed libgdbm-dev at 1.8.3-13.1
==> default: - install version 1.8.3-13.1 of package libgdbm-dev
==> default: [2017-05-04T04:12:41-04:00] INFO: Building ruby_rbenv_ruby[2.3.3 (vagrant)] (vagrant), this could take a while...
==> default: * ruby_rbenv_script[rbenv install 2.3.3 (vagrant)] action run
==> default:
==> default: * script[rbenv install 2.3.3 (vagrant)] action run
==> default:
==> default: [execute] Downloading ruby-2.3.3.tar.bz2...
==> default: -> https://cache.ruby-lang.org/pub/ruby/2.3/ruby-2.3.3.tar.bz2
==> default: Installing ruby-2.3.3...
==> default: Installed ruby-2.3.3 to /home/vagrant/.rbenv/versions/2.3.3
==> default:
==> default: [2017-05-04T04:17:50-04:00] INFO: script[rbenv install 2.3.3 (vagrant)] ran successfully
==> default: - execute "bash" "/tmp/chef-script20170504-3292-guw8rd"
==> default:
==> default: * apt_package[autoconf] action nothing (skipped due to action :nothing)
==> default: * apt_package[bison] action nothing (skipped due to action :nothing)
==> default: * apt_package[build-essential] action nothing
==> default: (skipped due to action :nothing)
==> default: * apt_package[libssl-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libyaml-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libreadline6-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[zlib1g-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libsqlite3-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libxml2-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libxslt1-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libc6-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libffi-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libgdbm3] action nothing (skipped due to action :nothing)
==> default: * apt_package[libgdbm-dev] action nothing (skipped due to action :nothing)
==> default: * ruby_rbenv_script[rbenv install 2.3.3 (vagrant)] action nothing (skipped due to action :nothing)
==> default:
==> default: * ruby_rbenv_global[2.3.3 (vagrant)] (vagrant) action create[2017-05-04T04:17:50-04:00] WARN: only_if block for ruby_rbenv_global[2.3.3 (vagrant)] (vagrant) returned "2.3.3", did you mean to run a command? If so use 'only_if "2.3.3"' in your code.
==> default:
==> default: * ruby_rbenv_script[rbenv global 2.3.3 (vagrant)] action run
==> default: * script[rbenv global 2.3.3 (vagrant)] action run
==> default: [2017-05-04T04:17:50-04:00] INFO: script[rbenv global 2.3.3 (vagrant)] ran successfully
==> default:
==> default: - execute "bash" "/tmp/chef-script20170504-3292-nu4jh"
==> default:
==> default:
==> default: * [2.3.3::bundler (vagrant)] (vagrant) action install
==> default:
==> default: ================================================================================
==> default: Error executing action `install` on resource '[2.3.3::bundler (vagrant)] (vagrant)'
==> default: ================================================================================
==> default:
==> default: Chef::Exceptions::FileNotFound
==> default: ------------------------------
==> default: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
==> default:
==> default: Cookbook Trace:
==> default: ---------------
==> default: /var/chef/cache/cookbooks/ruby_rbenv/libraries/chef_provider_package_rbenvrubygems.rb:57:in `initialize'
==> default:
==> default: Resource Declaration:
==> default: ---------------------
==> default: # In /var/chef/cache/cookbooks/ruby_rbenv/recipes/user.rb
==> default:
==> default: 69: rbenv_gem "#{gem['name']} (#{rbenv_user['user']})" do
==> default: 70: package_name gem['name']
==> default: 71: user rbenv_user['user']
==> default: 72: root_path rbenv_user['root_path'] if rbenv_user['root_path']
==> default: 73: rbenv_version rubie
==> default: 74:
==> default: 75: %w(version action options source).each do |attr|
==> default: 76: send(attr, gem[attr]) if gem[attr]
==> default: 77: end
==> default: 78: end
==> default: 79: end
==> default:
==> default: Compiled Resource:
==> default: ------------------
==> default: # Declared in /var/chef/cache/cookbooks/ruby_rbenv/recipes/user.rb:69:in `block (3 levels) in from_file'
==> default:
==> default: ruby_rbenv_gem("bundler (vagrant)") do
==> default: provider Chef::Provider::Package::RbenvRubygems
==> default: action [:install]
==> default: default_guard_interpreter :default
==> default: declared_type :rbenv_gem
==> default: cookbook_name "ruby_rbenv"
==> default: recipe_name "user"
==> default: package_name "bundler"
==> default: user "vagrant"
==> default: rbenv_version "2.3.3"
==> default: end
==> default:
==> default: System Info:
==> default: ------------
==> default: chef_version=13.0.118
==> default: platform=ubuntu
==> default: platform_version=15.10
==> default: ruby=ruby 2.4.1p111 (2017-03-22 revision 58053) [x86_64-linux]
==> default: program_name=chef-solo worker: ppid=3286;start=04:08:13;
==> default: executable=/opt/chef/bin/chef-solo
==> default:
==> default: [2017-05-04T04:17:50-04:00] INFO: Running queued delayed notifications before re-raising exception
==> default: Running handlers:
==> default: [2017-05-04T04:17:50-04:00] ERROR: Running exception handlers
==> default: Running handlers complete
==> default: [2017-05-04T04:17:50-04:00] ERROR: Exception handlers complete
==> default: Chef Client failed. 26 resources updated in 09 minutes 37 seconds
==> default: [2017-05-04T04:17:50-04:00] FATAL: Stacktrace dumped to /var/chef/cache/chef-stacktrace.out
==> default: [2017-05-04T04:17:50-04:00] FATAL: Please provide the contents of the stacktrace.out file if you file a bug report
==> default: [2017-05-04T04:17:50-04:00] ERROR: [2.3.3::bundler (vagrant)] (vagrant) (ruby_rbenv::user line 69) had an error: Chef::Exceptions::FileNotFound: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
==> default: [2017-05-04T04:17:50-04:00] FATAL: Chef::Exceptions::ChildConvergeError: Chef run process exited unsuccessfully (exit code 1)
Chef never successfully completed! Any errors should be visible in the
output above. Please fix your recipes so that they properly complete.
```
Stack out:
```
Generated at 2017-05-04 04:17:50 -0400
Chef::Exceptions::FileNotFound: [2.3.3::bundler (vagrant)] (vagrant) (ruby_rbenv::user line 69) had an error: Chef::Exceptions::FileNotFound: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:416:in `find_gem_by_path'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:387:in `initialize'
/var/chef/cache/cookbooks/ruby_rbenv/libraries/chef_provider_package_rbenvrubygems.rb:57:in `initialize'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `new'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `provider_for_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:591:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:69:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block (2 levels) in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `each'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:94:in `block in execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:114:in `call_iterator_block'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:85:in `step'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:103:in `iterate'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:55:in `each_with_index'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:92:in `execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:96:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:715:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `catch'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:749:in `converge_and_save'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:286:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:291:in `block in fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:244:in `block in run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/local_mode.rb:44:in `with_server_connectivity'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:232:in `run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:464:in `sleep_then_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:451:in `block in interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `loop'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:434:in `run_application'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:59:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/solo.rb:225:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/bin/chef-solo:25:in `<top (required)>'
/usr/bin/chef-solo:58:in `load'
/usr/bin/chef-solo:58:in `<main>'
>>>> Caused by Chef::Exceptions::FileNotFound: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:416:in `find_gem_by_path'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:387:in `initialize'
/var/chef/cache/cookbooks/ruby_rbenv/libraries/chef_provider_package_rbenvrubygems.rb:57:in `initialize'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `new'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `provider_for_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:591:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:69:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block (2 levels) in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `each'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:94:in `block in execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:114:in `call_iterator_block'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:85:in `step'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:103:in `iterate'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:55:in `each_with_index'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:92:in `execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:96:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:715:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `catch'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:749:in `converge_and_save'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:286:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:291:in `block in fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:244:in `block in run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/local_mode.rb:44:in `with_server_connectivity'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:232:in `run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:464:in `sleep_then_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:451:in `block in interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `loop'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:434:in `run_application'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:59:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/solo.rb:225:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/bin/chef-solo:25:in `<top (required)>'
/usr/bin/chef-solo:58:in `load'
/usr/bin/chef-solo:58:in `<main>'
```
|
1.0
|
Unable to find 'gem' binary in path - ### Cookbook version
Latest I believe, not 100% sure how to check.
### Chef-client version
13.0.118
### Platform Details
Vagrant 1.9.3, Windows 10, Hyper V
### Scenario:
Trying to run Vagrant Up
### Steps to Reproduce:
Running this vagrant file:
```
Vagrant.configure("2") do |config|
# The most common configuration options are documented and commented below.
# For a complete reference, please see the online documentation at
# https://docs.vagrantup.com.
# Set the provider order.
config.vm.provider "hyperv"
config.vm.provider "virtualbox"
# Choose Box
config.vm.box = "remyjette/wily64"
config.vm.boot_timeout = 2000
config.vm.network :private_network, ip: "192.168.2.45"
# Configure ports.
config.vm.network :forwarded_port, guest: 5000, host: 5000 # Rails Web Server
config.vm.network :forwarded_port, guest: 5432, host: 5432 # Postgres
# config.vm.synced_folder "../data", "/vagrant_data"
config.vm.synced_folder ".", "/vagrant", type: "smb", mount_options: ["mfsymlinks"]
# Setup box details.
config.vm.provider :hyperv do |hv|
hv.memory = 2048
hv.cpus = 2
end
# Run chef to configure software.
config.vm.provision :chef_solo do |chef|
chef.cookbooks_path = ["cookbooks"]
# Ruby
chef.add_recipe "ruby_build"
chef.add_recipe "ruby_rbenv::user"
# Install Ruby 2.3.3 and Bundler
chef.json = {
# Configure Ruby.
rbenv: {
user_installs: [{
user: 'vagrant',
rubies: ["2.3.3"],
global: "2.3.3",
gems: {
'2.3.3' => [
{ name: 'bundler'},
{ name: 'rails' },
{ name: 'rake' },
{ name: 'taps' },
{ name: 'puma' },
],
}
}]
},
}
end
end
```
### Expected Result:
Installation of ruby 2.3.3
### Actual Result:
System errors with:
```
==> default: - change owner from '' to 'vagrant'
==> default: * bash[Initialize rbenv (vagrant)] action run
==> default:
==> default: [execute]
==> default: export PATH="/home/vagrant/.rbenv/shims:${PATH}"
==> default: source "/home/vagrant/.rbenv/libexec/../completions/rbenv.bash"
==> default: rbenv rehash 2>/dev/null
==> default: rbenv() {
==> default: typeset command
==> default: command="$1"
==> default: if [ "$#" -gt 0 ]; then
==> default: shift
==> default: fi
==> default:
==> default: case "$command" in
==> default: rehash|shell)
==> default: eval `rbenv "sh-$command" "$@"`;;
==> default: *)
==> default: command rbenv "$command" "$@";;
==> default: esac
==> default: }
==> default: [2017-05-04T04:08:36-04:00] INFO: bash[Initialize rbenv (vagrant)] ran successfully
==> default: - execute "bash" "/tmp/chef-script20170504-3292-4pz37q"
==> default: * log[rbenv-post-init-vagrant] action write
==> default: [2017-05-04T04:08:36-04:00] INFO: rbenv-post-init-vagrant
==> default:
==> default: * ruby_block[Add rbenv to PATH] action run
==> default: [2017-05-04T04:08:36-04:00] INFO: ruby_block[Add rbenv to PATH] called
==> default:
==> default: - execute the ruby block Add rbenv to PATH
==> default: Recipe: ruby_rbenv::user
==> default: * ruby_rbenv_ruby[2.3.3 (vagrant)] (vagrant) action install
==> default:
==> default: * apt_package[autoconf] action install
==> default: [2017-05-04T04:08:45-04:00] INFO: apt_package[autoconf] installed autoconf at 2.69-8
==> default: - install version 2.69-8 of package autoconf
==> default: * apt_package[bison] action install
==> default: [2017-05-04T04:08:51-04:00] INFO: apt_package[bison] installed bison at 2:3.0.2.dfsg-2
==> default:
==> default: - install version 2:3.0.2.dfsg-2 of package bison
==> default: * apt_package[build-essential] action install
==> default: [2017-05-04T04:11:23-04:00] INFO: apt_package[build-essential] installed build-essential at 12.1ubuntu2
==> default:
==> default: - install version 12.1ubuntu2 of package build-essential
==> default: * apt_package[libssl-dev] action install
==> default: [2017-05-04T04:11:44-04:00] INFO: apt_package[libssl-dev] installed libssl-dev at 1.0.2d-0ubuntu1.5
==> default:
==> default: - install version 1.0.2d-0ubuntu1.5 of package libssl-dev
==> default: * apt_package[libyaml-dev] action install
==> default: [2017-05-04T04:11:49-04:00] INFO: apt_package[libyaml-dev] installed libyaml-dev at 0.1.6-3
==> default:
==> default: - install version 0.1.6-3 of package libyaml-dev
==> default: * apt_package[libreadline6-dev] action install
==> default: [2017-05-04T04:11:53-04:00] INFO: apt_package[libreadline6-dev] installed libreadline6-dev at 6.3-8ubuntu1
==> default: - install version 6.3-8ubuntu1 of package libreadline6-dev
==> default: * apt_package[zlib1g-dev] action install
==> default: (up to date)
==> default: * apt_package[libsqlite3-dev] action install
==> default: [2017-05-04T04:11:57-04:00] INFO: apt_package[libsqlite3-dev] installed libsqlite3-dev at 3.8.11.1-1
==> default:
==> default: - install version 3.8.11.1-1 of package libsqlite3-dev
==> default: * apt_package[libxml2-dev] action install
==> default: [2017-05-04T04:12:28-04:00] INFO: apt_package[libxml2-dev] installed libxml2-dev at 2.9.2+zdfsg1-4ubuntu0.4
==> default:
==> default:
==> default: - install version 2.9.2+zdfsg1-4ubuntu0.4 of package libxml2-dev
==> default: * apt_package[libxslt1-dev] action install
==> default: [2017-05-04T04:12:34-04:00] INFO: apt_package[libxslt1-dev] installed libxslt1-dev at 1.1.28-2build2
==> default:
==> default: - install version 1.1.28-2build2 of package libxslt1-dev
==> default: * apt_package[libc6-dev] action install
==> default: (up to date)
==> default: * apt_package[libffi-dev] action install
==> default: [2017-05-04T04:12:38-04:00] INFO: apt_package[libffi-dev] installed libffi-dev at 3.2.1-3
==> default:
==> default: - install version 3.2.1-3 of package libffi-dev
==> default: * apt_package[libgdbm3] action install
==> default: (up to date)
==> default: * apt_package[libgdbm-dev] action install
==> default: [2017-05-04T04:12:41-04:00] INFO: apt_package[libgdbm-dev] installed libgdbm-dev at 1.8.3-13.1
==> default: - install version 1.8.3-13.1 of package libgdbm-dev
==> default: [2017-05-04T04:12:41-04:00] INFO: Building ruby_rbenv_ruby[2.3.3 (vagrant)] (vagrant), this could take a while...
==> default: * ruby_rbenv_script[rbenv install 2.3.3 (vagrant)] action run
==> default:
==> default: * script[rbenv install 2.3.3 (vagrant)] action run
==> default:
==> default: [execute] Downloading ruby-2.3.3.tar.bz2...
==> default: -> https://cache.ruby-lang.org/pub/ruby/2.3/ruby-2.3.3.tar.bz2
==> default: Installing ruby-2.3.3...
==> default: Installed ruby-2.3.3 to /home/vagrant/.rbenv/versions/2.3.3
==> default:
==> default: [2017-05-04T04:17:50-04:00] INFO: script[rbenv install 2.3.3 (vagrant)] ran successfully
==> default: - execute "bash" "/tmp/chef-script20170504-3292-guw8rd"
==> default:
==> default: * apt_package[autoconf] action nothing (skipped due to action :nothing)
==> default: * apt_package[bison] action nothing (skipped due to action :nothing)
==> default: * apt_package[build-essential] action nothing
==> default: (skipped due to action :nothing)
==> default: * apt_package[libssl-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libyaml-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libreadline6-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[zlib1g-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libsqlite3-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libxml2-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libxslt1-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libc6-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libffi-dev] action nothing (skipped due to action :nothing)
==> default: * apt_package[libgdbm3] action nothing (skipped due to action :nothing)
==> default: * apt_package[libgdbm-dev] action nothing (skipped due to action :nothing)
==> default: * ruby_rbenv_script[rbenv install 2.3.3 (vagrant)] action nothing (skipped due to action :nothing)
==> default:
==> default: * ruby_rbenv_global[2.3.3 (vagrant)] (vagrant) action create[2017-05-04T04:17:50-04:00] WARN: only_if block for ruby_rbenv_global[2.3.3 (vagrant)] (vagrant) returned "2.3.3", did you mean to run a command? If so use 'only_if "2.3.3"' in your code.
==> default:
==> default: * ruby_rbenv_script[rbenv global 2.3.3 (vagrant)] action run
==> default: * script[rbenv global 2.3.3 (vagrant)] action run
==> default: [2017-05-04T04:17:50-04:00] INFO: script[rbenv global 2.3.3 (vagrant)] ran successfully
==> default:
==> default: - execute "bash" "/tmp/chef-script20170504-3292-nu4jh"
==> default:
==> default:
==> default: * [2.3.3::bundler (vagrant)] (vagrant) action install
==> default:
==> default: ================================================================================
==> default: Error executing action `install` on resource '[2.3.3::bundler (vagrant)] (vagrant)'
==> default: ================================================================================
==> default:
==> default: Chef::Exceptions::FileNotFound
==> default: ------------------------------
==> default: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
==> default:
==> default: Cookbook Trace:
==> default: ---------------
==> default: /var/chef/cache/cookbooks/ruby_rbenv/libraries/chef_provider_package_rbenvrubygems.rb:57:in `initialize'
==> default:
==> default: Resource Declaration:
==> default: ---------------------
==> default: # In /var/chef/cache/cookbooks/ruby_rbenv/recipes/user.rb
==> default:
==> default: 69: rbenv_gem "#{gem['name']} (#{rbenv_user['user']})" do
==> default: 70: package_name gem['name']
==> default: 71: user rbenv_user['user']
==> default: 72: root_path rbenv_user['root_path'] if rbenv_user['root_path']
==> default: 73: rbenv_version rubie
==> default: 74:
==> default: 75: %w(version action options source).each do |attr|
==> default: 76: send(attr, gem[attr]) if gem[attr]
==> default: 77: end
==> default: 78: end
==> default: 79: end
==> default:
==> default: Compiled Resource:
==> default: ------------------
==> default: # Declared in /var/chef/cache/cookbooks/ruby_rbenv/recipes/user.rb:69:in `block (3 levels) in from_file'
==> default:
==> default: ruby_rbenv_gem("bundler (vagrant)") do
==> default: provider Chef::Provider::Package::RbenvRubygems
==> default: action [:install]
==> default: default_guard_interpreter :default
==> default: declared_type :rbenv_gem
==> default: cookbook_name "ruby_rbenv"
==> default: recipe_name "user"
==> default: package_name "bundler"
==> default: user "vagrant"
==> default: rbenv_version "2.3.3"
==> default: end
==> default:
==> default: System Info:
==> default: ------------
==> default: chef_version=13.0.118
==> default: platform=ubuntu
==> default: platform_version=15.10
==> default: ruby=ruby 2.4.1p111 (2017-03-22 revision 58053) [x86_64-linux]
==> default: program_name=chef-solo worker: ppid=3286;start=04:08:13;
==> default: executable=/opt/chef/bin/chef-solo
==> default:
==> default: [2017-05-04T04:17:50-04:00] INFO: Running queued delayed notifications before re-raising exception
==> default: Running handlers:
==> default: [2017-05-04T04:17:50-04:00] ERROR: Running exception handlers
==> default: Running handlers complete
==> default: [2017-05-04T04:17:50-04:00] ERROR: Exception handlers complete
==> default: Chef Client failed. 26 resources updated in 09 minutes 37 seconds
==> default: [2017-05-04T04:17:50-04:00] FATAL: Stacktrace dumped to /var/chef/cache/chef-stacktrace.out
==> default: [2017-05-04T04:17:50-04:00] FATAL: Please provide the contents of the stacktrace.out file if you file a bug report
==> default: [2017-05-04T04:17:50-04:00] ERROR: [2.3.3::bundler (vagrant)] (vagrant) (ruby_rbenv::user line 69) had an error: Chef::Exceptions::FileNotFound: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
==> default: [2017-05-04T04:17:50-04:00] FATAL: Chef::Exceptions::ChildConvergeError: Chef run process exited unsuccessfully (exit code 1)
Chef never successfully completed! Any errors should be visible in the
output above. Please fix your recipes so that they properly complete.
```
Stack out:
```
Generated at 2017-05-04 04:17:50 -0400
Chef::Exceptions::FileNotFound: [2.3.3::bundler (vagrant)] (vagrant) (ruby_rbenv::user line 69) had an error: Chef::Exceptions::FileNotFound: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:416:in `find_gem_by_path'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:387:in `initialize'
/var/chef/cache/cookbooks/ruby_rbenv/libraries/chef_provider_package_rbenvrubygems.rb:57:in `initialize'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `new'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `provider_for_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:591:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:69:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block (2 levels) in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `each'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:94:in `block in execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:114:in `call_iterator_block'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:85:in `step'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:103:in `iterate'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:55:in `each_with_index'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:92:in `execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:96:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:715:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `catch'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:749:in `converge_and_save'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:286:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:291:in `block in fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:244:in `block in run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/local_mode.rb:44:in `with_server_connectivity'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:232:in `run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:464:in `sleep_then_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:451:in `block in interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `loop'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:434:in `run_application'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:59:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/solo.rb:225:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/bin/chef-solo:25:in `<top (required)>'
/usr/bin/chef-solo:58:in `load'
/usr/bin/chef-solo:58:in `<main>'
>>>> Caused by Chef::Exceptions::FileNotFound: Unable to find 'gem' binary in path: /usr/local/rbenv/shims:/usr/local/rbenv/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:416:in `find_gem_by_path'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/provider/package/rubygems.rb:387:in `initialize'
/var/chef/cache/cookbooks/ruby_rbenv/libraries/chef_provider_package_rbenvrubygems.rb:57:in `initialize'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `new'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:1399:in `provider_for_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource.rb:591:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:69:in `run_action'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block (2 levels) in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `each'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:97:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:94:in `block in execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:114:in `call_iterator_block'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:85:in `step'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:103:in `iterate'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/stepable_iterator.rb:55:in `each_with_index'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/resource_collection/resource_list.rb:92:in `execute_each_resource'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/runner.rb:96:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:715:in `block in converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `catch'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:710:in `converge'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:749:in `converge_and_save'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/client.rb:286:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:291:in `block in fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:279:in `fork_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:244:in `block in run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/local_mode.rb:44:in `with_server_connectivity'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:232:in `run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:464:in `sleep_then_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:451:in `block in interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `loop'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:450:in `interval_run_chef_client'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/client.rb:434:in `run_application'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application.rb:59:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/lib/chef/application/solo.rb:225:in `run'
/opt/chef/embedded/lib/ruby/gems/2.4.0/gems/chef-13.0.118/bin/chef-solo:25:in `<top (required)>'
/usr/bin/chef-solo:58:in `load'
/usr/bin/chef-solo:58:in `<main>'
```
|
defect
|
unable to find gem binary in path cookbook version latest i believe not sure how to check chef client version platform details vagrant windows hyper v scenario trying to run vagrant up steps to reproduce running this vagrant file vagrant configure do config the most common configuration options are documented and commented below for a complete reference please see the online documentation at set the provider order config vm provider hyperv config vm provider virtualbox choose box config vm box remyjette config vm boot timeout config vm network private network ip configure ports config vm network forwarded port guest host rails web server config vm network forwarded port guest host postgres config vm synced folder data vagrant data config vm synced folder vagrant type smb mount options setup box details config vm provider hyperv do hv hv memory hv cpus end run chef to configure software config vm provision chef solo do chef chef cookbooks path ruby chef add recipe ruby build chef add recipe ruby rbenv user install ruby and bundler chef json configure ruby rbenv user installs user vagrant rubies global gems name bundler name rails name rake name taps name puma end end expected result installation of ruby actual result system errors with default change owner from to vagrant default bash action run default default default export path home vagrant rbenv shims path default source home vagrant rbenv libexec completions rbenv bash default rbenv rehash dev null default rbenv default typeset command default command default if then default shift default fi default default case command in default rehash shell default eval rbenv sh command default default command rbenv command default esac default default info bash ran successfully default execute bash tmp chef default log action write default info rbenv post init vagrant default default ruby block action run default info ruby block called default default execute the ruby block add rbenv to path default recipe ruby rbenv user default ruby rbenv ruby vagrant action install default default apt package action install default info apt package installed autoconf at default install version of package autoconf default apt package action install default info apt package installed bison at dfsg default default install version dfsg of package bison default apt package action install default info apt package installed build essential at default default install version of package build essential default apt package action install default info apt package installed libssl dev at default default install version of package libssl dev default apt package action install default info apt package installed libyaml dev at default default install version of package libyaml dev default apt package action install default info apt package installed dev at default install version of package dev default apt package action install default up to date default apt package action install default info apt package installed dev at default default install version of package dev default apt package action install default info apt package installed dev at default default default install version of package dev default apt package action install default info apt package installed dev at default default install version of package dev default apt package action install default up to date default apt package action install default info apt package installed libffi dev at default default install version of package libffi dev default apt package action install default up to date default apt package action install default info apt package installed libgdbm dev at default install version of package libgdbm dev default info building ruby rbenv ruby vagrant this could take a while default ruby rbenv script action run default default script action run default default downloading ruby tar default default installing ruby default installed ruby to home vagrant rbenv versions default default info script ran successfully default execute bash tmp chef default default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing default skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default apt package action nothing skipped due to action nothing default ruby rbenv script action nothing skipped due to action nothing default default ruby rbenv global vagrant action create warn only if block for ruby rbenv global vagrant returned did you mean to run a command if so use only if in your code default default ruby rbenv script action run default script action run default info script ran successfully default default execute bash tmp chef default default default vagrant action install default default default error executing action install on resource vagrant default default default chef exceptions filenotfound default default unable to find gem binary in path usr local rbenv shims usr local rbenv bin usr local sbin usr local bin usr sbin usr bin sbin bin default default cookbook trace default default var chef cache cookbooks ruby rbenv libraries chef provider package rbenvrubygems rb in initialize default default resource declaration default default in var chef cache cookbooks ruby rbenv recipes user rb default default rbenv gem gem rbenv user do default package name gem default user rbenv user default root path rbenv user if rbenv user default rbenv version rubie default default w version action options source each do attr default send attr gem if gem default end default end default end default default compiled resource default default declared in var chef cache cookbooks ruby rbenv recipes user rb in block levels in from file default default ruby rbenv gem bundler vagrant do default provider chef provider package rbenvrubygems default action default default guard interpreter default default declared type rbenv gem default cookbook name ruby rbenv default recipe name user default package name bundler default user vagrant default rbenv version default end default default system info default default chef version default platform ubuntu default platform version default ruby ruby revision default program name chef solo worker ppid start default executable opt chef bin chef solo default default info running queued delayed notifications before re raising exception default running handlers default error running exception handlers default running handlers complete default error exception handlers complete default chef client failed resources updated in minutes seconds default fatal stacktrace dumped to var chef cache chef stacktrace out default fatal please provide the contents of the stacktrace out file if you file a bug report default error vagrant ruby rbenv user line had an error chef exceptions filenotfound unable to find gem binary in path usr local rbenv shims usr local rbenv bin usr local sbin usr local bin usr sbin usr bin sbin bin default fatal chef exceptions childconvergeerror chef run process exited unsuccessfully exit code chef never successfully completed any errors should be visible in the output above please fix your recipes so that they properly complete stack out generated at chef exceptions filenotfound vagrant ruby rbenv user line had an error chef exceptions filenotfound unable to find gem binary in path usr local rbenv shims usr local rbenv bin usr local sbin usr local bin usr sbin usr bin sbin bin opt chef embedded lib ruby gems gems chef lib chef provider package rubygems rb in find gem by path opt chef embedded lib ruby gems gems chef lib chef provider package rubygems rb in initialize var chef cache cookbooks ruby rbenv libraries chef provider package rbenvrubygems rb in initialize opt chef embedded lib ruby gems gems chef lib chef resource rb in new opt chef embedded lib ruby gems gems chef lib chef resource rb in provider for action opt chef embedded lib ruby gems gems chef lib chef resource rb in run action opt chef embedded lib ruby gems gems chef lib chef runner rb in run action opt chef embedded lib ruby gems gems chef lib chef runner rb in block levels in converge opt chef embedded lib ruby gems gems chef lib chef runner rb in each opt chef embedded lib ruby gems gems chef lib chef runner rb in block in converge opt chef embedded lib ruby gems gems chef lib chef resource collection resource list rb in block in execute each resource opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in call iterator block opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in step opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in iterate opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in each with index opt chef embedded lib ruby gems gems chef lib chef resource collection resource list rb in execute each resource opt chef embedded lib ruby gems gems chef lib chef runner rb in converge opt chef embedded lib ruby gems gems chef lib chef client rb in block in converge opt chef embedded lib ruby gems gems chef lib chef client rb in catch opt chef embedded lib ruby gems gems chef lib chef client rb in converge opt chef embedded lib ruby gems gems chef lib chef client rb in converge and save opt chef embedded lib ruby gems gems chef lib chef client rb in run opt chef embedded lib ruby gems gems chef lib chef application rb in block in fork chef client opt chef embedded lib ruby gems gems chef lib chef application rb in fork opt chef embedded lib ruby gems gems chef lib chef application rb in fork chef client opt chef embedded lib ruby gems gems chef lib chef application rb in block in run chef client opt chef embedded lib ruby gems gems chef lib chef local mode rb in with server connectivity opt chef embedded lib ruby gems gems chef lib chef application rb in run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in sleep then run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in block in interval run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in loop opt chef embedded lib ruby gems gems chef lib chef application client rb in interval run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in run application opt chef embedded lib ruby gems gems chef lib chef application rb in run opt chef embedded lib ruby gems gems chef lib chef application solo rb in run opt chef embedded lib ruby gems gems chef bin chef solo in usr bin chef solo in load usr bin chef solo in caused by chef exceptions filenotfound unable to find gem binary in path usr local rbenv shims usr local rbenv bin usr local sbin usr local bin usr sbin usr bin sbin bin opt chef embedded lib ruby gems gems chef lib chef provider package rubygems rb in find gem by path opt chef embedded lib ruby gems gems chef lib chef provider package rubygems rb in initialize var chef cache cookbooks ruby rbenv libraries chef provider package rbenvrubygems rb in initialize opt chef embedded lib ruby gems gems chef lib chef resource rb in new opt chef embedded lib ruby gems gems chef lib chef resource rb in provider for action opt chef embedded lib ruby gems gems chef lib chef resource rb in run action opt chef embedded lib ruby gems gems chef lib chef runner rb in run action opt chef embedded lib ruby gems gems chef lib chef runner rb in block levels in converge opt chef embedded lib ruby gems gems chef lib chef runner rb in each opt chef embedded lib ruby gems gems chef lib chef runner rb in block in converge opt chef embedded lib ruby gems gems chef lib chef resource collection resource list rb in block in execute each resource opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in call iterator block opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in step opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in iterate opt chef embedded lib ruby gems gems chef lib chef resource collection stepable iterator rb in each with index opt chef embedded lib ruby gems gems chef lib chef resource collection resource list rb in execute each resource opt chef embedded lib ruby gems gems chef lib chef runner rb in converge opt chef embedded lib ruby gems gems chef lib chef client rb in block in converge opt chef embedded lib ruby gems gems chef lib chef client rb in catch opt chef embedded lib ruby gems gems chef lib chef client rb in converge opt chef embedded lib ruby gems gems chef lib chef client rb in converge and save opt chef embedded lib ruby gems gems chef lib chef client rb in run opt chef embedded lib ruby gems gems chef lib chef application rb in block in fork chef client opt chef embedded lib ruby gems gems chef lib chef application rb in fork opt chef embedded lib ruby gems gems chef lib chef application rb in fork chef client opt chef embedded lib ruby gems gems chef lib chef application rb in block in run chef client opt chef embedded lib ruby gems gems chef lib chef local mode rb in with server connectivity opt chef embedded lib ruby gems gems chef lib chef application rb in run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in sleep then run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in block in interval run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in loop opt chef embedded lib ruby gems gems chef lib chef application client rb in interval run chef client opt chef embedded lib ruby gems gems chef lib chef application client rb in run application opt chef embedded lib ruby gems gems chef lib chef application rb in run opt chef embedded lib ruby gems gems chef lib chef application solo rb in run opt chef embedded lib ruby gems gems chef bin chef solo in usr bin chef solo in load usr bin chef solo in
| 1
|
206,882
| 23,405,675,487
|
IssuesEvent
|
2022-08-12 12:35:54
|
KOSASIH/nevatronic
|
https://api.github.com/repos/KOSASIH/nevatronic
|
opened
|
gatsby-plugin-manifest-4.15.0.tgz: 1 vulnerabilities (highest severity is: 5.5)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>gatsby-plugin-manifest-4.15.0.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/file-type/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/KOSASIH/nevatronic/commit/f827072a4983d8b8bbfd01d9bc0f37a0403faca0">f827072a4983d8b8bbfd01d9bc0f37a0403faca0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-36313](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36313) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | file-type-16.5.3.tgz | Transitive | 4.15.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-36313</summary>
### Vulnerable Library - <b>file-type-16.5.3.tgz</b></p>
<p>Detect the file type of a Buffer/Uint8Array/ArrayBuffer</p>
<p>Library home page: <a href="https://registry.npmjs.org/file-type/-/file-type-16.5.3.tgz">https://registry.npmjs.org/file-type/-/file-type-16.5.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/file-type/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-plugin-manifest-4.15.0.tgz (Root Library)
- gatsby-core-utils-3.15.0.tgz
- :x: **file-type-16.5.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/KOSASIH/nevatronic/commit/f827072a4983d8b8bbfd01d9bc0f37a0403faca0">f827072a4983d8b8bbfd01d9bc0f37a0403faca0</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue was discovered in the file-type package before 16.5.4 and 17.x before 17.1.3 for Node.js. A malformed MKV file could cause the file type detector to get caught in an infinite loop. This would make the application become unresponsive and could be used to cause a DoS attack.
<p>Publish Date: 2022-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36313>CVE-2022-36313</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-07-21</p>
<p>Fix Resolution (file-type): 16.5.4</p>
<p>Direct dependency fix Resolution (gatsby-plugin-manifest): 4.15.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
gatsby-plugin-manifest-4.15.0.tgz: 1 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>gatsby-plugin-manifest-4.15.0.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/file-type/package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/KOSASIH/nevatronic/commit/f827072a4983d8b8bbfd01d9bc0f37a0403faca0">f827072a4983d8b8bbfd01d9bc0f37a0403faca0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-36313](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36313) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | file-type-16.5.3.tgz | Transitive | 4.15.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-36313</summary>
### Vulnerable Library - <b>file-type-16.5.3.tgz</b></p>
<p>Detect the file type of a Buffer/Uint8Array/ArrayBuffer</p>
<p>Library home page: <a href="https://registry.npmjs.org/file-type/-/file-type-16.5.3.tgz">https://registry.npmjs.org/file-type/-/file-type-16.5.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/file-type/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-plugin-manifest-4.15.0.tgz (Root Library)
- gatsby-core-utils-3.15.0.tgz
- :x: **file-type-16.5.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/KOSASIH/nevatronic/commit/f827072a4983d8b8bbfd01d9bc0f37a0403faca0">f827072a4983d8b8bbfd01d9bc0f37a0403faca0</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue was discovered in the file-type package before 16.5.4 and 17.x before 17.1.3 for Node.js. A malformed MKV file could cause the file type detector to get caught in an infinite loop. This would make the application become unresponsive and could be used to cause a DoS attack.
<p>Publish Date: 2022-07-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36313>CVE-2022-36313</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-07-21</p>
<p>Fix Resolution (file-type): 16.5.4</p>
<p>Direct dependency fix Resolution (gatsby-plugin-manifest): 4.15.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_defect
|
gatsby plugin manifest tgz vulnerabilities highest severity is vulnerable library gatsby plugin manifest tgz path to dependency file package json path to vulnerable library node modules file type package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium file type tgz transitive details cve vulnerable library file type tgz detect the file type of a buffer arraybuffer library home page a href path to dependency file package json path to vulnerable library node modules file type package json dependency hierarchy gatsby plugin manifest tgz root library gatsby core utils tgz x file type tgz vulnerable library found in head commit a href found in base branch main vulnerability details an issue was discovered in the file type package before and x before for node js a malformed mkv file could cause the file type detector to get caught in an infinite loop this would make the application become unresponsive and could be used to cause a dos attack publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution file type direct dependency fix resolution gatsby plugin manifest step up your open source security game with mend
| 0
|
142,895
| 11,499,526,879
|
IssuesEvent
|
2020-02-12 14:06:21
|
akka/akka
|
https://api.github.com/repos/akka/akka
|
opened
|
Add test using Tagged in persistence tck
|
1 - triaged help wanted t:persistence t:testing
|
To verify that the journal unwraps the replayed events. Detect issue like https://github.com/akka/akka/issues/28552
Possibility to opt out if journal doesn't support tagging.
|
1.0
|
Add test using Tagged in persistence tck - To verify that the journal unwraps the replayed events. Detect issue like https://github.com/akka/akka/issues/28552
Possibility to opt out if journal doesn't support tagging.
|
non_defect
|
add test using tagged in persistence tck to verify that the journal unwraps the replayed events detect issue like possibility to opt out if journal doesn t support tagging
| 0
|
96,861
| 16,168,287,799
|
IssuesEvent
|
2021-05-01 23:51:31
|
gabriel-milan/uptime-bot
|
https://api.github.com/repos/gabriel-milan/uptime-bot
|
opened
|
CVE-2021-28092 (High) detected in is-svg-3.0.0.tgz
|
security vulnerability
|
## CVE-2021-28092 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary>
<p>Check if a string or buffer is SVG</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p>
<p>Path to dependency file: uptime-bot/client/package.json</p>
<p>Path to vulnerable library: uptime-bot/client/node_modules/postcss-svgo/node_modules/is-svg/package.json</p>
<p>
Dependency Hierarchy:
- cli-service-4.5.11.tgz (Root Library)
- optimize-cssnano-plugin-1.0.6.tgz
- cssnano-preset-default-4.0.7.tgz
- postcss-svgo-4.0.2.tgz
- :x: **is-svg-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gabriel-milan/uptime-bot/commit/216b2d1977764ebebd21770dadc4261dd1f6d51c">216b2d1977764ebebd21770dadc4261dd1f6d51c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time.
<p>Publish Date: 2021-03-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p>
<p>Release Date: 2021-03-12</p>
<p>Fix Resolution: v4.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-28092 (High) detected in is-svg-3.0.0.tgz - ## CVE-2021-28092 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-svg-3.0.0.tgz</b></p></summary>
<p>Check if a string or buffer is SVG</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz">https://registry.npmjs.org/is-svg/-/is-svg-3.0.0.tgz</a></p>
<p>Path to dependency file: uptime-bot/client/package.json</p>
<p>Path to vulnerable library: uptime-bot/client/node_modules/postcss-svgo/node_modules/is-svg/package.json</p>
<p>
Dependency Hierarchy:
- cli-service-4.5.11.tgz (Root Library)
- optimize-cssnano-plugin-1.0.6.tgz
- cssnano-preset-default-4.0.7.tgz
- postcss-svgo-4.0.2.tgz
- :x: **is-svg-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gabriel-milan/uptime-bot/commit/216b2d1977764ebebd21770dadc4261dd1f6d51c">216b2d1977764ebebd21770dadc4261dd1f6d51c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The is-svg package 2.1.0 through 4.2.1 for Node.js uses a regular expression that is vulnerable to Regular Expression Denial of Service (ReDoS). If an attacker provides a malicious string, is-svg will get stuck processing the input for a very long time.
<p>Publish Date: 2021-03-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092>CVE-2021-28092</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28092</a></p>
<p>Release Date: 2021-03-12</p>
<p>Fix Resolution: v4.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in is svg tgz cve high severity vulnerability vulnerable library is svg tgz check if a string or buffer is svg library home page a href path to dependency file uptime bot client package json path to vulnerable library uptime bot client node modules postcss svgo node modules is svg package json dependency hierarchy cli service tgz root library optimize cssnano plugin tgz cssnano preset default tgz postcss svgo tgz x is svg tgz vulnerable library found in head commit a href found in base branch master vulnerability details the is svg package through for node js uses a regular expression that is vulnerable to regular expression denial of service redos if an attacker provides a malicious string is svg will get stuck processing the input for a very long time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
63,760
| 17,930,769,624
|
IssuesEvent
|
2021-09-10 08:56:57
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
Space Settings Dialog do not support restricted join rule
|
T-Defect S-Minor A-Spaces A-Space-Settings O-Intermediate
|
### Steps to reproduce
1. Create a sub space
2. Right Click on the created sub space
3. Got to the visibility tab
### What happened?
You only have option to set public or private (invite only)
No option to set as Restricted
Also if the subspace is actually restricted, the UI will show it as public
<img width="436" alt="image" src="https://user-images.githubusercontent.com/9841565/130999242-802283b6-9a98-4abc-bded-766e5a980875.png">
### What did you expect?
Would expect to have the 3 options like we have for rooms
<img width="699" alt="image" src="https://user-images.githubusercontent.com/9841565/130999302-de1db439-b284-4480-aed0-b6175520308b.png">
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
develop.element.io
### Homeserver
matrix.org
### Have you submitted a rageshake?
No
|
1.0
|
Space Settings Dialog do not support restricted join rule - ### Steps to reproduce
1. Create a sub space
2. Right Click on the created sub space
3. Got to the visibility tab
### What happened?
You only have option to set public or private (invite only)
No option to set as Restricted
Also if the subspace is actually restricted, the UI will show it as public
<img width="436" alt="image" src="https://user-images.githubusercontent.com/9841565/130999242-802283b6-9a98-4abc-bded-766e5a980875.png">
### What did you expect?
Would expect to have the 3 options like we have for rooms
<img width="699" alt="image" src="https://user-images.githubusercontent.com/9841565/130999302-de1db439-b284-4480-aed0-b6175520308b.png">
### Operating system
_No response_
### Application version
_No response_
### How did you install the app?
develop.element.io
### Homeserver
matrix.org
### Have you submitted a rageshake?
No
|
defect
|
space settings dialog do not support restricted join rule steps to reproduce create a sub space right click on the created sub space got to the visibility tab what happened you only have option to set public or private invite only no option to set as restricted also if the subspace is actually restricted the ui will show it as public img width alt image src what did you expect would expect to have the options like we have for rooms img width alt image src operating system no response application version no response how did you install the app develop element io homeserver matrix org have you submitted a rageshake no
| 1
|
59,575
| 17,023,166,122
|
IssuesEvent
|
2021-07-03 00:40:15
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Duplicating segments & nodes
|
Component: potlatch (flash editor) Priority: major Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 12.23pm, Tuesday, 22nd May 2007]**
I'm not sure under what circumstances this occurs, and it may be when I myself do something wrong, but I have noticed duplicated nodes, and segments.
Occasionally I have noticed these duplicated segments form a way as well.
It is possible that these occur when either a way is split using Potlatch, or when a new way is joined onto an existing one with a new node. (maybe its not exactly the fault of Potlatch but due to the server being unresponsive [lots of 500 Internal server Errors being reported by JOSM today] and not all of Potlatch's edits get uploaded to the server?)
|
1.0
|
Duplicating segments & nodes - **[Submitted to the original trac issue database at 12.23pm, Tuesday, 22nd May 2007]**
I'm not sure under what circumstances this occurs, and it may be when I myself do something wrong, but I have noticed duplicated nodes, and segments.
Occasionally I have noticed these duplicated segments form a way as well.
It is possible that these occur when either a way is split using Potlatch, or when a new way is joined onto an existing one with a new node. (maybe its not exactly the fault of Potlatch but due to the server being unresponsive [lots of 500 Internal server Errors being reported by JOSM today] and not all of Potlatch's edits get uploaded to the server?)
|
defect
|
duplicating segments nodes i m not sure under what circumstances this occurs and it may be when i myself do something wrong but i have noticed duplicated nodes and segments occasionally i have noticed these duplicated segments form a way as well it is possible that these occur when either a way is split using potlatch or when a new way is joined onto an existing one with a new node maybe its not exactly the fault of potlatch but due to the server being unresponsive and not all of potlatch s edits get uploaded to the server
| 1
|
1,403
| 2,603,847,293
|
IssuesEvent
|
2015-02-24 18:16:17
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳病毒疣的初期症状
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳病毒疣的初期症状〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:16
|
1.0
|
沈阳病毒疣的初期症状 - ```
沈阳病毒疣的初期症状〓沈陽軍區政治部醫院性病〓TEL:024-3
1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。�
��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌�
��歷史悠久、設備精良、技術權威、專家云集,是預防、保健
、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲��
�部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、�
��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空
軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體��
�等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:16
|
defect
|
沈阳病毒疣的初期症状 沈阳病毒疣的初期症状〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
| 1
|
413,267
| 12,062,143,122
|
IssuesEvent
|
2020-04-16 02:06:44
|
pkleskovic/a1map-angular
|
https://api.github.com/repos/pkleskovic/a1map-angular
|
closed
|
Separate local storage stuff into a separate service
|
Priority: Low Status: Abandoned Type: Maintenance
|
This is just a code styling issue, should be done if there is any time at the end of the dev cycle.
|
1.0
|
Separate local storage stuff into a separate service - This is just a code styling issue, should be done if there is any time at the end of the dev cycle.
|
non_defect
|
separate local storage stuff into a separate service this is just a code styling issue should be done if there is any time at the end of the dev cycle
| 0
|
270,588
| 28,967,854,714
|
IssuesEvent
|
2023-05-10 09:08:02
|
samqws-mendplatform/easybuggy
|
https://api.github.com/repos/samqws-mendplatform/easybuggy
|
opened
|
CVE-2019-17571 (Critical) detected in log4j-1.2.13.jar
|
Mend: dependency security vulnerability
|
## CVE-2019-17571 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- slf4j-log4j12-1.5.0.jar (Root Library)
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samqws-mendplatform/easybuggy/commit/cf985b0a40175770ef863bac7c7715e4aeef8302">cf985b0a40175770ef863bac7c7715e4aeef8302</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E">https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: log4j-manual - 1.2.17-16;log4j-javadoc - 1.2.17-16;log4j - 1.2.17-16,1.2.17-16</p>
</p>
</details>
<p></p>
|
True
|
CVE-2019-17571 (Critical) detected in log4j-1.2.13.jar - ## CVE-2019-17571 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- slf4j-log4j12-1.5.0.jar (Root Library)
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samqws-mendplatform/easybuggy/commit/cf985b0a40175770ef863bac7c7715e4aeef8302">cf985b0a40175770ef863bac7c7715e4aeef8302</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E">https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: log4j-manual - 1.2.17-16;log4j-javadoc - 1.2.17-16;log4j - 1.2.17-16,1.2.17-16</p>
</p>
</details>
<p></p>
|
non_defect
|
cve critical detected in jar cve critical severity vulnerability vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository jar dependency hierarchy jar root library x jar vulnerable library found in head commit a href found in base branch master vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution manual javadoc
| 0
|
69,974
| 22,773,109,783
|
IssuesEvent
|
2022-07-08 12:01:15
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
opened
|
When DataTable is updated via ajax event, the value of the selectAll checkbox changes
|
:lady_beetle: defect 8.0.19 11.0.6 10.0.14
|
Reported By PRO User;
> When all items within the table are selected but some are removed afterwards, those that were removed are not acknowledged.
listForm:tableList_selection: "@all" -> not affected by unchecking
Updating the table on rowUnselectCheckbox event changes "@all" into a list of row ids that are checked which is ultimately a wanted behaviour, but updating the table is not a viable solution.
|
1.0
|
When DataTable is updated via ajax event, the value of the selectAll checkbox changes - Reported By PRO User;
> When all items within the table are selected but some are removed afterwards, those that were removed are not acknowledged.
listForm:tableList_selection: "@all" -> not affected by unchecking
Updating the table on rowUnselectCheckbox event changes "@all" into a list of row ids that are checked which is ultimately a wanted behaviour, but updating the table is not a viable solution.
|
defect
|
when datatable is updated via ajax event the value of the selectall checkbox changes reported by pro user when all items within the table are selected but some are removed afterwards those that were removed are not acknowledged listform tablelist selection all not affected by unchecking updating the table on rowunselectcheckbox event changes all into a list of row ids that are checked which is ultimately a wanted behaviour but updating the table is not a viable solution
| 1
|
106,491
| 16,677,004,341
|
IssuesEvent
|
2021-06-07 17:29:31
|
XG-xin/bolt-testing2
|
https://api.github.com/repos/XG-xin/bolt-testing2
|
closed
|
CVE-2020-11022 (Medium) detected in jquery-3.4.1.min.js, phpunit/php-code-coverage-7.0.14 - autoclosed
|
security vulnerability
|
## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.4.1.min.js</b>, <b>phpunit/php-code-coverage-7.0.14</b></p></summary>
<p>
<details><summary><b>jquery-3.4.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.4.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.4.1/jquery.min.js</a></p>
<p>Path to vulnerable library: bolt-testing2/vendor/phpunit/php-code-coverage/src/Report/Html/Renderer/Template/js/jquery.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.4.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>phpunit/php-code-coverage-7.0.14</b></p></summary>
<p>Library that provides collection, processing, and rendering functionality for PHP code coverage information.</p>
<p>Library home page: <a href="https://api.github.com/repos/sebastianbergmann/php-code-coverage/zipball/bb7c9a210c72e4709cdde67f8b7362f672f2225c">https://api.github.com/repos/sebastianbergmann/php-code-coverage/zipball/bb7c9a210c72e4709cdde67f8b7362f672f2225c</a></p>
<p>
Dependency Hierarchy:
- phpcompatibility/php-compatibility-9.3.5 (Root Library)
- phpunit/phpunit-8.5.15
- :x: **phpunit/php-code-coverage-7.0.14** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/XG-xin/bolt-testing2/commits/51a861a0c933482376f9d7c3282f164a91f084cf">51a861a0c933482376f9d7c3282f164a91f084cf</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-11022 (Medium) detected in jquery-3.4.1.min.js, phpunit/php-code-coverage-7.0.14 - autoclosed - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-3.4.1.min.js</b>, <b>phpunit/php-code-coverage-7.0.14</b></p></summary>
<p>
<details><summary><b>jquery-3.4.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.4.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.4.1/jquery.min.js</a></p>
<p>Path to vulnerable library: bolt-testing2/vendor/phpunit/php-code-coverage/src/Report/Html/Renderer/Template/js/jquery.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.4.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>phpunit/php-code-coverage-7.0.14</b></p></summary>
<p>Library that provides collection, processing, and rendering functionality for PHP code coverage information.</p>
<p>Library home page: <a href="https://api.github.com/repos/sebastianbergmann/php-code-coverage/zipball/bb7c9a210c72e4709cdde67f8b7362f672f2225c">https://api.github.com/repos/sebastianbergmann/php-code-coverage/zipball/bb7c9a210c72e4709cdde67f8b7362f672f2225c</a></p>
<p>
Dependency Hierarchy:
- phpcompatibility/php-compatibility-9.3.5 (Root Library)
- phpunit/phpunit-8.5.15
- :x: **phpunit/php-code-coverage-7.0.14** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/XG-xin/bolt-testing2/commits/51a861a0c933482376f9d7c3282f164a91f084cf">51a861a0c933482376f9d7c3282f164a91f084cf</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in jquery min js phpunit php code coverage autoclosed cve medium severity vulnerability vulnerable libraries jquery min js phpunit php code coverage jquery min js javascript library for dom operations library home page a href path to vulnerable library bolt vendor phpunit php code coverage src report html renderer template js jquery min js dependency hierarchy x jquery min js vulnerable library phpunit php code coverage library that provides collection processing and rendering functionality for php code coverage information library home page a href dependency hierarchy phpcompatibility php compatibility root library phpunit phpunit x phpunit php code coverage vulnerable library found in head commit a href found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
268,578
| 23,381,298,677
|
IssuesEvent
|
2022-08-11 09:42:43
|
IDgis/geoportaal-test
|
https://api.github.com/repos/IDgis/geoportaal-test
|
opened
|
Aanpassen geavanceerd zoeken met nieuwe onderzoeken
|
gebruikerstest onderzoeksbank
|
Het standaard zoeken kon voorheen op basis van een 'gewoon' zoekterm. Daarnaast was een een optie voor uitgebreid zoeken waarbij ook op ISO onderwerpen gefilterd kon worden. Dit is nu niet meer van toepassing op onderzoeken.
Dit betekent de volgende wijzigingen voor het nieuwe geavanceerde zoeken:
- De optie om op ISO onderwerpen te zoeken is verwijderd
- Er is een zoekingang gekomen om te filteren op type onderzoek
- Er is een zoekingang gekomen om te filteren op thema
Uiteraard kunnen de verschillende opties ook gecombineerd worden voor een nog specifiekere zoekvraag.
|
1.0
|
Aanpassen geavanceerd zoeken met nieuwe onderzoeken - Het standaard zoeken kon voorheen op basis van een 'gewoon' zoekterm. Daarnaast was een een optie voor uitgebreid zoeken waarbij ook op ISO onderwerpen gefilterd kon worden. Dit is nu niet meer van toepassing op onderzoeken.
Dit betekent de volgende wijzigingen voor het nieuwe geavanceerde zoeken:
- De optie om op ISO onderwerpen te zoeken is verwijderd
- Er is een zoekingang gekomen om te filteren op type onderzoek
- Er is een zoekingang gekomen om te filteren op thema
Uiteraard kunnen de verschillende opties ook gecombineerd worden voor een nog specifiekere zoekvraag.
|
non_defect
|
aanpassen geavanceerd zoeken met nieuwe onderzoeken het standaard zoeken kon voorheen op basis van een gewoon zoekterm daarnaast was een een optie voor uitgebreid zoeken waarbij ook op iso onderwerpen gefilterd kon worden dit is nu niet meer van toepassing op onderzoeken dit betekent de volgende wijzigingen voor het nieuwe geavanceerde zoeken de optie om op iso onderwerpen te zoeken is verwijderd er is een zoekingang gekomen om te filteren op type onderzoek er is een zoekingang gekomen om te filteren op thema uiteraard kunnen de verschillende opties ook gecombineerd worden voor een nog specifiekere zoekvraag
| 0
|
36,862
| 8,170,663,450
|
IssuesEvent
|
2018-08-27 08:33:25
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
closed
|
IPv6 support on EC2
|
Team: Core Type: Defect
|
AWS EC2 [supports](https://aws.amazon.com/blogs/aws/new-ipv6-support-for-ec2-instances-in-virtual-private-clouds/) IPv6, but the Hazelcast explicitly [forces using IPv4](https://github.com/hazelcast/hazelcast/blob/v3.9/hazelcast/src/main/java/com/hazelcast/instance/DefaultAddressPicker.java#L367-L375) when AWS join is configured.
```java
private boolean preferIPv4Stack() {
boolean preferIPv4Stack = Boolean.getBoolean("java.net.preferIPv4Stack")
|| hazelcastProperties.getBoolean(GroupProperty.PREFER_IPv4_STACK);
// AWS does not support IPv6
JoinConfig join = config.getNetworkConfig().getJoin();
AwsConfig awsConfig = join.getAwsConfig();
boolean awsEnabled = awsConfig != null && awsConfig.isEnabled();
return preferIPv4Stack || awsEnabled;
}
```
**Suggested fix:**
Don't provide the configuration based artificial intelligence (e.g. cloud-provider based). Use just the system (and/or Hazelcast) properties to control the IPv6 support. Users should know themselves if their environment supports the IPv6.
|
1.0
|
IPv6 support on EC2 - AWS EC2 [supports](https://aws.amazon.com/blogs/aws/new-ipv6-support-for-ec2-instances-in-virtual-private-clouds/) IPv6, but the Hazelcast explicitly [forces using IPv4](https://github.com/hazelcast/hazelcast/blob/v3.9/hazelcast/src/main/java/com/hazelcast/instance/DefaultAddressPicker.java#L367-L375) when AWS join is configured.
```java
private boolean preferIPv4Stack() {
boolean preferIPv4Stack = Boolean.getBoolean("java.net.preferIPv4Stack")
|| hazelcastProperties.getBoolean(GroupProperty.PREFER_IPv4_STACK);
// AWS does not support IPv6
JoinConfig join = config.getNetworkConfig().getJoin();
AwsConfig awsConfig = join.getAwsConfig();
boolean awsEnabled = awsConfig != null && awsConfig.isEnabled();
return preferIPv4Stack || awsEnabled;
}
```
**Suggested fix:**
Don't provide the configuration based artificial intelligence (e.g. cloud-provider based). Use just the system (and/or Hazelcast) properties to control the IPv6 support. Users should know themselves if their environment supports the IPv6.
|
defect
|
support on aws but the hazelcast explicitly when aws join is configured java private boolean boolean boolean getboolean java net hazelcastproperties getboolean groupproperty prefer stack aws does not support joinconfig join config getnetworkconfig getjoin awsconfig awsconfig join getawsconfig boolean awsenabled awsconfig null awsconfig isenabled return awsenabled suggested fix don t provide the configuration based artificial intelligence e g cloud provider based use just the system and or hazelcast properties to control the support users should know themselves if their environment supports the
| 1
|
15,928
| 10,409,595,081
|
IssuesEvent
|
2019-09-13 09:10:22
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
Have the possibility to add an Azure VNET/Subnet to access an Azure Container Registery (Firewalls and virtual networks)
|
enhancement service/container-registry
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
We want to restrain the network access to our ACR (Azure Container Registry). For this, from the portal we have the ability to use the feature called Firewalls and virtual networks and select a specific VNET/Subnet to be able to access the ACR. However, Terraform does not seem to be able to provide us an attribute to configure that.
Currently, it appears that we can only use network_rule_set that supports the following:
- default_action
- ip_rule
So we can only specify CIDR or a public IP but what about a specific subnet inside a VNET? Because we can configure that for other services like Account Storage using the virtual_network_subnet_ids :
```
network_rules {
default_action = "Deny"
bypass = ["AzureServices"]
virtual_network_subnet_ids = ["${azurerm_subnet.test.id}"]
}
```
### New or Affected Resource(s)
Terraform v0.12.7
+ provider.azurerm v1.33.1
### Terraform Configuration
```
network_rule_set {
default_action = "Deny"
}
```
Thank you in advance :)
|
1.0
|
Have the possibility to add an Azure VNET/Subnet to access an Azure Container Registery (Firewalls and virtual networks) - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
We want to restrain the network access to our ACR (Azure Container Registry). For this, from the portal we have the ability to use the feature called Firewalls and virtual networks and select a specific VNET/Subnet to be able to access the ACR. However, Terraform does not seem to be able to provide us an attribute to configure that.
Currently, it appears that we can only use network_rule_set that supports the following:
- default_action
- ip_rule
So we can only specify CIDR or a public IP but what about a specific subnet inside a VNET? Because we can configure that for other services like Account Storage using the virtual_network_subnet_ids :
```
network_rules {
default_action = "Deny"
bypass = ["AzureServices"]
virtual_network_subnet_ids = ["${azurerm_subnet.test.id}"]
}
```
### New or Affected Resource(s)
Terraform v0.12.7
+ provider.azurerm v1.33.1
### Terraform Configuration
```
network_rule_set {
default_action = "Deny"
}
```
Thank you in advance :)
|
non_defect
|
have the possibility to add an azure vnet subnet to access an azure container registery firewalls and virtual networks community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description we want to restrain the network access to our acr azure container registry for this from the portal we have the ability to use the feature called firewalls and virtual networks and select a specific vnet subnet to be able to access the acr however terraform does not seem to be able to provide us an attribute to configure that currently it appears that we can only use network rule set that supports the following default action ip rule so we can only specify cidr or a public ip but what about a specific subnet inside a vnet because we can configure that for other services like account storage using the virtual network subnet ids network rules default action deny bypass virtual network subnet ids new or affected resource s terraform provider azurerm terraform configuration network rule set default action deny thank you in advance
| 0
|
34,473
| 7,451,837,138
|
IssuesEvent
|
2018-03-29 05:37:55
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Piirangutega KÜ kuvamine annab viga
|
P: high R: fixed T: defect
|
**Reported by jaak tamre on 25 Apr 2013 10:06 UTC**
'''Object'''
Kirjeldusüksuse detailvaade
http://dev.dira.teepub/et/description_unit/view/101000020708/
'''Description'''
```
An exception has been thrown during the rendering of a template ("Invalid ext date string: 2025-01-01 00:00:00") in DiraDescriptionUnitBundle:View:view.html.twig at line 392.
```
DescriptionUnitAccessTerms validSince peale hakkab seda karjuma. Kas seal peaks ikka olema laiendatud kuupäev või on midagi valesti läinud?
|
1.0
|
Piirangutega KÜ kuvamine annab viga - **Reported by jaak tamre on 25 Apr 2013 10:06 UTC**
'''Object'''
Kirjeldusüksuse detailvaade
http://dev.dira.teepub/et/description_unit/view/101000020708/
'''Description'''
```
An exception has been thrown during the rendering of a template ("Invalid ext date string: 2025-01-01 00:00:00") in DiraDescriptionUnitBundle:View:view.html.twig at line 392.
```
DescriptionUnitAccessTerms validSince peale hakkab seda karjuma. Kas seal peaks ikka olema laiendatud kuupäev või on midagi valesti läinud?
|
defect
|
piirangutega kü kuvamine annab viga reported by jaak tamre on apr utc object kirjeldusüksuse detailvaade description an exception has been thrown during the rendering of a template invalid ext date string in diradescriptionunitbundle view view html twig at line descriptionunitaccessterms validsince peale hakkab seda karjuma kas seal peaks ikka olema laiendatud kuupäev või on midagi valesti läinud
| 1
|
360,582
| 25,296,726,396
|
IssuesEvent
|
2022-11-17 07:27:11
|
WordPress/Documentation-Issue-Tracker
|
https://api.github.com/repos/WordPress/Documentation-Issue-Tracker
|
closed
|
Plugin handbook: Incorrect formatted code snippets on "Activation / Deactivation Hooks" page
|
developer documentation tracking issue plugins
|
## Issue Description
Some code snippet on https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/ are incorrect formatted.
## Example of an incorrect formatted code snippet
<img width="682" alt="Screenshot 2022-11-05 at 11 56 16" src="https://user-images.githubusercontent.com/3323310/200101893-7544c0fd-a3b1-463e-81fa-3456356a15e9.png">
## URL of the Page with the Issue
- https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/
## Section of Page with the issue
- [ ] [Activation](https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/#activation)
- [ ] [Deactivation](https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/#deactivation)
- [ ] [Example](https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/#example)
## Why is this a problem?
These code snippets might have been created before the existence of the Code Syntax Block.
## Suggested Fix
Using the same Code Syntax Block, as used for working code snippets such as https://developer.wordpress.org/themes/block-themes/converting-a-classic-theme-to-a-block-theme/#adding-block-template-parts-in-classic-themes, should solve this problem.
|
1.0
|
Plugin handbook: Incorrect formatted code snippets on "Activation / Deactivation Hooks" page - ## Issue Description
Some code snippet on https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/ are incorrect formatted.
## Example of an incorrect formatted code snippet
<img width="682" alt="Screenshot 2022-11-05 at 11 56 16" src="https://user-images.githubusercontent.com/3323310/200101893-7544c0fd-a3b1-463e-81fa-3456356a15e9.png">
## URL of the Page with the Issue
- https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/
## Section of Page with the issue
- [ ] [Activation](https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/#activation)
- [ ] [Deactivation](https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/#deactivation)
- [ ] [Example](https://developer.wordpress.org/plugins/plugin-basics/activation-deactivation-hooks/#example)
## Why is this a problem?
These code snippets might have been created before the existence of the Code Syntax Block.
## Suggested Fix
Using the same Code Syntax Block, as used for working code snippets such as https://developer.wordpress.org/themes/block-themes/converting-a-classic-theme-to-a-block-theme/#adding-block-template-parts-in-classic-themes, should solve this problem.
|
non_defect
|
plugin handbook incorrect formatted code snippets on activation deactivation hooks page issue description some code snippet on are incorrect formatted example of an incorrect formatted code snippet img width alt screenshot at src url of the page with the issue section of page with the issue why is this a problem these code snippets might have been created before the existence of the code syntax block suggested fix using the same code syntax block as used for working code snippets such as should solve this problem
| 0
|
255,480
| 19,304,288,078
|
IssuesEvent
|
2021-12-13 09:52:07
|
NetAppDocs/ontap-metrocluster
|
https://api.github.com/repos/NetAppDocs/ontap-metrocluster
|
closed
|
Add BES firmware upgrade info for 3.8.0.2
|
documentation
|
Page: [Configuring Broadcom IP switches](https://docs.netapp.com/us-en/ontap-metrocluster/install-ip/task_switch_config_broadcom.html)
In the "Downloading and installing the Broadcom switch EFOS software" section we don't have info for 3.8.0.2, which should be pretty similar to 3.7.0.4 (in terms of upgrade from older releases)
|
1.0
|
Add BES firmware upgrade info for 3.8.0.2 - Page: [Configuring Broadcom IP switches](https://docs.netapp.com/us-en/ontap-metrocluster/install-ip/task_switch_config_broadcom.html)
In the "Downloading and installing the Broadcom switch EFOS software" section we don't have info for 3.8.0.2, which should be pretty similar to 3.7.0.4 (in terms of upgrade from older releases)
|
non_defect
|
add bes firmware upgrade info for page in the downloading and installing the broadcom switch efos software section we don t have info for which should be pretty similar to in terms of upgrade from older releases
| 0
|
2,965
| 2,607,967,918
|
IssuesEvent
|
2015-02-26 00:43:12
|
chrsmithdemos/leveldb
|
https://api.github.com/repos/chrsmithdemos/leveldb
|
closed
|
Need an option to tell LevelDB to perform fsync periodly
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Open a database
2. ls -l in the database directory
3. Perform several Put operation
4. Wait for 10 seconds
5. ls -l in the same directory
What is the expected output? What do you see instead?
The size of *.log should change. But it is not, all files remain unmodified.
What version of the product are you using? On what operating system?
1.8.0, Linux
Please provide any additional information below.
* Feature required
Provide an option(sync, etc) when Open a database, if it is set to 1(one
second), a background thread will perform fsync on opened files every second.
As Linux kernel actually commits writes after 30 seconds, this is a long time,
may cause a lot data loss on machine failure.
```
-----
Original issue reported on code.google.com by `wuzuy...@gmail.com` on 22 Mar 2013 at 5:30
|
1.0
|
Need an option to tell LevelDB to perform fsync periodly - ```
What steps will reproduce the problem?
1. Open a database
2. ls -l in the database directory
3. Perform several Put operation
4. Wait for 10 seconds
5. ls -l in the same directory
What is the expected output? What do you see instead?
The size of *.log should change. But it is not, all files remain unmodified.
What version of the product are you using? On what operating system?
1.8.0, Linux
Please provide any additional information below.
* Feature required
Provide an option(sync, etc) when Open a database, if it is set to 1(one
second), a background thread will perform fsync on opened files every second.
As Linux kernel actually commits writes after 30 seconds, this is a long time,
may cause a lot data loss on machine failure.
```
-----
Original issue reported on code.google.com by `wuzuy...@gmail.com` on 22 Mar 2013 at 5:30
|
defect
|
need an option to tell leveldb to perform fsync periodly what steps will reproduce the problem open a database ls l in the database directory perform several put operation wait for seconds ls l in the same directory what is the expected output what do you see instead the size of log should change but it is not all files remain unmodified what version of the product are you using on what operating system linux please provide any additional information below feature required provide an option sync etc when open a database if it is set to one second a background thread will perform fsync on opened files every second as linux kernel actually commits writes after seconds this is a long time may cause a lot data loss on machine failure original issue reported on code google com by wuzuy gmail com on mar at
| 1
|
160,270
| 13,785,302,456
|
IssuesEvent
|
2020-10-08 22:34:27
|
informalsystems/tendermint-rs
|
https://api.github.com/repos/informalsystems/tendermint-rs
|
closed
|
docs: Document testing of rpc crate
|
documentation
|
We need comprehensive documentation for the ways in which we're testing the `rpc` crate in its [README](https://github.com/informalsystems/tendermint-rs/blob/master/rpc/README.md).
|
1.0
|
docs: Document testing of rpc crate - We need comprehensive documentation for the ways in which we're testing the `rpc` crate in its [README](https://github.com/informalsystems/tendermint-rs/blob/master/rpc/README.md).
|
non_defect
|
docs document testing of rpc crate we need comprehensive documentation for the ways in which we re testing the rpc crate in its
| 0
|
370,987
| 10,959,647,869
|
IssuesEvent
|
2019-11-27 11:52:58
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
int.search.tb.ask.com - see bug description
|
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important
|
<!-- @browser: Firefox 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:71.0) Gecko/20100101 Firefox/71.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://int.search.tb.ask.com/search/GGmain.jhtml?ct=ARS&n=7858e0ad&p2=%5EXP%5Expw344%5ETTAB02%5Ein&pg=GGmain&pn=1&ptb=465E6819-F78C-44AA-920A-64BA5AC12E86&qs=&si=21160&ss=sub&st=sb&searchfor=Google+Search&feedurl=%252Fars%252Ffeedback%253ForiginalQuery%253Dgoogle%2526relatedQuery%253Dgoogle%252Bsearch&tpr=jre10&ots=1574850778289
**Browser / Version**: Firefox 71.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: do not like it
**Steps to Reproduce**:
fhfftgghhhhhhjjjjjjjjjjjjjjjjjjj
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191114160003</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
<p>Console Messages:</p>
<pre>
[{'level': 'log', 'log': ['JQMIGRATE: Migrate is installed, version 3.0.1'], 'uri': 'https://code.jquery.com/jquery-migrate-3.0.1.min.js', 'pos': '33:19'}, {'level': 'warn', 'log': ['Use of Mutation Events is deprecated. Use MutationObserver instead.'], 'uri': 'https://code.jquery.com/jquery-3.3.1.min.js', 'pos': '2:40257'}, {'level': 'warn', 'log': ["Content Security Policy: Ignoring 'unsafe-inline' within script-src: strict-dynamic specified"], 'uri': '', 'pos': '0:0'}, {'level': 'warn', 'log': ['Content Security Policy: Ignoring http: within script-src: strict-dynamic specified'], 'uri': '', 'pos': '0:0'}, {'level': 'warn', 'log': ['Content Security Policy: Ignoring https: within script-src: strict-dynamic specified'], 'uri': '', 'pos': '0:0'}]
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
int.search.tb.ask.com - see bug description - <!-- @browser: Firefox 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:71.0) Gecko/20100101 Firefox/71.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://int.search.tb.ask.com/search/GGmain.jhtml?ct=ARS&n=7858e0ad&p2=%5EXP%5Expw344%5ETTAB02%5Ein&pg=GGmain&pn=1&ptb=465E6819-F78C-44AA-920A-64BA5AC12E86&qs=&si=21160&ss=sub&st=sb&searchfor=Google+Search&feedurl=%252Fars%252Ffeedback%253ForiginalQuery%253Dgoogle%2526relatedQuery%253Dgoogle%252Bsearch&tpr=jre10&ots=1574850778289
**Browser / Version**: Firefox 71.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: do not like it
**Steps to Reproduce**:
fhfftgghhhhhhjjjjjjjjjjjjjjjjjjj
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191114160003</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
<p>Console Messages:</p>
<pre>
[{'level': 'log', 'log': ['JQMIGRATE: Migrate is installed, version 3.0.1'], 'uri': 'https://code.jquery.com/jquery-migrate-3.0.1.min.js', 'pos': '33:19'}, {'level': 'warn', 'log': ['Use of Mutation Events is deprecated. Use MutationObserver instead.'], 'uri': 'https://code.jquery.com/jquery-3.3.1.min.js', 'pos': '2:40257'}, {'level': 'warn', 'log': ["Content Security Policy: Ignoring 'unsafe-inline' within script-src: strict-dynamic specified"], 'uri': '', 'pos': '0:0'}, {'level': 'warn', 'log': ['Content Security Policy: Ignoring http: within script-src: strict-dynamic specified'], 'uri': '', 'pos': '0:0'}, {'level': 'warn', 'log': ['Content Security Policy: Ignoring https: within script-src: strict-dynamic specified'], 'uri': '', 'pos': '0:0'}]
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_defect
|
int search tb ask com see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description do not like it steps to reproduce fhfftgghhhhhhjjjjjjjjjjjjjjjjjjj browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false console messages uri pos level warn log uri pos level warn log uri pos level warn log uri pos level warn log uri pos from with ❤️
| 0
|
741,302
| 25,787,964,985
|
IssuesEvent
|
2022-12-09 22:52:49
|
bounswe/bounswe2022group1
|
https://api.github.com/repos/bounswe/bounswe2022group1
|
opened
|
Completing Missing Tasks After Milestone 2
|
Priority: High Type: Task Status: In Progress Frontend
|
**Issue Description:**
I didn't complete some of my duties for Milestone 2. I am working on them to catch up with our project and be coherent with my group members' works.
**Tasks to Do & Duration:**
- [ ] add issue labels
- [ ] add reviewer
- [ ] add related links
- [ ] look up short nextjs tutorial - **30min**
- [ ] move latest folder structure to development and frontend branches - **15min**
- [ ] merge latest changes in learningspace-detailpage branch to frontend branch - **15dk**
- [ ] change background theme of LS - **15dk**
- [ ] take contents' component of page as parameters and adding them to the suitable places in a LS detail page - **45dk**
- [ ] form the content's (text, video, picture) structure fits to the description in related API - **45dk**
- [ ] link API (content-get) to the endpoints - **45dk**
- [ ] implement join button to handle some membership and leaving from LS cases - **60dk**
*Task Deadline: 11/12/2022 15:00*
*Final Situation:*
*Reviewer: @egetaga @omerozdemir1 @ecesrkn @uskudarli*
|
1.0
|
Completing Missing Tasks After Milestone 2 - **Issue Description:**
I didn't complete some of my duties for Milestone 2. I am working on them to catch up with our project and be coherent with my group members' works.
**Tasks to Do & Duration:**
- [ ] add issue labels
- [ ] add reviewer
- [ ] add related links
- [ ] look up short nextjs tutorial - **30min**
- [ ] move latest folder structure to development and frontend branches - **15min**
- [ ] merge latest changes in learningspace-detailpage branch to frontend branch - **15dk**
- [ ] change background theme of LS - **15dk**
- [ ] take contents' component of page as parameters and adding them to the suitable places in a LS detail page - **45dk**
- [ ] form the content's (text, video, picture) structure fits to the description in related API - **45dk**
- [ ] link API (content-get) to the endpoints - **45dk**
- [ ] implement join button to handle some membership and leaving from LS cases - **60dk**
*Task Deadline: 11/12/2022 15:00*
*Final Situation:*
*Reviewer: @egetaga @omerozdemir1 @ecesrkn @uskudarli*
|
non_defect
|
completing missing tasks after milestone issue description i didn t complete some of my duties for milestone i am working on them to catch up with our project and be coherent with my group members works tasks to do duration add issue labels add reviewer add related links look up short nextjs tutorial move latest folder structure to development and frontend branches merge latest changes in learningspace detailpage branch to frontend branch change background theme of ls take contents component of page as parameters and adding them to the suitable places in a ls detail page form the content s text video picture structure fits to the description in related api link api content get to the endpoints implement join button to handle some membership and leaving from ls cases task deadline final situation reviewer egetaga ecesrkn uskudarli
| 0
|
14,606
| 2,829,610,108
|
IssuesEvent
|
2015-05-23 02:06:28
|
awesomebing1/fuzzdb
|
https://api.github.com/repos/awesomebing1/fuzzdb
|
closed
|
http://www.hidesertstar.com/calendar/event_7b6c28c8-6cd4-11e4-9752-cb6a3e4f46f4.html
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
http://www.hidesertstar.com/calendar/event_7b6c28c8-6cd4-11e4-9752-cb6a3e4f46f4.
html
What version of the
prohttp://www.hidesertstar.com/calendar/event_7b6c28c8-6cd4-11e4-9752-cb6a3e4f46
f4.htmlduct are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `sabujhos...@gmail.com` on 15 Nov 2014 at 2:47
|
1.0
|
http://www.hidesertstar.com/calendar/event_7b6c28c8-6cd4-11e4-9752-cb6a3e4f46f4.html - ```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
http://www.hidesertstar.com/calendar/event_7b6c28c8-6cd4-11e4-9752-cb6a3e4f46f4.
html
What version of the
prohttp://www.hidesertstar.com/calendar/event_7b6c28c8-6cd4-11e4-9752-cb6a3e4f46
f4.htmlduct are you using? On what operating system?
Please provide any additional information below.
```
Original issue reported on code.google.com by `sabujhos...@gmail.com` on 15 Nov 2014 at 2:47
|
defect
|
what steps will reproduce the problem what is the expected output what do you see instead html what version of the pro htmlduct are you using on what operating system please provide any additional information below original issue reported on code google com by sabujhos gmail com on nov at
| 1
|
55,841
| 14,706,061,926
|
IssuesEvent
|
2021-01-04 19:12:46
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
opened
|
kernel slab high memory usage during scrub OOM kill other applications
|
Status: Triage Needed Type: Defect
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name |Ubuntu
Distribution Version |18.04
Linux Kernel |5.4.0-58-generic
Architecture |amd64
ZFS Version |0.8.3-1ubuntu12.5
SPL Version |0.8.3-1ubuntu12.5
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
We run the HWE Ubuntu kernel which therefore means we get the 0.8.* version of zfs/spl, our issue is probably the same as:
https://github.com/openzfs/zfs/issues/8662
We run MySQL (Mariadb, actually) using zfs volumes for data and backup space (separate volumes), we run a scrub from cron every 4 weeks which takes ~4 hours, on our replicas the scrub generally completes without issue but with the primary we have seen MySQL crash (OOM killed on the last crash)
The servers are Intel Xeon Gold, with 512Gb RAM, disks are 6 x Intel S4510 SSD 3.8Tb in 3 x mirrored sets
### Describe how to reproduce the problem
Start a scrub on the data volume, then watch meminfo for Unreclaim usage:
```
zpool scrub mysqldata
```
```
Every 2.0s: cat /proc/meminfo | grep claim
Mon Jan 4 19:04:27 2021
KReclaimable: 2442512 kB
SReclaimable: 2442512 kB
SUnreclaim: 1932272 kB
```
after 30s later:
```
Every 2.0s: cat /proc/meminfo | grep claim
Mon Jan 4 19:05:02 2021
KReclaimable: 2442976 kB
SReclaimable: 2442976 kB
SUnreclaim: 7637196 kB
```
Then issue the stop:
```
zpool scrub -s mysqldata
```
Check again:
```
Every 2.0s: cat /proc/meminfo | grep claim
Mon Jan 4 19:06:05 2021
KReclaimable: 2442976 kB
SReclaimable: 2442976 kB
SUnreclaim: 1970984 kB
```
I was unable to alter the behaviour of the `SUnreclaim` meminfo value by changing any of `/sys/module/zfs/parameters/zfs_scan_mem_lim_fact`, `/sys/module/zfs/parameters/zfs_scan_mem_lim_soft_fact` or by adding `/sys/module/zfs/parameters/zfs_scrub_delay` (permission denied as root)
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
|
1.0
|
kernel slab high memory usage during scrub OOM kill other applications - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name |Ubuntu
Distribution Version |18.04
Linux Kernel |5.4.0-58-generic
Architecture |amd64
ZFS Version |0.8.3-1ubuntu12.5
SPL Version |0.8.3-1ubuntu12.5
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
We run the HWE Ubuntu kernel which therefore means we get the 0.8.* version of zfs/spl, our issue is probably the same as:
https://github.com/openzfs/zfs/issues/8662
We run MySQL (Mariadb, actually) using zfs volumes for data and backup space (separate volumes), we run a scrub from cron every 4 weeks which takes ~4 hours, on our replicas the scrub generally completes without issue but with the primary we have seen MySQL crash (OOM killed on the last crash)
The servers are Intel Xeon Gold, with 512Gb RAM, disks are 6 x Intel S4510 SSD 3.8Tb in 3 x mirrored sets
### Describe how to reproduce the problem
Start a scrub on the data volume, then watch meminfo for Unreclaim usage:
```
zpool scrub mysqldata
```
```
Every 2.0s: cat /proc/meminfo | grep claim
Mon Jan 4 19:04:27 2021
KReclaimable: 2442512 kB
SReclaimable: 2442512 kB
SUnreclaim: 1932272 kB
```
after 30s later:
```
Every 2.0s: cat /proc/meminfo | grep claim
Mon Jan 4 19:05:02 2021
KReclaimable: 2442976 kB
SReclaimable: 2442976 kB
SUnreclaim: 7637196 kB
```
Then issue the stop:
```
zpool scrub -s mysqldata
```
Check again:
```
Every 2.0s: cat /proc/meminfo | grep claim
Mon Jan 4 19:06:05 2021
KReclaimable: 2442976 kB
SReclaimable: 2442976 kB
SUnreclaim: 1970984 kB
```
I was unable to alter the behaviour of the `SUnreclaim` meminfo value by changing any of `/sys/module/zfs/parameters/zfs_scan_mem_lim_fact`, `/sys/module/zfs/parameters/zfs_scan_mem_lim_soft_fact` or by adding `/sys/module/zfs/parameters/zfs_scrub_delay` (permission denied as root)
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
|
defect
|
kernel slab high memory usage during scrub oom kill other applications thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu distribution version linux kernel generic architecture zfs version spl version commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing we run the hwe ubuntu kernel which therefore means we get the version of zfs spl our issue is probably the same as we run mysql mariadb actually using zfs volumes for data and backup space separate volumes we run a scrub from cron every weeks which takes hours on our replicas the scrub generally completes without issue but with the primary we have seen mysql crash oom killed on the last crash the servers are intel xeon gold with ram disks are x intel ssd in x mirrored sets describe how to reproduce the problem start a scrub on the data volume then watch meminfo for unreclaim usage zpool scrub mysqldata every cat proc meminfo grep claim mon jan kreclaimable kb sreclaimable kb sunreclaim kb after later every cat proc meminfo grep claim mon jan kreclaimable kb sreclaimable kb sunreclaim kb then issue the stop zpool scrub s mysqldata check again every cat proc meminfo grep claim mon jan kreclaimable kb sreclaimable kb sunreclaim kb i was unable to alter the behaviour of the sunreclaim meminfo value by changing any of sys module zfs parameters zfs scan mem lim fact sys module zfs parameters zfs scan mem lim soft fact or by adding sys module zfs parameters zfs scrub delay permission denied as root include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
| 1
|
666,228
| 22,346,751,857
|
IssuesEvent
|
2022-06-15 08:30:02
|
HonkingGoose/vale-plain-language-guidelines
|
https://api.github.com/repos/HonkingGoose/vale-plain-language-guidelines
|
closed
|
Create forms
|
type:refactor priority-4-low status:ready
|
Todo:
- [ ] Bug report form
- [ ] Feature requests form
- [ ] Refactoring form
Base forms on the style we use for `renovatebot/renovate` repository.
|
1.0
|
Create forms - Todo:
- [ ] Bug report form
- [ ] Feature requests form
- [ ] Refactoring form
Base forms on the style we use for `renovatebot/renovate` repository.
|
non_defect
|
create forms todo bug report form feature requests form refactoring form base forms on the style we use for renovatebot renovate repository
| 0
|
1,212
| 2,601,759,759
|
IssuesEvent
|
2015-02-24 00:34:35
|
chrsmith/bwapi
|
https://api.github.com/repos/chrsmith/bwapi
|
closed
|
Find method to retrieve accurate latency information
|
auto-migrated Component-Logic Maintainability Object-Game Priority-Medium Type-Defect Usability
|
```
Currently GetLatencyFrames, GetLatencyTime, GetRemainingLatencyFrames, and
GetRemainingLatencyTime return inaccurate values.
The data was never processed correctly, and only appeared correct on some game
modes.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 13 Dec 2010 at 6:47
|
1.0
|
Find method to retrieve accurate latency information - ```
Currently GetLatencyFrames, GetLatencyTime, GetRemainingLatencyFrames, and
GetRemainingLatencyTime return inaccurate values.
The data was never processed correctly, and only appeared correct on some game
modes.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 13 Dec 2010 at 6:47
|
defect
|
find method to retrieve accurate latency information currently getlatencyframes getlatencytime getremaininglatencyframes and getremaininglatencytime return inaccurate values the data was never processed correctly and only appeared correct on some game modes original issue reported on code google com by aheinerm on dec at
| 1
|
86,460
| 8,037,047,767
|
IssuesEvent
|
2018-07-30 11:14:25
|
resin-io-modules/resin-procbots
|
https://api.github.com/repos/resin-io-modules/resin-procbots
|
closed
|
SyncBot: Support threads that shift between synced inboxes
|
flow/testing
|
When a thread moves from one synced inbox to another, SyncBot incorrectly finds the association with the previous inbox, and nothing then syncs.
The search for a context should be inbox specific.
|
1.0
|
SyncBot: Support threads that shift between synced inboxes - When a thread moves from one synced inbox to another, SyncBot incorrectly finds the association with the previous inbox, and nothing then syncs.
The search for a context should be inbox specific.
|
non_defect
|
syncbot support threads that shift between synced inboxes when a thread moves from one synced inbox to another syncbot incorrectly finds the association with the previous inbox and nothing then syncs the search for a context should be inbox specific
| 0
|
568,564
| 16,983,089,594
|
IssuesEvent
|
2021-06-30 11:21:20
|
bryntum/support
|
https://api.github.com/repos/bryntum/support
|
closed
|
Infinite scroll along the time axis
|
epic feature request high-priority in progress resolved
|
- [ ] autoGrowTimeAxis config on Scheduler, reacting to ScrollManager reaching end or start point of the timeaxis (on by default for resize, drag drop, drag create interactions)
- [ ] Scrolling will use sliding window approach
- [ ] A demo showing incremental load as you scroll along the time axis
- [ ] Configurable buffer window size (bufferCoef in Ext Scheduler)?
https://www.bryntum.com/forum/viewtopic.php?p=74185#p74185
https://www.bryntum.com/forum/viewtopic.php?p=73065#p73065
https://www.bryntum.com/forum/viewtopic.php?p=69153#p69153
https://www.bryntum.com/forum/viewtopic.php?p=63455#p63455
https://www.bryntum.com/forum/viewtopic.php?p=62135#p62135
https://www.bryntum.com/forum/viewtopic.php?p=74578#p74578
|
1.0
|
Infinite scroll along the time axis - - [ ] autoGrowTimeAxis config on Scheduler, reacting to ScrollManager reaching end or start point of the timeaxis (on by default for resize, drag drop, drag create interactions)
- [ ] Scrolling will use sliding window approach
- [ ] A demo showing incremental load as you scroll along the time axis
- [ ] Configurable buffer window size (bufferCoef in Ext Scheduler)?
https://www.bryntum.com/forum/viewtopic.php?p=74185#p74185
https://www.bryntum.com/forum/viewtopic.php?p=73065#p73065
https://www.bryntum.com/forum/viewtopic.php?p=69153#p69153
https://www.bryntum.com/forum/viewtopic.php?p=63455#p63455
https://www.bryntum.com/forum/viewtopic.php?p=62135#p62135
https://www.bryntum.com/forum/viewtopic.php?p=74578#p74578
|
non_defect
|
infinite scroll along the time axis autogrowtimeaxis config on scheduler reacting to scrollmanager reaching end or start point of the timeaxis on by default for resize drag drop drag create interactions scrolling will use sliding window approach a demo showing incremental load as you scroll along the time axis configurable buffer window size buffercoef in ext scheduler
| 0
|
144,673
| 22,470,480,087
|
IssuesEvent
|
2022-06-22 07:41:04
|
AvaloniaUI/Avalonia
|
https://api.github.com/repos/AvaloniaUI/Avalonia
|
closed
|
Should PointerPointProperties be public in PointerEventArgs?
|
by-design
|
**Describe the bug**
I am trying to find out whether the left button is down whilst the mouse/pointer is moved within a Templated Control
**To Reproduce**
```
public class ArrowAngleControl : TemplatedControl
{
public ArrowAngleControl()
{
PointerMoved += ArrowAngleControl_PointerMoved;
}
private void ArrowAngleControl_PointerMoved(object? sender, global::Avalonia.Input.PointerEventArgs e)
{
bool isMouseDown = e.Properties.IsLeftButtonPressed; // e.Properties does not exist
}
}
```
**Expected behavior**
The API describes the PointerPointProperties here [http://reference.avaloniaui.net/api/Avalonia.Input/PointerEventArgs/](http://reference.avaloniaui.net/api/Avalonia.Input/PointerEventArgs/), however it is not accessible as it is set to protected [https://github.com/AvaloniaUI/Avalonia/blob/6314f4f34a0fad3d5bc46d40165beb0f7c4693fe/src/Avalonia.Base/Input/PointerEventArgs.cs#L143](https://github.com/AvaloniaUI/Avalonia/blob/6314f4f34a0fad3d5bc46d40165beb0f7c4693fe/src/Avalonia.Base/Input/PointerEventArgs.cs#L143)
Furthermore the comment for the e.InputModifiers deprecated, says to use PointerPointProperties
**Screenshots**
**Desktop (please complete the following information):**
- OS: [e.g. Windows, Mac, Linux (State distribution)]
- Version 0.10.15
|
1.0
|
Should PointerPointProperties be public in PointerEventArgs? - **Describe the bug**
I am trying to find out whether the left button is down whilst the mouse/pointer is moved within a Templated Control
**To Reproduce**
```
public class ArrowAngleControl : TemplatedControl
{
public ArrowAngleControl()
{
PointerMoved += ArrowAngleControl_PointerMoved;
}
private void ArrowAngleControl_PointerMoved(object? sender, global::Avalonia.Input.PointerEventArgs e)
{
bool isMouseDown = e.Properties.IsLeftButtonPressed; // e.Properties does not exist
}
}
```
**Expected behavior**
The API describes the PointerPointProperties here [http://reference.avaloniaui.net/api/Avalonia.Input/PointerEventArgs/](http://reference.avaloniaui.net/api/Avalonia.Input/PointerEventArgs/), however it is not accessible as it is set to protected [https://github.com/AvaloniaUI/Avalonia/blob/6314f4f34a0fad3d5bc46d40165beb0f7c4693fe/src/Avalonia.Base/Input/PointerEventArgs.cs#L143](https://github.com/AvaloniaUI/Avalonia/blob/6314f4f34a0fad3d5bc46d40165beb0f7c4693fe/src/Avalonia.Base/Input/PointerEventArgs.cs#L143)
Furthermore the comment for the e.InputModifiers deprecated, says to use PointerPointProperties
**Screenshots**
**Desktop (please complete the following information):**
- OS: [e.g. Windows, Mac, Linux (State distribution)]
- Version 0.10.15
|
non_defect
|
should pointerpointproperties be public in pointereventargs describe the bug i am trying to find out whether the left button is down whilst the mouse pointer is moved within a templated control to reproduce public class arrowanglecontrol templatedcontrol public arrowanglecontrol pointermoved arrowanglecontrol pointermoved private void arrowanglecontrol pointermoved object sender global avalonia input pointereventargs e bool ismousedown e properties isleftbuttonpressed e properties does not exist expected behavior the api describes the pointerpointproperties here however it is not accessible as it is set to protected furthermore the comment for the e inputmodifiers deprecated says to use pointerpointproperties screenshots desktop please complete the following information os version
| 0
|
41,688
| 10,567,682,596
|
IssuesEvent
|
2019-10-06 06:51:26
|
markoteittinen/custom-maps
|
https://api.github.com/repos/markoteittinen/custom-maps
|
closed
|
remove library dependency
|
Priority-Medium Type-Defect auto-migrated
|
```
Hi,
ive tryed to install costommaps but ive got the following error:
E/PackageManager( 424): Package com.custommapsapp.android requires unavailable
shared library com.google.android.maps; failing!
W/PackageManager( 424): Package couldn't be installed in
/data/app/com.custommapsapp.android-1.apk
so please remove the dependencies of the propitiatory library
com.google.android.maps
thanks
```
Original issue reported on code.google.com by `tre...@googlemail.com` on 27 Jul 2014 at 3:03
|
1.0
|
remove library dependency - ```
Hi,
ive tryed to install costommaps but ive got the following error:
E/PackageManager( 424): Package com.custommapsapp.android requires unavailable
shared library com.google.android.maps; failing!
W/PackageManager( 424): Package couldn't be installed in
/data/app/com.custommapsapp.android-1.apk
so please remove the dependencies of the propitiatory library
com.google.android.maps
thanks
```
Original issue reported on code.google.com by `tre...@googlemail.com` on 27 Jul 2014 at 3:03
|
defect
|
remove library dependency hi ive tryed to install costommaps but ive got the following error e packagemanager package com custommapsapp android requires unavailable shared library com google android maps failing w packagemanager package couldn t be installed in data app com custommapsapp android apk so please remove the dependencies of the propitiatory library com google android maps thanks original issue reported on code google com by tre googlemail com on jul at
| 1
|
65,071
| 8,786,958,302
|
IssuesEvent
|
2018-12-20 17:07:30
|
krichtof/testboard
|
https://api.github.com/repos/krichtof/testboard
|
opened
|
Indiquer l'intention du projet
|
documentation
|
Permettre à n'importe qui de savoir pourquoi ce projet a été créé
Comment : Indiquer l'intention dans un fichier README
|
1.0
|
Indiquer l'intention du projet - Permettre à n'importe qui de savoir pourquoi ce projet a été créé
Comment : Indiquer l'intention dans un fichier README
|
non_defect
|
indiquer l intention du projet permettre à n importe qui de savoir pourquoi ce projet a été créé comment indiquer l intention dans un fichier readme
| 0
|
52,853
| 13,225,167,814
|
IssuesEvent
|
2020-08-17 20:37:35
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
noise-generator and very late MCHits/MCPEs (Trac #473)
|
Migrated from Trac combo simulation defect
|
For maps with very late MCHits/MCPEs noise-generator will fill the RAM until the node dies (60gb and more). MCPE times are in the range of seconds to hours after the initial neutrino interaction. They can be generated by clsim in geant4 mode.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/473">https://code.icecube.wisc.edu/projects/icecube/ticket/473</a>, reported by vehringand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-03-11T14:43:34",
"_ts": "1426085014817363",
"description": "For maps with very late MCHits/MCPEs noise-generator will fill the RAM until the node dies (60gb and more). MCPE times are in the range of seconds to hours after the initial neutrino interaction. They can be generated by clsim in geant4 mode.",
"reporter": "vehring",
"cc": "",
"resolution": "wontfix",
"time": "2013-11-18T23:04:03",
"component": "combo simulation",
"summary": "noise-generator and very late MCHits/MCPEs",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
noise-generator and very late MCHits/MCPEs (Trac #473) - For maps with very late MCHits/MCPEs noise-generator will fill the RAM until the node dies (60gb and more). MCPE times are in the range of seconds to hours after the initial neutrino interaction. They can be generated by clsim in geant4 mode.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/473">https://code.icecube.wisc.edu/projects/icecube/ticket/473</a>, reported by vehringand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-03-11T14:43:34",
"_ts": "1426085014817363",
"description": "For maps with very late MCHits/MCPEs noise-generator will fill the RAM until the node dies (60gb and more). MCPE times are in the range of seconds to hours after the initial neutrino interaction. They can be generated by clsim in geant4 mode.",
"reporter": "vehring",
"cc": "",
"resolution": "wontfix",
"time": "2013-11-18T23:04:03",
"component": "combo simulation",
"summary": "noise-generator and very late MCHits/MCPEs",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
noise generator and very late mchits mcpes trac for maps with very late mchits mcpes noise generator will fill the ram until the node dies and more mcpe times are in the range of seconds to hours after the initial neutrino interaction they can be generated by clsim in mode migrated from json status closed changetime ts description for maps with very late mchits mcpes noise generator will fill the ram until the node dies and more mcpe times are in the range of seconds to hours after the initial neutrino interaction they can be generated by clsim in mode reporter vehring cc resolution wontfix time component combo simulation summary noise generator and very late mchits mcpes priority normal keywords milestone owner olivas type defect
| 1
|
156,418
| 24,619,632,803
|
IssuesEvent
|
2022-10-15 19:04:44
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
opened
|
Templates: make it easier to do template specific style changes
|
[Type] Enhancement Needs Design Feedback [Type] Feedback Global Styles [Feature] Site Editor
|
## What problem does this address?
In the [seventeenth call for testing for the FSE Outreach Program](https://make.wordpress.org/test/2022/09/15/fse-program-testing-call-17-guiding-the-gutenberg-gallery/#comment-2749), the following feedback came in that reflects a likely common action someone might want to take: change the color scheme for a single template.
>When I created a template for the “Submission Opportunities” category, I thought I’d change the color scheme for this category from the Styles menu in the sidebar. The Style is not just applied to this single category, but to the whole site. Is that correct? Since I was creating a template for this category, and the call for testing instructions encouraged me to push things a bit, I thought I’d select a whole different color scheme just for this specific category, expecting the style to be only applied to the category and not the whole site.
Here's a quick video showing how it _seems_ at first that changing Styles does the trick, without a clear warning that it'll change the whole site:
https://user-images.githubusercontent.com/26996883/196003569-578a35cb-600a-45cd-9a54-96194ea4c51d.mov
In many ways, it doesn't seem like it should be this hard to make template specific changes to aspects as basic as background color without changing the entire site.
## What is your proposed solution?
Part of this is due to the lack of clarity when saving in multi entity saving (https://github.com/WordPress/gutenberg/issues/29388) but the other piece of this is around the visibility of Styles combined with the difficult in making _template specific_ changes. Curious what others think and what can be done as right now it seems like multiple layers need to be considered across global styles, saving changes, etc.
|
1.0
|
Templates: make it easier to do template specific style changes - ## What problem does this address?
In the [seventeenth call for testing for the FSE Outreach Program](https://make.wordpress.org/test/2022/09/15/fse-program-testing-call-17-guiding-the-gutenberg-gallery/#comment-2749), the following feedback came in that reflects a likely common action someone might want to take: change the color scheme for a single template.
>When I created a template for the “Submission Opportunities” category, I thought I’d change the color scheme for this category from the Styles menu in the sidebar. The Style is not just applied to this single category, but to the whole site. Is that correct? Since I was creating a template for this category, and the call for testing instructions encouraged me to push things a bit, I thought I’d select a whole different color scheme just for this specific category, expecting the style to be only applied to the category and not the whole site.
Here's a quick video showing how it _seems_ at first that changing Styles does the trick, without a clear warning that it'll change the whole site:
https://user-images.githubusercontent.com/26996883/196003569-578a35cb-600a-45cd-9a54-96194ea4c51d.mov
In many ways, it doesn't seem like it should be this hard to make template specific changes to aspects as basic as background color without changing the entire site.
## What is your proposed solution?
Part of this is due to the lack of clarity when saving in multi entity saving (https://github.com/WordPress/gutenberg/issues/29388) but the other piece of this is around the visibility of Styles combined with the difficult in making _template specific_ changes. Curious what others think and what can be done as right now it seems like multiple layers need to be considered across global styles, saving changes, etc.
|
non_defect
|
templates make it easier to do template specific style changes what problem does this address in the the following feedback came in that reflects a likely common action someone might want to take change the color scheme for a single template when i created a template for the “submission opportunities” category i thought i’d change the color scheme for this category from the styles menu in the sidebar the style is not just applied to this single category but to the whole site is that correct since i was creating a template for this category and the call for testing instructions encouraged me to push things a bit i thought i’d select a whole different color scheme just for this specific category expecting the style to be only applied to the category and not the whole site here s a quick video showing how it seems at first that changing styles does the trick without a clear warning that it ll change the whole site in many ways it doesn t seem like it should be this hard to make template specific changes to aspects as basic as background color without changing the entire site what is your proposed solution part of this is due to the lack of clarity when saving in multi entity saving but the other piece of this is around the visibility of styles combined with the difficult in making template specific changes curious what others think and what can be done as right now it seems like multiple layers need to be considered across global styles saving changes etc
| 0
|
232,408
| 25,578,514,872
|
IssuesEvent
|
2022-12-01 01:05:54
|
CodeChung/codechung
|
https://api.github.com/repos/CodeChung/codechung
|
opened
|
CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz
|
security vulnerability
|
## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- axios-5.9.2.tgz (Root Library)
- proxy-1.3.3.tgz
- http-proxy-middleware-0.19.1.tgz
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- source-map-resolve-0.5.3.tgz
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/CodeChung/codechung/commit/de1c2b0d1a23367b161c2d995029f9693bd8a155">de1c2b0d1a23367b161c2d995029f9693bd8a155</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-38900 (High) detected in decode-uri-component-0.2.0.tgz - ## CVE-2022-38900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>decode-uri-component-0.2.0.tgz</b></p></summary>
<p>A better decodeURIComponent</p>
<p>Library home page: <a href="https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz">https://registry.npmjs.org/decode-uri-component/-/decode-uri-component-0.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/decode-uri-component/package.json</p>
<p>
Dependency Hierarchy:
- axios-5.9.2.tgz (Root Library)
- proxy-1.3.3.tgz
- http-proxy-middleware-0.19.1.tgz
- micromatch-3.1.10.tgz
- snapdragon-0.8.2.tgz
- source-map-resolve-0.5.3.tgz
- :x: **decode-uri-component-0.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/CodeChung/codechung/commit/de1c2b0d1a23367b161c2d995029f9693bd8a155">de1c2b0d1a23367b161c2d995029f9693bd8a155</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
decode-uri-component 0.2.0 is vulnerable to Improper Input Validation resulting in DoS.
<p>Publish Date: 2022-11-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-38900>CVE-2022-38900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in decode uri component tgz cve high severity vulnerability vulnerable library decode uri component tgz a better decodeuricomponent library home page a href path to dependency file package json path to vulnerable library node modules decode uri component package json dependency hierarchy axios tgz root library proxy tgz http proxy middleware tgz micromatch tgz snapdragon tgz source map resolve tgz x decode uri component tgz vulnerable library found in head commit a href found in base branch master vulnerability details decode uri component is vulnerable to improper input validation resulting in dos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
187,939
| 22,046,055,583
|
IssuesEvent
|
2022-05-30 01:55:29
|
artsking/linux-4.1.15
|
https://api.github.com/repos/artsking/linux-4.1.15
|
closed
|
CVE-2020-25656 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed
|
security vulnerability
|
## CVE-2020-25656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/keyboard.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/keyboard.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel. A use-after-free was found in the way the console subsystem was using ioctls KDGKBSENT and KDSKBSENT. A local user could use this flaw to get read memory access out of bounds. The highest threat from this vulnerability is to data confidentiality.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25656>CVE-2020-25656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gregkh/linux/commit/82e61c3909db51d91b9d3e2071557b6435018b80">https://github.com/gregkh/linux/commit/82e61c3909db51d91b9d3e2071557b6435018b80</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: 4.14.204, 4.19.155, 4.4.242, 4.9.242, 5.4.75, 5.9.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-25656 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2020-25656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/keyboard.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/keyboard.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel. A use-after-free was found in the way the console subsystem was using ioctls KDGKBSENT and KDSKBSENT. A local user could use this flaw to get read memory access out of bounds. The highest threat from this vulnerability is to data confidentiality.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25656>CVE-2020-25656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/gregkh/linux/commit/82e61c3909db51d91b9d3e2071557b6435018b80">https://github.com/gregkh/linux/commit/82e61c3909db51d91b9d3e2071557b6435018b80</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: 4.14.204, 4.19.155, 4.4.242, 4.9.242, 5.4.75, 5.9.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers tty vt keyboard c drivers tty vt keyboard c vulnerability details a flaw was found in the linux kernel a use after free was found in the way the console subsystem was using ioctls kdgkbsent and kdskbsent a local user could use this flaw to get read memory access out of bounds the highest threat from this vulnerability is to data confidentiality publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
15,405
| 5,956,115,111
|
IssuesEvent
|
2017-05-28 13:52:00
|
Framstag/libosmscout
|
https://api.github.com/repos/Framstag/libosmscout
|
closed
|
Compiler warnings in GenCoordDat.cpp
|
build enhancement help wanted
|
See https://ci.appveyor.com/project/Framstag/libosmscout/build/job/644tpk6kyk8u0q3h
```
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::FindDuplicateCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:108:16: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
Id oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::StoreCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:283:19: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
OSMId oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
```
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/44932709-compiler-warnings-in-gencoorddat-cpp?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
Compiler warnings in GenCoordDat.cpp - See https://ci.appveyor.com/project/Framstag/libosmscout/build/job/644tpk6kyk8u0q3h
```
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::FindDuplicateCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:108:16: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
Id oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp: In member function 'bool osmscout::CoordDataGenerator::StoreCoordinates(const osmscout::TypeConfig&, const osmscout::ImportParameter&, osmscout::Progress&, std::unordered_map<long long unsigned int, unsigned char>&) const':
C:/projects/libosmscout/libosmscout-import/src/osmscout/import/GenCoordDat.cpp:283:19: warning: unused variable 'oldUpperLimit' [-Wunused-variable]
OSMId oldUpperLimit=currentUpperLimit;
^~~~~~~~~~~~~
```
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/44932709-compiler-warnings-in-gencoorddat-cpp?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F22109906&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_defect
|
compiler warnings in gencoorddat cpp see c projects libosmscout libosmscout import src osmscout import gencoorddat cpp in member function bool osmscout coorddatagenerator findduplicatecoordinates const osmscout typeconfig const osmscout importparameter osmscout progress std unordered map const c projects libosmscout libosmscout import src osmscout import gencoorddat cpp warning unused variable oldupperlimit id oldupperlimit currentupperlimit c projects libosmscout libosmscout import src osmscout import gencoorddat cpp in member function bool osmscout coorddatagenerator storecoordinates const osmscout typeconfig const osmscout importparameter osmscout progress std unordered map const c projects libosmscout libosmscout import src osmscout import gencoorddat cpp warning unused variable oldupperlimit osmid oldupperlimit currentupperlimit want to back this issue we accept bounties via
| 0
|
347,401
| 31,162,196,950
|
IssuesEvent
|
2023-08-16 16:49:22
|
prysmaticlabs/prysm
|
https://api.github.com/repos/prysmaticlabs/prysm
|
opened
|
Enable deneb e2e test
|
E2E Tests Deneb
|
Typically engine API has backward compatibiliy. The issue Deneb e2e test is facing now is due same version by adding a new argument (NewPayloadV3 + eip4788). If Prysm uses NewPayloadV3 with eip4788 and geth doesnt. Prysm won’t be able to interop with geth. Currently geth only has eip4788 on lightclient’s devnet8 branch. This is only needed for deneb e2e tests.
To move forward, e2e changes will remain in a separate branch. All the other bigger refactors to block processing, sync, etc can be transformed into `develop`. Typically in every fork, e2e changes are always the last to be in anyway
|
1.0
|
Enable deneb e2e test - Typically engine API has backward compatibiliy. The issue Deneb e2e test is facing now is due same version by adding a new argument (NewPayloadV3 + eip4788). If Prysm uses NewPayloadV3 with eip4788 and geth doesnt. Prysm won’t be able to interop with geth. Currently geth only has eip4788 on lightclient’s devnet8 branch. This is only needed for deneb e2e tests.
To move forward, e2e changes will remain in a separate branch. All the other bigger refactors to block processing, sync, etc can be transformed into `develop`. Typically in every fork, e2e changes are always the last to be in anyway
|
non_defect
|
enable deneb test typically engine api has backward compatibiliy the issue deneb test is facing now is due same version by adding a new argument if prysm uses with and geth doesnt prysm won’t be able to interop with geth currently geth only has on lightclient’s branch this is only needed for deneb tests to move forward changes will remain in a separate branch all the other bigger refactors to block processing sync etc can be transformed into develop typically in every fork changes are always the last to be in anyway
| 0
|
59,821
| 17,023,256,827
|
IssuesEvent
|
2021-07-03 01:05:11
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
History window broken codepage
|
Component: merkaartor Priority: minor Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 1.38pm, Sunday, 1st June 2008]**
History window show way history info in not UTF-8 codepage.
|
1.0
|
History window broken codepage - **[Submitted to the original trac issue database at 1.38pm, Sunday, 1st June 2008]**
History window show way history info in not UTF-8 codepage.
|
defect
|
history window broken codepage history window show way history info in not utf codepage
| 1
|
81,629
| 31,158,034,602
|
IssuesEvent
|
2023-08-16 14:17:12
|
vector-im/element-x-ios
|
https://api.github.com/repos/vector-im/element-x-ios
|
opened
|
Its not possible to create a DM or invite a user with a full MXID on a new account
|
T-Defect
|
### Steps to reproduce
1. Sign in to EX on a new account (that hasn't accepted identity server terms) or use OIDC to create a new account.
2. Click the prominent Start Chat button.
3. Enter a full mxid to tell your colleague "wooo, it worked, I'm signed into Element"
### Outcome
#### What did you expect?
To be able to start a chat.
#### What happened instead?

```
2023-08-16T14:00:42.162374Z DEBUG root:send{homeserver="https://my.homeserver/" server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5, V1_6] config=RequestConfig { timeout: 30s } request_id="REQ-27" request_size="197 B" path="/_matrix/client/v3/createRoom" status=403 response_size="389 B"}: matrix_sdk::http_client: crates/matrix-sdk/src/http_client/mod.rs:208: Error while sending request: Api(Server(ClientApi(Error { status_code: 403, authenticate: None, body: Standard { kind: _Custom { errcode: "M_CONSENT_NOT_GIVEN", extra: Extra({"consent_uri": String("https://my.homeserver/_matrix/consent?u=myuserid&h=6757a23ece6b4347206dabe6eadd5cf652ac75db4ca10af5061ca1b83151ebd3")}) }, message: "To continue using this homeserver you must review and agree to our terms and conditions at https://my.homeserver/_matrix/consent?u= myuserid&h=6757a23ece6b4347206dabe6eadd5cf652ac75db4ca10af5061ca1b83151ebd3." } })))
```
### Your phone model
Simulator
### Operating system version
16.4
### Application version
1.2.4 `develop`
### Homeserver
element.io
### Will you send logs?
No
|
1.0
|
Its not possible to create a DM or invite a user with a full MXID on a new account - ### Steps to reproduce
1. Sign in to EX on a new account (that hasn't accepted identity server terms) or use OIDC to create a new account.
2. Click the prominent Start Chat button.
3. Enter a full mxid to tell your colleague "wooo, it worked, I'm signed into Element"
### Outcome
#### What did you expect?
To be able to start a chat.
#### What happened instead?

```
2023-08-16T14:00:42.162374Z DEBUG root:send{homeserver="https://my.homeserver/" server_versions=[V1_0, V1_1, V1_2, V1_3, V1_4, V1_5, V1_6] config=RequestConfig { timeout: 30s } request_id="REQ-27" request_size="197 B" path="/_matrix/client/v3/createRoom" status=403 response_size="389 B"}: matrix_sdk::http_client: crates/matrix-sdk/src/http_client/mod.rs:208: Error while sending request: Api(Server(ClientApi(Error { status_code: 403, authenticate: None, body: Standard { kind: _Custom { errcode: "M_CONSENT_NOT_GIVEN", extra: Extra({"consent_uri": String("https://my.homeserver/_matrix/consent?u=myuserid&h=6757a23ece6b4347206dabe6eadd5cf652ac75db4ca10af5061ca1b83151ebd3")}) }, message: "To continue using this homeserver you must review and agree to our terms and conditions at https://my.homeserver/_matrix/consent?u= myuserid&h=6757a23ece6b4347206dabe6eadd5cf652ac75db4ca10af5061ca1b83151ebd3." } })))
```
### Your phone model
Simulator
### Operating system version
16.4
### Application version
1.2.4 `develop`
### Homeserver
element.io
### Will you send logs?
No
|
defect
|
its not possible to create a dm or invite a user with a full mxid on a new account steps to reproduce sign in to ex on a new account that hasn t accepted identity server terms or use oidc to create a new account click the prominent start chat button enter a full mxid to tell your colleague wooo it worked i m signed into element outcome what did you expect to be able to start a chat what happened instead debug root send homeserver server versions config requestconfig timeout request id req request size b path matrix client createroom status response size b matrix sdk http client crates matrix sdk src http client mod rs error while sending request api server clientapi error status code authenticate none body standard kind custom errcode m consent not given extra extra consent uri string message to continue using this homeserver you must review and agree to our terms and conditions at myuserid h your phone model simulator operating system version application version develop homeserver element io will you send logs no
| 1
|
71,623
| 23,729,826,247
|
IssuesEvent
|
2022-08-31 00:01:54
|
Gogo1951/Groupie
|
https://api.github.com/repos/Gogo1951/Groupie
|
closed
|
Ensure Warcraft Logs Link works for Servers with Multiple-Word-Names
|
Priority - 4 Low Type - Defect
|

The space kills the URL.
This will happen on real servers too.
VERY LOW prio to fix this at this time.

|
1.0
|
Ensure Warcraft Logs Link works for Servers with Multiple-Word-Names - 
The space kills the URL.
This will happen on real servers too.
VERY LOW prio to fix this at this time.

|
defect
|
ensure warcraft logs link works for servers with multiple word names the space kills the url this will happen on real servers too very low prio to fix this at this time
| 1
|
70,347
| 23,131,922,819
|
IssuesEvent
|
2022-07-28 11:10:38
|
vector-im/element-android
|
https://api.github.com/repos/vector-im/element-android
|
closed
|
`Show all rooms in Home` preference has no effect
|
T-Defect A-Spaces S-Major O-Frequent A-Settings
|
### Steps to reproduce
1. Have rooms in spaces
2. In settings -> preferences disable `Show all rooms in Home`
3. Open the home space, expect to NOT see the rooms from other spaces
### Outcome
#### What did you expect?
For rooms within other spaces to be filtered out from home when the preference is disabled
#### What happened instead?
Element android
| PREFERENCE | SPACE | ROOM LIST | OTHER SPACE |
| --- | --- | --- | --- |
|||
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
_No response_
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
Yes
|
1.0
|
`Show all rooms in Home` preference has no effect - ### Steps to reproduce
1. Have rooms in spaces
2. In settings -> preferences disable `Show all rooms in Home`
3. Open the home space, expect to NOT see the rooms from other spaces
### Outcome
#### What did you expect?
For rooms within other spaces to be filtered out from home when the preference is disabled
#### What happened instead?
Element android
| PREFERENCE | SPACE | ROOM LIST | OTHER SPACE |
| --- | --- | --- | --- |
|||
### Your phone model
_No response_
### Operating system version
_No response_
### Application version and app store
_No response_
### Homeserver
_No response_
### Will you send logs?
No
### Are you willing to provide a PR?
Yes
|
defect
|
show all rooms in home preference has no effect steps to reproduce have rooms in spaces in settings preferences disable show all rooms in home open the home space expect to not see the rooms from other spaces outcome what did you expect for rooms within other spaces to be filtered out from home when the preference is disabled what happened instead element android preference space room list other space your phone model no response operating system version no response application version and app store no response homeserver no response will you send logs no are you willing to provide a pr yes
| 1
|
190,923
| 22,173,483,766
|
IssuesEvent
|
2022-06-06 05:19:36
|
Satheesh575555/linux-4.19.72
|
https://api.github.com/repos/Satheesh575555/linux-4.19.72
|
reopened
|
CVE-2021-30002 (Medium) detected in linuxlinux-4.19.236
|
security vulnerability
|
## CVE-2021-30002 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.19.72/commit/ca82789c9f44a15d0b5166020b5c08fc8685cb69">ca82789c9f44a15d0b5166020b5c08fc8685cb69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/v4l2-core/v4l2-ioctl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/v4l2-core/v4l2-ioctl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.11.3 when a webcam device exists. video_usercopy in drivers/media/v4l2-core/v4l2-ioctl.c has a memory leak for large arguments, aka CID-fb18802a338b.
<p>Publish Date: 2021-04-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-30002>CVE-2021-30002</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-30002">https://www.linuxkernelcves.com/cves/CVE-2021-30002</a></p>
<p>Release Date: 2021-04-02</p>
<p>Fix Resolution: v4.4.260, v4.9.260, v4.14.224, v4.19.179, v5.4.103, v5.10.21, v5.11.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-30002 (Medium) detected in linuxlinux-4.19.236 - ## CVE-2021-30002 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.19.72/commit/ca82789c9f44a15d0b5166020b5c08fc8685cb69">ca82789c9f44a15d0b5166020b5c08fc8685cb69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/v4l2-core/v4l2-ioctl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/v4l2-core/v4l2-ioctl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.11.3 when a webcam device exists. video_usercopy in drivers/media/v4l2-core/v4l2-ioctl.c has a memory leak for large arguments, aka CID-fb18802a338b.
<p>Publish Date: 2021-04-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-30002>CVE-2021-30002</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-30002">https://www.linuxkernelcves.com/cves/CVE-2021-30002</a></p>
<p>Release Date: 2021-04-02</p>
<p>Fix Resolution: v4.4.260, v4.9.260, v4.14.224, v4.19.179, v5.4.103, v5.10.21, v5.11.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers media core ioctl c drivers media core ioctl c vulnerability details an issue was discovered in the linux kernel before when a webcam device exists video usercopy in drivers media core ioctl c has a memory leak for large arguments aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
50,148
| 13,187,348,413
|
IssuesEvent
|
2020-08-13 03:07:39
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
I3Time does not correctly add and subtract across year boundries (Trac #204)
|
Migrated from Trac dataclasses defect
|
MAX_DAQTIME in I3Time.h is incorrect and so adding and subtracting across year boundries does not work. Unit test "year_transitions" added to I3TimeTest.cxx.
In [2]: t1=dataclasses.I3Time()
In [3]: t2=dataclasses.I3Time()
In [4]: t1.SetUTCCalDate(2007,12,31,23,59,59,0)
In [5]: t2.SetUTCCalDate(2008,1,1,0,0,0,0)
In [6]: ( t1+icetray.I3Units.s ) == t2
Out[6]: False
In [7]: t1
Out[7]: I3Time(2007,315359990000000000)
In [8]: t2
Out[8]: I3Time(2008,0)
In [9]: print t1
2007-12-31 23:59:59.000,000,000,0 UTC
In [10]: print t2
2008-01-01 00:00:00.000,000,000,0 UTC
In [11]: print t1+icetray.I3Units.s
2008-01-01 00:00:00.000,000,000,0 UTC
In [12]: t1+icetray.I3Units.s
Out[12]: I3Time(2007,315360000000000000)
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/204
, reported by kjmeagher and owned by blaufuss_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2010-06-02T16:24:15",
"description": "MAX_DAQTIME in I3Time.h is incorrect and so adding and subtracting across year boundries does not work. Unit test \"year_transitions\" added to I3TimeTest.cxx.\n\n\nIn [2]: t1=dataclasses.I3Time()\n\nIn [3]: t2=dataclasses.I3Time()\n\nIn [4]: t1.SetUTCCalDate(2007,12,31,23,59,59,0)\n\nIn [5]: t2.SetUTCCalDate(2008,1,1,0,0,0,0)\n\nIn [6]: ( t1+icetray.I3Units.s ) == t2\n\nOut[6]: False\n\nIn [7]: t1\n\nOut[7]: I3Time(2007,315359990000000000)\n\nIn [8]: t2\n\nOut[8]: I3Time(2008,0)\n\nIn [9]: print t1\n\n2007-12-31 23:59:59.000,000,000,0 UTC\n\nIn [10]: print t2\n\n2008-01-01 00:00:00.000,000,000,0 UTC\n\nIn [11]: print t1+icetray.I3Units.s\n\n2008-01-01 00:00:00.000,000,000,0 UTC\n\nIn [12]: t1+icetray.I3Units.s\n\nOut[12]: I3Time(2007,315360000000000000)\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1275495855000000",
"component": "dataclasses",
"summary": "I3Time does not correctly add and subtract across year boundries",
"priority": "normal",
"keywords": "",
"time": "2010-04-14T15:13:51",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
I3Time does not correctly add and subtract across year boundries (Trac #204) - MAX_DAQTIME in I3Time.h is incorrect and so adding and subtracting across year boundries does not work. Unit test "year_transitions" added to I3TimeTest.cxx.
In [2]: t1=dataclasses.I3Time()
In [3]: t2=dataclasses.I3Time()
In [4]: t1.SetUTCCalDate(2007,12,31,23,59,59,0)
In [5]: t2.SetUTCCalDate(2008,1,1,0,0,0,0)
In [6]: ( t1+icetray.I3Units.s ) == t2
Out[6]: False
In [7]: t1
Out[7]: I3Time(2007,315359990000000000)
In [8]: t2
Out[8]: I3Time(2008,0)
In [9]: print t1
2007-12-31 23:59:59.000,000,000,0 UTC
In [10]: print t2
2008-01-01 00:00:00.000,000,000,0 UTC
In [11]: print t1+icetray.I3Units.s
2008-01-01 00:00:00.000,000,000,0 UTC
In [12]: t1+icetray.I3Units.s
Out[12]: I3Time(2007,315360000000000000)
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/204
, reported by kjmeagher and owned by blaufuss_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2010-06-02T16:24:15",
"description": "MAX_DAQTIME in I3Time.h is incorrect and so adding and subtracting across year boundries does not work. Unit test \"year_transitions\" added to I3TimeTest.cxx.\n\n\nIn [2]: t1=dataclasses.I3Time()\n\nIn [3]: t2=dataclasses.I3Time()\n\nIn [4]: t1.SetUTCCalDate(2007,12,31,23,59,59,0)\n\nIn [5]: t2.SetUTCCalDate(2008,1,1,0,0,0,0)\n\nIn [6]: ( t1+icetray.I3Units.s ) == t2\n\nOut[6]: False\n\nIn [7]: t1\n\nOut[7]: I3Time(2007,315359990000000000)\n\nIn [8]: t2\n\nOut[8]: I3Time(2008,0)\n\nIn [9]: print t1\n\n2007-12-31 23:59:59.000,000,000,0 UTC\n\nIn [10]: print t2\n\n2008-01-01 00:00:00.000,000,000,0 UTC\n\nIn [11]: print t1+icetray.I3Units.s\n\n2008-01-01 00:00:00.000,000,000,0 UTC\n\nIn [12]: t1+icetray.I3Units.s\n\nOut[12]: I3Time(2007,315360000000000000)\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1275495855000000",
"component": "dataclasses",
"summary": "I3Time does not correctly add and subtract across year boundries",
"priority": "normal",
"keywords": "",
"time": "2010-04-14T15:13:51",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
defect
|
does not correctly add and subtract across year boundries trac max daqtime in h is incorrect and so adding and subtracting across year boundries does not work unit test year transitions added to cxx in dataclasses in dataclasses in setutccaldate in setutccaldate in icetray s out false in out in out in print utc in print utc in print icetray s utc in icetray s out migrated from reported by kjmeagher and owned by blaufuss json status closed changetime description max daqtime in h is incorrect and so adding and subtracting across year boundries does not work unit test year transitions added to cxx n n nin dataclasses n nin dataclasses n nin setutccaldate n nin setutccaldate n nin icetray s n nout false n nin n nout n nin n nout n nin print n utc n nin print n utc n nin print icetray s n utc n nin icetray s n nout n reporter kjmeagher cc resolution fixed ts component dataclasses summary does not correctly add and subtract across year boundries priority normal keywords time milestone owner blaufuss type defect
| 1
|
324,589
| 27,812,005,859
|
IssuesEvent
|
2023-03-18 08:14:35
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix elementwise.test_exp2
|
Sub Task Ivy API Experimental Failing Test
|
| | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
|
1.0
|
Fix elementwise.test_exp2 - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4453070356/jobs/7821287461" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_experimental/test_core/test_elementwise.py::test_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-03-18T03:27:54.0813107Z E TypeError: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0813676Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0813979Z E
2023-03-18T03:27:54.0814176Z E Invoked with:
2023-03-18T03:27:54.0820417Z E ivy.utils.exceptions.IvyBackendException: jax: to_numpy: (): incompatible function arguments. The following argument types are supported:
2023-03-18T03:27:54.0821017Z E 1. (self: xla::PyBuffer::pyobject, dtype: object = None, context: object = None) -> object
2023-03-18T03:27:54.0821316Z E
2023-03-18T03:27:54.0821510Z E Invoked with:
2023-03-18T03:27:54.0821747Z E Falsifying example: test_exp2(
2023-03-18T03:27:54.0822115Z E dtype_and_x=(['bfloat16'], [array([-1], dtype=bfloat16)]),
2023-03-18T03:27:54.0822416Z E test_flags=FunctionTestFlags(
2023-03-18T03:27:54.0822675Z E num_positional_args=1,
2023-03-18T03:27:54.0822907Z E with_out=False,
2023-03-18T03:27:54.0823142Z E instance_method=False,
2023-03-18T03:27:54.0823381Z E test_gradients=False,
2023-03-18T03:27:54.0823617Z E test_compile=False,
2023-03-18T03:27:54.0823854Z E as_variable=[False],
2023-03-18T03:27:54.0824095Z E native_arrays=[False],
2023-03-18T03:27:54.0824331Z E container=[False],
2023-03-18T03:27:54.0824545Z E ),
2023-03-18T03:27:54.0824787Z E fn_name='exp2',
2023-03-18T03:27:54.0825089Z E ground_truth_backend='tensorflow',
2023-03-18T03:27:54.0825637Z E backend_fw=<module 'ivy.functional.backends.jax' from '/ivy/ivy/functional/backends/jax/__init__.py'>,
2023-03-18T03:27:54.0826018Z E on_device='cpu',
2023-03-18T03:27:54.0826217Z E )
2023-03-18T03:27:54.0826394Z E
2023-03-18T03:27:54.0826916Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2RkAAMoBacZAABkAAU=') as a decorator on your test case
</details>
|
non_defect
|
fix elementwise test tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test functional test experimental test core test elementwise py test e typeerror incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e ivy utils exceptions ivybackendexception jax to numpy incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e falsifying example test e dtype and x dtype e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile false e as variable e native arrays e container e e fn name e ground truth backend tensorflow e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test functional test experimental test core test elementwise py test e typeerror incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e ivy utils exceptions ivybackendexception jax to numpy incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e falsifying example test e dtype and x dtype e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile false e as variable e native arrays e container e e fn name e ground truth backend tensorflow e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test functional test experimental test core test elementwise py test e typeerror incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e ivy utils exceptions ivybackendexception jax to numpy incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e falsifying example test e dtype and x dtype e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile false e as variable e native arrays e container e e fn name e ground truth backend tensorflow e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test functional test experimental test core test elementwise py test e typeerror incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e ivy utils exceptions ivybackendexception jax to numpy incompatible function arguments the following argument types are supported e self xla pybuffer pyobject dtype object none context object none object e e invoked with e falsifying example test e dtype and x dtype e test flags functiontestflags e num positional args e with out false e instance method false e test gradients false e test compile false e as variable e native arrays e container e e fn name e ground truth backend tensorflow e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
| 0
|
74,413
| 25,118,708,458
|
IssuesEvent
|
2022-11-09 05:48:08
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Homeserver URL does not appear to be a valid Matrix homeserver
|
T-Defect
|
### Steps to reproduce
Hello, I set up a server on Debian\nginx in the local network, if I go through the web interface (https://matrix.test.rus/), then I get to the server without problems, if the desktop client Element or some other is used, I get an error:
"URL does not appear to be a valid Matrix homeserver"
What could be the problem ?
My Nginx config:
server {
listen 443 ssl http2;
listen [::]:443 ssl http2;
listen 8448 ssl http2 default_server;
listen [::]:8448 ssl http2 default_server;
server_name matrix.test.rus;
ssl_certificate /etc/ssl/private/matrix.crt;
ssl_certificate_key /etc/ssl/private/matrix.key;
ssl_session_timeout 1d;
ssl_session_cache shared:SSL:50m;
ssl_session_tickets off;
ssl_protocols TLSv1.2 TLSv1.3;
ssl_prefer_server_ciphers on;
root /var/www/element;
index index.html;
location ~* ^(\/_matrix|\/_synapse\/client) {
proxy_pass http://127.0.0.1:8008;
proxy_set_header X-Forwarded-For $remote_addr;
proxy_set_header X-Forwarded-Proto $scheme;
proxy_set_header Host $host;
}
location /.well-known/matrix/client {
default_type application/json;
add_header Access-Control-Allow-Origin *;
return 200 '{ "m.homeserver": { "base_url": "https://matrix.test.rus:443" } }';
}
location /.well-known/matrix/server {
default_type application/json;
add_header Access-Control-Allow-Origin *;
return 200 '{"m.server": "matrix.test.rus:443"}';
}
}

### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
Debian 11
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
local matrix.test.rus
### Will you send logs?
Yes
|
1.0
|
Homeserver URL does not appear to be a valid Matrix homeserver - ### Steps to reproduce
Hello, I set up a server on Debian\nginx in the local network, if I go through the web interface (https://matrix.test.rus/), then I get to the server without problems, if the desktop client Element or some other is used, I get an error:
"URL does not appear to be a valid Matrix homeserver"
What could be the problem ?
My Nginx config:
server {
listen 443 ssl http2;
listen [::]:443 ssl http2;
listen 8448 ssl http2 default_server;
listen [::]:8448 ssl http2 default_server;
server_name matrix.test.rus;
ssl_certificate /etc/ssl/private/matrix.crt;
ssl_certificate_key /etc/ssl/private/matrix.key;
ssl_session_timeout 1d;
ssl_session_cache shared:SSL:50m;
ssl_session_tickets off;
ssl_protocols TLSv1.2 TLSv1.3;
ssl_prefer_server_ciphers on;
root /var/www/element;
index index.html;
location ~* ^(\/_matrix|\/_synapse\/client) {
proxy_pass http://127.0.0.1:8008;
proxy_set_header X-Forwarded-For $remote_addr;
proxy_set_header X-Forwarded-Proto $scheme;
proxy_set_header Host $host;
}
location /.well-known/matrix/client {
default_type application/json;
add_header Access-Control-Allow-Origin *;
return 200 '{ "m.homeserver": { "base_url": "https://matrix.test.rus:443" } }';
}
location /.well-known/matrix/server {
default_type application/json;
add_header Access-Control-Allow-Origin *;
return 200 '{"m.server": "matrix.test.rus:443"}';
}
}

### Outcome
#### What did you expect?
#### What happened instead?
### Operating system
Debian 11
### Application version
_No response_
### How did you install the app?
_No response_
### Homeserver
local matrix.test.rus
### Will you send logs?
Yes
|
defect
|
homeserver url does not appear to be a valid matrix homeserver steps to reproduce hello i set up a server on debian nginx in the local network if i go through the web interface then i get to the server without problems if the desktop client element or some other is used i get an error url does not appear to be a valid matrix homeserver what could be the problem my nginx config server listen ssl listen ssl listen ssl default server listen ssl default server server name matrix test rus ssl certificate etc ssl private matrix crt ssl certificate key etc ssl private matrix key ssl session timeout ssl session cache shared ssl ssl session tickets off ssl protocols ssl prefer server ciphers on root var www element index index html location matrix synapse client proxy pass proxy set header x forwarded for remote addr proxy set header x forwarded proto scheme proxy set header host host location well known matrix client default type application json add header access control allow origin return m homeserver base url location well known matrix server default type application json add header access control allow origin return m server matrix test rus outcome what did you expect what happened instead operating system debian application version no response how did you install the app no response homeserver local matrix test rus will you send logs yes
| 1
|
222,154
| 17,036,856,900
|
IssuesEvent
|
2021-07-05 08:19:47
|
handsontable/handsontable
|
https://api.github.com/repos/handsontable/handsontable
|
closed
|
New docs: A blink effect when entering the dark theme
|
Type: Bug Type: Documentation
|
### Description
<!--- Tell us what happens and what should happen -->
My browser remembers that I want to use the dark theme in the documentation. However, when I open the docs in a fresh tab, it "blinks" with a white theme first, and only then, after 200-300 ms, it switches to the dark theme. It looks bad.
### Steps to reproduce
<!--- Provide steps to reproduce this issue -->
1. Open new docs, choose a dark theme, close new docs
2. Open new docs in a new tab/window
3. See how it blinks at the first load
|
1.0
|
New docs: A blink effect when entering the dark theme - ### Description
<!--- Tell us what happens and what should happen -->
My browser remembers that I want to use the dark theme in the documentation. However, when I open the docs in a fresh tab, it "blinks" with a white theme first, and only then, after 200-300 ms, it switches to the dark theme. It looks bad.
### Steps to reproduce
<!--- Provide steps to reproduce this issue -->
1. Open new docs, choose a dark theme, close new docs
2. Open new docs in a new tab/window
3. See how it blinks at the first load
|
non_defect
|
new docs a blink effect when entering the dark theme description my browser remembers that i want to use the dark theme in the documentation however when i open the docs in a fresh tab it blinks with a white theme first and only then after ms it switches to the dark theme it looks bad steps to reproduce open new docs choose a dark theme close new docs open new docs in a new tab window see how it blinks at the first load
| 0
|
41,559
| 10,516,858,512
|
IssuesEvent
|
2019-09-28 20:45:07
|
fuzzdb-project/fuzzdb
|
https://api.github.com/repos/fuzzdb-project/fuzzdb
|
closed
|
JSON fuzzing
|
Priority-Medium Type-Defect auto-migrated
|
```
I made a quick dictionary that I think would be helpful when someone wants to
fuzz JSON structure.
```
Original issue reported on code.google.com by `arteau.o...@gmail.com` on 18 Feb 2015 at 7:16
Attachments:
- [fuzz_json.patch](https://storage.googleapis.com/google-code-attachments/fuzzdb/issue-127/comment-0/fuzz_json.patch)
|
1.0
|
JSON fuzzing - ```
I made a quick dictionary that I think would be helpful when someone wants to
fuzz JSON structure.
```
Original issue reported on code.google.com by `arteau.o...@gmail.com` on 18 Feb 2015 at 7:16
Attachments:
- [fuzz_json.patch](https://storage.googleapis.com/google-code-attachments/fuzzdb/issue-127/comment-0/fuzz_json.patch)
|
defect
|
json fuzzing i made a quick dictionary that i think would be helpful when someone wants to fuzz json structure original issue reported on code google com by arteau o gmail com on feb at attachments
| 1
|
57,938
| 11,810,577,767
|
IssuesEvent
|
2020-03-19 16:42:44
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
Replace local allocations with stack allocated spans
|
api-suggestion area-System.Runtime code-analyzer
|
Flag places where known small temporary arrays of primitives (e.g. with a small constant length / where the total size of `sizeof(T)*length` can be determined to be less than some threshold) not inside any loop and not passed around could be replaced by span-based `stackalloc`.
**Category**: Performance
|
1.0
|
Replace local allocations with stack allocated spans - Flag places where known small temporary arrays of primitives (e.g. with a small constant length / where the total size of `sizeof(T)*length` can be determined to be less than some threshold) not inside any loop and not passed around could be replaced by span-based `stackalloc`.
**Category**: Performance
|
non_defect
|
replace local allocations with stack allocated spans flag places where known small temporary arrays of primitives e g with a small constant length where the total size of sizeof t length can be determined to be less than some threshold not inside any loop and not passed around could be replaced by span based stackalloc category performance
| 0
|
233,170
| 18,951,902,154
|
IssuesEvent
|
2021-11-18 15:57:23
|
geoffhumphrey/brewcompetitiononlineentry
|
https://api.github.com/repos/geoffhumphrey/brewcompetitiononlineentry
|
opened
|
2.0.0 or Later - BJCP XML Report Repeats Entries
|
bug in latest master commit confirmed v2.3.2
|
Version: 2.0.0 or later
Description of Issue:
The BJCP is now rejecting XML reports submitted to them if there are multiple entries for a single name. Previously, the XML report separated Judges, Stewards, and Staff. If a person happened to serve as both a Judge and Staff member, for example, their information was repeated, as such:
```
<JudgeData>
<JudgeName>John Doe</JudgeName>
<JudgeID>A0000</JudgeID>
<JudgeRole>Judge</JudgeRole>
<JudgePts>1.0</JudgePts>
<NonJudgePts>0.0</NonJudgePts>
</JudgeData>
<JudgeData>
<JudgeName>John Doe</JudgeName>
<JudgeID>A0000</JudgeID>
<JudgeRole>Staff</JudgeRole>
<JudgePts>0.0</JudgePts>
<NonJudgePts>0.5</NonJudgePts>
</JudgeData>
```
With the fix in the latest master commit and in the next version, the roles will be combined into a single entry, like this:
```
<JudgeData>
<JudgeName>John Doe</JudgeName>
<JudgeID>A0000</JudgeID>
<JudgeRole>Staff + Judge</JudgeRole>
<JudgePts>1.0</JudgePts>
<NonJudgePts>0.5</NonJudgePts>
</JudgeData>
```
If a fix is needed for version 2.1.18 or later, download the attached zip, unpack, and upload the files to their respective sub-folders in your BCOE&M directory:
\includes\db\output_staff_points.db.php
\output\export\export.output.php
For enhancements, please prepend "Enhancement - " to the title. For issues/bugs, please prepend the BCOE&M version number the title
[bcoem_bjcp_xml_fix_2.1.18+.zip](https://github.com/geoffhumphrey/brewcompetitiononlineentry/files/7564111/bcoem_bjcp_xml_fix_2.1.18%2B.zip)
.
|
1.0
|
2.0.0 or Later - BJCP XML Report Repeats Entries - Version: 2.0.0 or later
Description of Issue:
The BJCP is now rejecting XML reports submitted to them if there are multiple entries for a single name. Previously, the XML report separated Judges, Stewards, and Staff. If a person happened to serve as both a Judge and Staff member, for example, their information was repeated, as such:
```
<JudgeData>
<JudgeName>John Doe</JudgeName>
<JudgeID>A0000</JudgeID>
<JudgeRole>Judge</JudgeRole>
<JudgePts>1.0</JudgePts>
<NonJudgePts>0.0</NonJudgePts>
</JudgeData>
<JudgeData>
<JudgeName>John Doe</JudgeName>
<JudgeID>A0000</JudgeID>
<JudgeRole>Staff</JudgeRole>
<JudgePts>0.0</JudgePts>
<NonJudgePts>0.5</NonJudgePts>
</JudgeData>
```
With the fix in the latest master commit and in the next version, the roles will be combined into a single entry, like this:
```
<JudgeData>
<JudgeName>John Doe</JudgeName>
<JudgeID>A0000</JudgeID>
<JudgeRole>Staff + Judge</JudgeRole>
<JudgePts>1.0</JudgePts>
<NonJudgePts>0.5</NonJudgePts>
</JudgeData>
```
If a fix is needed for version 2.1.18 or later, download the attached zip, unpack, and upload the files to their respective sub-folders in your BCOE&M directory:
\includes\db\output_staff_points.db.php
\output\export\export.output.php
For enhancements, please prepend "Enhancement - " to the title. For issues/bugs, please prepend the BCOE&M version number the title
[bcoem_bjcp_xml_fix_2.1.18+.zip](https://github.com/geoffhumphrey/brewcompetitiononlineentry/files/7564111/bcoem_bjcp_xml_fix_2.1.18%2B.zip)
.
|
non_defect
|
or later bjcp xml report repeats entries version or later description of issue the bjcp is now rejecting xml reports submitted to them if there are multiple entries for a single name previously the xml report separated judges stewards and staff if a person happened to serve as both a judge and staff member for example their information was repeated as such john doe judge john doe staff with the fix in the latest master commit and in the next version the roles will be combined into a single entry like this john doe staff judge if a fix is needed for version or later download the attached zip unpack and upload the files to their respective sub folders in your bcoe m directory includes db output staff points db php output export export output php for enhancements please prepend enhancement to the title for issues bugs please prepend the bcoe m version number the title
| 0
|
51,817
| 13,211,318,906
|
IssuesEvent
|
2020-08-15 22:16:46
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
opened
|
everything - busted include guards (Trac #1115)
|
Incomplete Migration Migrated from Trac defect other
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1115">https://code.icecube.wisc.edu/projects/icecube/ticket/1115</a>, reported by nega</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-12-09T16:41:34",
"_ts": "1449679294743515",
"description": "The following files have missing or broken include guards. Please add a comment listing the files you've fixed and I'll strike them from the list.\n\nclast/private/clast/converter/I3CLastFitParamsConverter.h\nclast/public/clast/I3CLastCalculator.h\nclsim/private/clsim/converter/I3MCHitConverterWithIDs.h\nclsim/private/clsim/converter/I3MCPEConverterWithIDs.h\nclsim/private/clsim/converter/I3PhotonConverter.h\nclsim/private/clsim/shadow/split_serializable_backport.h\nclsim/private/geant4/TrkCerenkov.hh\nclsim/private/geant4/TrkDetectorConstruction.hh\nclsim/private/geant4/TrkEventAction.hh\nclsim/private/geant4/TrkPrimaryGeneratorAction.hh\nclsim/private/geant4/TrkStackingAction.hh\nclsim/private/geant4/TrkUISessionToQueue.hh\nclsim/private/geant4/TrkUserEventInformation.hh\nclsim/private/pybindings/const_ptr_helpers.h\nclsim/private/pybindings/python_gil_holder.h\n~~cmake/tool-patches/boost-1.33.1/boost/foreach.hpp~~\n~~cmake/tool-patches/boost-1.33.1/boost/python/object/pointer_holder.hpp~~\ncmake/tool-patches/common/I3/intel_compatibility.h\ncramer-rao/private/cramer-rao/converter/I3CramerRaoParamsConverter.h\ncscd-llh/private/converter/I3CscdLlhFitParamsConverter.h\n~~dataclasses/private/pybindings/I3MCTreePhysicsLibraryWrappers.hh~~\n~~dataclasses/private/pybindings/I3Vectors/I3Vectors.h~~\n~~dataclasses/public/dataclasses/payload/I3SuperDSTTrigger.h~~\n~~dataclasses/public/dataclasses/physics/detail/I3MCTree_impl.h~~\n~~dipolefit/private/dipolefit/converter/I3DipoleFitParamsConverter.h~~\n~~!DomTools/public/DomTools/I3DistanceMap.h~~\ndouble-muon/private/test/TestFrame.h\n~~dst-extractor/public/dst-extractor/dst07/I3DSTExtractor07.h~~\n~~dst-extractor/public/dst-extractor/dst08/I3DSTExtractor08.h~~\n~~dst-extractor/public/dst-extractor/dst09/I3DSTExtractor09.h~~\n~~dst-extractor/public/dst-extractor/dst10/I3DSTExtractor10.h~~\n~~dst-extractor/public/dst-extractor/dst11/I3DSTExtractor11.h~~\n~~dst-extractor/public/dst-extractor/dst12/I3DSTExtractor12.h~~\n~~dst-extractor/public/dst-extractor/dst13/I3DSTExtractor13.h~~\n~~dst-extractor/public/dst-extractor/I3DSTExtractor.h~~\n~~fill-ratio/private/fill-ratio/converter/I3FillRatioInfoConverter.h~~\nfilter-tools/public/filter-tools/CreateFilterMask.h\nfilter-tools/public/filter-tools/FilterCheckModule.h\nfilter-tools/public/filter-tools/FilterMaskFilter.h\nfilter-tools/public/filter-tools/ParticleCheck.h\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopUserSteppingAction.h\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopUserTrackingAction.h\ng4-tankresponse/private/g4-tankresponse/triangle/triangle.h\ngulliver/private/gulliver/converter/I3LogLikelihoodFitParamsConverter.h\ngulliver/public/gulliver/I3PDFBase.h\ngulliver/public/gulliver/utilities/ieee754.h\nhdfwriter/private/hdfwriter/hdf5_opt.h\n~~healpix-cxx/public/healpix-cxx/cfitsio/compress.h\nhealpix-cxx/public/healpix-cxx/cfitsio/eval_defs.h\nhealpix-cxx/public/healpix-cxx/cfitsio/eval_tab.h\nhealpix-cxx/public/healpix-cxx/cfitsio/f77_wrap.h\nhealpix-cxx/public/healpix-cxx/cfitsio/fpack.h\nhealpix-cxx/public/healpix-cxx/cfitsio/group.h\nhealpix-cxx/public/healpix-cxx/cfitsio/region.h\nhealpix-cxx/public/healpix-cxx/cxxsupport/geom_utils.h~~\n~~!IceHive/private/test/TestHelpers.h~~\n~~icepick/public/icepick/I3UnFilter.h~~\nicetray/public/icetray/python/std_map_indexing_suite.hpp\n~~icetray/public/icetray/python/std_vector_indexing_suite.hpp~~\nicetray/public/icetray/test/ConstructorTest.h\n~~interfaces/private/pybindings/install.h~~\nlinefit/private/linefit/converter/I3LineFitParamsConverter.h\nmillipede/private/millipede/converter/MillipedeFitParamsConverter.h\nmue/private/mue/interpol.h\nmue/private/mue/llhreco.h\nmue/private/mue/miniball.h\nmue/private/mue/reader.h\nmue/private/muex/miniball.h\n!MuonGun/private/test/common.h\n~~neutrino-generator/private/test/TestUtils.h~~\nophelia/private/ophelia/converter/I3OpheliaFirstGuessTrackConverter.h\nparaboloid/private/paraboloid/I3ParaboloidFitParamsConverter.h\npayload-parsing/private/payload-parsing/I3DAQDataTrimmer.h\nphotonics-service/public/photonics-service/I3PhotonicsL2Reader.h\nphys-services/private/converter/I3CutValuesConverter.h\nphys-services/private/converter/I3EventInfoConverter.h\nphys-services/private/converter/I3RecoInfoConverter.h\nphys-services/private/geo-selector/Miniball.h\nphys-services/public/phys-services/geo-selector/I3GeoTrimmers.h\nportia/private/portia/converter/I3PortiaEventConverter.h\nppc/private/ppc/ppc.h\nshield/private/shield/converter/I3ShieldHitRecordConverter.h\nsimclasses/private/simclasses/converter/I3MCPulseListConverter.h\nsimclasses/private/simclasses/converter/I3MMCTrackListConverter.h\nsimclasses/private/simclasses/I3SequenceOpOStream.h\n~~tableio/private/pybindings/const_ptr_helpers.h~~\n~~tableio/private/pybindings/I3TableRow_detail.h~~\n~~tableio/private/tableio/converter/I3IndexColumnsGenerator.h~~\n~~tableio/private/tableio/converter/PythonConverter.h~~\n~~tableio/private/tableio/utils/I3ConverterBundle.h~~\n~~tableio/public/tableio/converter/I3FilterResultMapConverter.h~~\n~~tensor-of-inertia/private/tensor-of-inertia/converter/I3TensorOfInertiaFitParamsConverter.h~~\n~~tensor-of-inertia/public/tensor-of-inertia/I3TensorOfInertiaCalculator.h~~\ntoprec/private/toprec/converter/I3TopLateralFitParamsConverter.h\ntoprec/private/toprec/converter/I3TopRecoPlaneFitParamsConverter.h\ntoprec/private/toprec/converter/SnowCorrectionDiagnosticsConverter.h\ntoprec/public/toprec/I3TopLateralFitParams.h\ntpx/private/tpx/converter/convert_I3IceTopBaseline.h\n~~trigger-sim/public/trigger-sim/utilities/DetectorStatusUtils.h~~\n!WaveCalibrator/private/WaveCalibrator/I3WaveCalibrator.h\n",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2015-08-14T17:22:58",
"component": "other",
"summary": "everything - busted include guards",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
everything - busted include guards (Trac #1115) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1115">https://code.icecube.wisc.edu/projects/icecube/ticket/1115</a>, reported by nega</summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-12-09T16:41:34",
"_ts": "1449679294743515",
"description": "The following files have missing or broken include guards. Please add a comment listing the files you've fixed and I'll strike them from the list.\n\nclast/private/clast/converter/I3CLastFitParamsConverter.h\nclast/public/clast/I3CLastCalculator.h\nclsim/private/clsim/converter/I3MCHitConverterWithIDs.h\nclsim/private/clsim/converter/I3MCPEConverterWithIDs.h\nclsim/private/clsim/converter/I3PhotonConverter.h\nclsim/private/clsim/shadow/split_serializable_backport.h\nclsim/private/geant4/TrkCerenkov.hh\nclsim/private/geant4/TrkDetectorConstruction.hh\nclsim/private/geant4/TrkEventAction.hh\nclsim/private/geant4/TrkPrimaryGeneratorAction.hh\nclsim/private/geant4/TrkStackingAction.hh\nclsim/private/geant4/TrkUISessionToQueue.hh\nclsim/private/geant4/TrkUserEventInformation.hh\nclsim/private/pybindings/const_ptr_helpers.h\nclsim/private/pybindings/python_gil_holder.h\n~~cmake/tool-patches/boost-1.33.1/boost/foreach.hpp~~\n~~cmake/tool-patches/boost-1.33.1/boost/python/object/pointer_holder.hpp~~\ncmake/tool-patches/common/I3/intel_compatibility.h\ncramer-rao/private/cramer-rao/converter/I3CramerRaoParamsConverter.h\ncscd-llh/private/converter/I3CscdLlhFitParamsConverter.h\n~~dataclasses/private/pybindings/I3MCTreePhysicsLibraryWrappers.hh~~\n~~dataclasses/private/pybindings/I3Vectors/I3Vectors.h~~\n~~dataclasses/public/dataclasses/payload/I3SuperDSTTrigger.h~~\n~~dataclasses/public/dataclasses/physics/detail/I3MCTree_impl.h~~\n~~dipolefit/private/dipolefit/converter/I3DipoleFitParamsConverter.h~~\n~~!DomTools/public/DomTools/I3DistanceMap.h~~\ndouble-muon/private/test/TestFrame.h\n~~dst-extractor/public/dst-extractor/dst07/I3DSTExtractor07.h~~\n~~dst-extractor/public/dst-extractor/dst08/I3DSTExtractor08.h~~\n~~dst-extractor/public/dst-extractor/dst09/I3DSTExtractor09.h~~\n~~dst-extractor/public/dst-extractor/dst10/I3DSTExtractor10.h~~\n~~dst-extractor/public/dst-extractor/dst11/I3DSTExtractor11.h~~\n~~dst-extractor/public/dst-extractor/dst12/I3DSTExtractor12.h~~\n~~dst-extractor/public/dst-extractor/dst13/I3DSTExtractor13.h~~\n~~dst-extractor/public/dst-extractor/I3DSTExtractor.h~~\n~~fill-ratio/private/fill-ratio/converter/I3FillRatioInfoConverter.h~~\nfilter-tools/public/filter-tools/CreateFilterMask.h\nfilter-tools/public/filter-tools/FilterCheckModule.h\nfilter-tools/public/filter-tools/FilterMaskFilter.h\nfilter-tools/public/filter-tools/ParticleCheck.h\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopUserSteppingAction.h\ng4-tankresponse/private/g4-tankresponse/g4classes/G4IceTopUserTrackingAction.h\ng4-tankresponse/private/g4-tankresponse/triangle/triangle.h\ngulliver/private/gulliver/converter/I3LogLikelihoodFitParamsConverter.h\ngulliver/public/gulliver/I3PDFBase.h\ngulliver/public/gulliver/utilities/ieee754.h\nhdfwriter/private/hdfwriter/hdf5_opt.h\n~~healpix-cxx/public/healpix-cxx/cfitsio/compress.h\nhealpix-cxx/public/healpix-cxx/cfitsio/eval_defs.h\nhealpix-cxx/public/healpix-cxx/cfitsio/eval_tab.h\nhealpix-cxx/public/healpix-cxx/cfitsio/f77_wrap.h\nhealpix-cxx/public/healpix-cxx/cfitsio/fpack.h\nhealpix-cxx/public/healpix-cxx/cfitsio/group.h\nhealpix-cxx/public/healpix-cxx/cfitsio/region.h\nhealpix-cxx/public/healpix-cxx/cxxsupport/geom_utils.h~~\n~~!IceHive/private/test/TestHelpers.h~~\n~~icepick/public/icepick/I3UnFilter.h~~\nicetray/public/icetray/python/std_map_indexing_suite.hpp\n~~icetray/public/icetray/python/std_vector_indexing_suite.hpp~~\nicetray/public/icetray/test/ConstructorTest.h\n~~interfaces/private/pybindings/install.h~~\nlinefit/private/linefit/converter/I3LineFitParamsConverter.h\nmillipede/private/millipede/converter/MillipedeFitParamsConverter.h\nmue/private/mue/interpol.h\nmue/private/mue/llhreco.h\nmue/private/mue/miniball.h\nmue/private/mue/reader.h\nmue/private/muex/miniball.h\n!MuonGun/private/test/common.h\n~~neutrino-generator/private/test/TestUtils.h~~\nophelia/private/ophelia/converter/I3OpheliaFirstGuessTrackConverter.h\nparaboloid/private/paraboloid/I3ParaboloidFitParamsConverter.h\npayload-parsing/private/payload-parsing/I3DAQDataTrimmer.h\nphotonics-service/public/photonics-service/I3PhotonicsL2Reader.h\nphys-services/private/converter/I3CutValuesConverter.h\nphys-services/private/converter/I3EventInfoConverter.h\nphys-services/private/converter/I3RecoInfoConverter.h\nphys-services/private/geo-selector/Miniball.h\nphys-services/public/phys-services/geo-selector/I3GeoTrimmers.h\nportia/private/portia/converter/I3PortiaEventConverter.h\nppc/private/ppc/ppc.h\nshield/private/shield/converter/I3ShieldHitRecordConverter.h\nsimclasses/private/simclasses/converter/I3MCPulseListConverter.h\nsimclasses/private/simclasses/converter/I3MMCTrackListConverter.h\nsimclasses/private/simclasses/I3SequenceOpOStream.h\n~~tableio/private/pybindings/const_ptr_helpers.h~~\n~~tableio/private/pybindings/I3TableRow_detail.h~~\n~~tableio/private/tableio/converter/I3IndexColumnsGenerator.h~~\n~~tableio/private/tableio/converter/PythonConverter.h~~\n~~tableio/private/tableio/utils/I3ConverterBundle.h~~\n~~tableio/public/tableio/converter/I3FilterResultMapConverter.h~~\n~~tensor-of-inertia/private/tensor-of-inertia/converter/I3TensorOfInertiaFitParamsConverter.h~~\n~~tensor-of-inertia/public/tensor-of-inertia/I3TensorOfInertiaCalculator.h~~\ntoprec/private/toprec/converter/I3TopLateralFitParamsConverter.h\ntoprec/private/toprec/converter/I3TopRecoPlaneFitParamsConverter.h\ntoprec/private/toprec/converter/SnowCorrectionDiagnosticsConverter.h\ntoprec/public/toprec/I3TopLateralFitParams.h\ntpx/private/tpx/converter/convert_I3IceTopBaseline.h\n~~trigger-sim/public/trigger-sim/utilities/DetectorStatusUtils.h~~\n!WaveCalibrator/private/WaveCalibrator/I3WaveCalibrator.h\n",
"reporter": "nega",
"cc": "",
"resolution": "fixed",
"time": "2015-08-14T17:22:58",
"component": "other",
"summary": "everything - busted include guards",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
|
defect
|
everything busted include guards trac migrated from json status closed changetime ts description the following files have missing or broken include guards please add a comment listing the files you ve fixed and i ll strike them from the list n nclast private clast converter h nclast public clast h nclsim private clsim converter h nclsim private clsim converter h nclsim private clsim converter h nclsim private clsim shadow split serializable backport h nclsim private trkcerenkov hh nclsim private trkdetectorconstruction hh nclsim private trkeventaction hh nclsim private trkprimarygeneratoraction hh nclsim private trkstackingaction hh nclsim private trkuisessiontoqueue hh nclsim private trkusereventinformation hh nclsim private pybindings const ptr helpers h nclsim private pybindings python gil holder h n cmake tool patches boost boost foreach hpp n cmake tool patches boost boost python object pointer holder hpp ncmake tool patches common intel compatibility h ncramer rao private cramer rao converter h ncscd llh private converter h n dataclasses private pybindings hh n dataclasses private pybindings h n dataclasses public dataclasses payload h n dataclasses public dataclasses physics detail impl h n dipolefit private dipolefit converter h n domtools public domtools h ndouble muon private test testframe h n dst extractor public dst extractor h n dst extractor public dst extractor h n dst extractor public dst extractor h n dst extractor public dst extractor h n dst extractor public dst extractor h n dst extractor public dst extractor h n dst extractor public dst extractor h n dst extractor public dst extractor h n fill ratio private fill ratio converter h nfilter tools public filter tools createfiltermask h nfilter tools public filter tools filtercheckmodule h nfilter tools public filter tools filtermaskfilter h nfilter tools public filter tools particlecheck h tankresponse private tankresponse h tankresponse private tankresponse h tankresponse private tankresponse triangle triangle h ngulliver private gulliver converter h ngulliver public gulliver h ngulliver public gulliver utilities h nhdfwriter private hdfwriter opt h n healpix cxx public healpix cxx cfitsio compress h nhealpix cxx public healpix cxx cfitsio eval defs h nhealpix cxx public healpix cxx cfitsio eval tab h nhealpix cxx public healpix cxx cfitsio wrap h nhealpix cxx public healpix cxx cfitsio fpack h nhealpix cxx public healpix cxx cfitsio group h nhealpix cxx public healpix cxx cfitsio region h nhealpix cxx public healpix cxx cxxsupport geom utils h n icehive private test testhelpers h n icepick public icepick h nicetray public icetray python std map indexing suite hpp n icetray public icetray python std vector indexing suite hpp nicetray public icetray test constructortest h n interfaces private pybindings install h nlinefit private linefit converter h nmillipede private millipede converter millipedefitparamsconverter h nmue private mue interpol h nmue private mue llhreco h nmue private mue miniball h nmue private mue reader h nmue private muex miniball h n muongun private test common h n neutrino generator private test testutils h nophelia private ophelia converter h nparaboloid private paraboloid h npayload parsing private payload parsing h nphotonics service public photonics service h nphys services private converter h nphys services private converter h nphys services private converter h nphys services private geo selector miniball h nphys services public phys services geo selector h nportia private portia converter h nppc private ppc ppc h nshield private shield converter h nsimclasses private simclasses converter h nsimclasses private simclasses converter h nsimclasses private simclasses h n tableio private pybindings const ptr helpers h n tableio private pybindings detail h n tableio private tableio converter h n tableio private tableio converter pythonconverter h n tableio private tableio utils h n tableio public tableio converter h n tensor of inertia private tensor of inertia converter h n tensor of inertia public tensor of inertia h ntoprec private toprec converter h ntoprec private toprec converter h ntoprec private toprec converter snowcorrectiondiagnosticsconverter h ntoprec public toprec h ntpx private tpx converter convert h n trigger sim public trigger sim utilities detectorstatusutils h n wavecalibrator private wavecalibrator h n reporter nega cc resolution fixed time component other summary everything busted include guards priority normal keywords milestone owner type defect
| 1
|
82,032
| 31,864,475,961
|
IssuesEvent
|
2023-09-15 13:15:47
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Sender is shown as ID instead of user name in shared files of encrypted conversation (Mac and Windows)
|
T-Defect
|
### Steps to reproduce
I have found an issue when checking shared files in encrypted conversation.
From Element Windows or MacOS client, select any encrypted conversation --> Room info --> Files. All shared files are listed with sender's user ID instead of user name.
As soon as I open the uploaded files dialog the matrix ID is displayed instead of the displayname. Not only in the files dialog but also in the chat the message where the file was uploaded and sent comes from a user ID not a displayname.
### Outcome
#### What did you expect?
Expected correct displayname.
#### What happened instead?
See matrix id of users which have shared files.
### Operating system
Windows, MacOS
### Application version
1.11.43
### How did you install the app?
element.io/get-started
### Homeserver
matrix.org
### Will you send logs?
No
|
1.0
|
Sender is shown as ID instead of user name in shared files of encrypted conversation (Mac and Windows) - ### Steps to reproduce
I have found an issue when checking shared files in encrypted conversation.
From Element Windows or MacOS client, select any encrypted conversation --> Room info --> Files. All shared files are listed with sender's user ID instead of user name.
As soon as I open the uploaded files dialog the matrix ID is displayed instead of the displayname. Not only in the files dialog but also in the chat the message where the file was uploaded and sent comes from a user ID not a displayname.
### Outcome
#### What did you expect?
Expected correct displayname.
#### What happened instead?
See matrix id of users which have shared files.
### Operating system
Windows, MacOS
### Application version
1.11.43
### How did you install the app?
element.io/get-started
### Homeserver
matrix.org
### Will you send logs?
No
|
defect
|
sender is shown as id instead of user name in shared files of encrypted conversation mac and windows steps to reproduce i have found an issue when checking shared files in encrypted conversation from element windows or macos client select any encrypted conversation room info files all shared files are listed with sender s user id instead of user name as soon as i open the uploaded files dialog the matrix id is displayed instead of the displayname not only in the files dialog but also in the chat the message where the file was uploaded and sent comes from a user id not a displayname outcome what did you expect expected correct displayname what happened instead see matrix id of users which have shared files operating system windows macos application version how did you install the app element io get started homeserver matrix org will you send logs no
| 1
|
2,922
| 3,251,149,117
|
IssuesEvent
|
2015-10-19 08:08:29
|
piwik/piwik
|
https://api.github.com/repos/piwik/piwik
|
closed
|
Use Piwik to measure how Piwik app is used by the community (opt-in)
|
c: Usability Major Task
|
just a weird idea:
track yourself / track piwik usages
would be interesting
- to show the power of piwik e.g. for every firstime user
- to show what is possible if integrated in best way
- to compare different piwik users
- to learn what new featuers may be needed for tracking such complex sites like piwik backend
- to optimize usability
- ...
with event tracking which is on the track it should be possible to get great inside of what one have used piwik for even with that many onpage actions piwik has.
'''enable via a simple checkbox in settings
-> automatic: new website "My Piwik" and start tracking immediately'''
- everything should be configured/setup in best way
- events hast to be implemented for all piwik standard actions
- possibility to share the results with piwik developers
- ....
what do you think ??
|
True
|
Use Piwik to measure how Piwik app is used by the community (opt-in) - just a weird idea:
track yourself / track piwik usages
would be interesting
- to show the power of piwik e.g. for every firstime user
- to show what is possible if integrated in best way
- to compare different piwik users
- to learn what new featuers may be needed for tracking such complex sites like piwik backend
- to optimize usability
- ...
with event tracking which is on the track it should be possible to get great inside of what one have used piwik for even with that many onpage actions piwik has.
'''enable via a simple checkbox in settings
-> automatic: new website "My Piwik" and start tracking immediately'''
- everything should be configured/setup in best way
- events hast to be implemented for all piwik standard actions
- possibility to share the results with piwik developers
- ....
what do you think ??
|
non_defect
|
use piwik to measure how piwik app is used by the community opt in just a weird idea track yourself track piwik usages would be interesting to show the power of piwik e g for every firstime user to show what is possible if integrated in best way to compare different piwik users to learn what new featuers may be needed for tracking such complex sites like piwik backend to optimize usability with event tracking which is on the track it should be possible to get great inside of what one have used piwik for even with that many onpage actions piwik has enable via a simple checkbox in settings automatic new website my piwik and start tracking immediately everything should be configured setup in best way events hast to be implemented for all piwik standard actions possibility to share the results with piwik developers what do you think
| 0
|
30,972
| 8,634,833,819
|
IssuesEvent
|
2018-11-22 18:40:23
|
covexo/devspace
|
https://api.github.com/repos/covexo/devspace
|
closed
|
When building images Dockerfiles outside the context cause an error, yet this is supported from Docker 18.03 onwards
|
area/building-images kind/bug priority/longterm
|
<!-- Please use this template for reporting bugs and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**What happened?**
Given the following image definition, and docker daemon version 18.09.0, build 4d60db4
images:
example:
name: issue/example
skipPush: true
build:
dockerfilePath: docker/example/Dockerfile
contextPath: ../example
The image was not built, instead the following error is output:
[INFO] Building image 'issue/example' with engine 'docker'
[DONE] √ Authentication successful (hub.docker.com)
Sending build context to Docker daemon 4.14MB
[FATAL] Error building image: Error during image build: Error response from daemon: Forbidden path outside the build context: ../dev-system/docker/example/Dockerfile ()
**What did you expect to happen instead?**
When using the same docker version 18.09.0, build 4d60db4 from the command line:
“docker build -f docker/example/Dockerfile ../example
The command succeeded and the image was built.
**How can we reproduce the bug?** (as minimally and precisely as possible)
Create a devspace config where an image configuration contains a Dockerfile outside of the context for that image build.
**Local Environment:**
- Operating System: linux
- Deployment method: kubectl apply
**Kubernetes Cluster:**
- Cloud Provider: Local single node Microk8s
- Kubernetes Version: v1.12.2
**Anything else we need to know?**
Essential do our dev system/process as we need specific development docker images (e.g. no npm prune --production for example) and we need to keep these in a seperate repository.
<!-- DO NOT EDIT BELOW THIS LINE -->
/kind bug
|
1.0
|
When building images Dockerfiles outside the context cause an error, yet this is supported from Docker 18.03 onwards - <!-- Please use this template for reporting bugs and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!-->
**What happened?**
Given the following image definition, and docker daemon version 18.09.0, build 4d60db4
images:
example:
name: issue/example
skipPush: true
build:
dockerfilePath: docker/example/Dockerfile
contextPath: ../example
The image was not built, instead the following error is output:
[INFO] Building image 'issue/example' with engine 'docker'
[DONE] √ Authentication successful (hub.docker.com)
Sending build context to Docker daemon 4.14MB
[FATAL] Error building image: Error during image build: Error response from daemon: Forbidden path outside the build context: ../dev-system/docker/example/Dockerfile ()
**What did you expect to happen instead?**
When using the same docker version 18.09.0, build 4d60db4 from the command line:
“docker build -f docker/example/Dockerfile ../example
The command succeeded and the image was built.
**How can we reproduce the bug?** (as minimally and precisely as possible)
Create a devspace config where an image configuration contains a Dockerfile outside of the context for that image build.
**Local Environment:**
- Operating System: linux
- Deployment method: kubectl apply
**Kubernetes Cluster:**
- Cloud Provider: Local single node Microk8s
- Kubernetes Version: v1.12.2
**Anything else we need to know?**
Essential do our dev system/process as we need specific development docker images (e.g. no npm prune --production for example) and we need to keep these in a seperate repository.
<!-- DO NOT EDIT BELOW THIS LINE -->
/kind bug
|
non_defect
|
when building images dockerfiles outside the context cause an error yet this is supported from docker onwards what happened given the following image definition and docker daemon version build images example name issue example skippush true build dockerfilepath docker example dockerfile contextpath example the image was not built instead the following error is output building image issue example with engine docker √ authentication successful hub docker com sending build context to docker daemon error building image error during image build error response from daemon forbidden path outside the build context dev system docker example dockerfile what did you expect to happen instead when using the same docker version build from the command line “docker build f docker example dockerfile example the command succeeded and the image was built how can we reproduce the bug as minimally and precisely as possible create a devspace config where an image configuration contains a dockerfile outside of the context for that image build local environment operating system linux deployment method kubectl apply kubernetes cluster cloud provider local single node kubernetes version anything else we need to know essential do our dev system process as we need specific development docker images e g no npm prune production for example and we need to keep these in a seperate repository kind bug
| 0
|
44,159
| 12,015,214,148
|
IssuesEvent
|
2020-04-10 13:29:05
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
[COGNITION] CONSIDER adding a heading before "Longer than 1 year"
|
508-defect-4 508-issue-cognition 508/Accessibility vsa vsa-public-websites
|
# [508-defect-4](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-4)
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
On Fiduciary claims — [https://preview-prod.vfs.va.gov/preview?nodeId=3016](https://preview-prod.vfs.va.gov/preview?nodeId=3016), **consider** adding a heading before "Longer than 1 year", as without the context it may be confusing for screen reader users who don't have the context of the visual design.
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to understand the context of the content as sighted users gain from visual design.
## Environment
* Operating System: all
* Browser: all
* Screenreading device: all
* Server destination: devshop?/prod?
## Steps to Recreate
1. Enter https://preview-prod.vfs.va.gov/preview?nodeId=3016 in browser, in the VA environment
2. Start screenreading device listed in Environment
3. Navigate to the section "Appeal to a Veterans Law Judge by requesting a Board Appeal"
4. Notice the first "About 1 year" has a heading "How long it take VA to make a decision?"
5. Verify the second "About 1 year" doesn't have the heading, and sounds abrupt and confusing.
## Possible Fixes (optional)
Add "How long it take VA to make a decision?" before the second instance of "About 1 year".
## WCAG or Vendor Guidance (optional)
* [Web Accessibility In Context](https://www.smashingmagazine.com/2019/06/web-accessibility-context/)
* [Designing For Accessibility And Inclusion](https://www.smashingmagazine.com/2018/04/designing-accessibility-inclusion/)
## Screenshots or Trace Logs
https://user-images.githubusercontent.com/57469/78830442-f4b89380-79b5-11ea-8952-33e6aec71c60.png
|
1.0
|
[COGNITION] CONSIDER adding a heading before "Longer than 1 year" - # [508-defect-4](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-4)
**Feedback framework**
- **❗️ Must** for if the feedback must be applied
- **⚠️Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Description
On Fiduciary claims — [https://preview-prod.vfs.va.gov/preview?nodeId=3016](https://preview-prod.vfs.va.gov/preview?nodeId=3016), **consider** adding a heading before "Longer than 1 year", as without the context it may be confusing for screen reader users who don't have the context of the visual design.
## Point of Contact
**VFS Point of Contact:** Jennifer
## Acceptance Criteria
As a screen reader user, I want to understand the context of the content as sighted users gain from visual design.
## Environment
* Operating System: all
* Browser: all
* Screenreading device: all
* Server destination: devshop?/prod?
## Steps to Recreate
1. Enter https://preview-prod.vfs.va.gov/preview?nodeId=3016 in browser, in the VA environment
2. Start screenreading device listed in Environment
3. Navigate to the section "Appeal to a Veterans Law Judge by requesting a Board Appeal"
4. Notice the first "About 1 year" has a heading "How long it take VA to make a decision?"
5. Verify the second "About 1 year" doesn't have the heading, and sounds abrupt and confusing.
## Possible Fixes (optional)
Add "How long it take VA to make a decision?" before the second instance of "About 1 year".
## WCAG or Vendor Guidance (optional)
* [Web Accessibility In Context](https://www.smashingmagazine.com/2019/06/web-accessibility-context/)
* [Designing For Accessibility And Inclusion](https://www.smashingmagazine.com/2018/04/designing-accessibility-inclusion/)
## Screenshots or Trace Logs
https://user-images.githubusercontent.com/57469/78830442-f4b89380-79b5-11ea-8952-33e6aec71c60.png
|
defect
|
consider adding a heading before longer than year feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description on fiduciary claims — consider adding a heading before longer than year as without the context it may be confusing for screen reader users who don t have the context of the visual design point of contact vfs point of contact jennifer acceptance criteria as a screen reader user i want to understand the context of the content as sighted users gain from visual design environment operating system all browser all screenreading device all server destination devshop prod steps to recreate enter in browser in the va environment start screenreading device listed in environment navigate to the section appeal to a veterans law judge by requesting a board appeal notice the first about year has a heading how long it take va to make a decision verify the second about year doesn t have the heading and sounds abrupt and confusing possible fixes optional add how long it take va to make a decision before the second instance of about year wcag or vendor guidance optional screenshots or trace logs
| 1
|
82,077
| 7,812,348,498
|
IssuesEvent
|
2018-06-12 13:09:10
|
philharmonic77/Blog-s-Comment
|
https://api.github.com/repos/philharmonic77/Blog-s-Comment
|
closed
|
聚类分析在有监督学习里的应用
|
gitment test.md
|
https://phil77.netlify.com/post/test/
先聚类再分类 参考文献:The Utility of Clustering in Prediction Tasks 简单数据集测试 数据集 kaggle-Titanic 先做简单的特征提取,得到训练数据集。
|
1.0
|
聚类分析在有监督学习里的应用 - https://phil77.netlify.com/post/test/
先聚类再分类 参考文献:The Utility of Clustering in Prediction Tasks 简单数据集测试 数据集 kaggle-Titanic 先做简单的特征提取,得到训练数据集。
|
non_defect
|
聚类分析在有监督学习里的应用 先聚类再分类 参考文献:the utility of clustering in prediction tasks 简单数据集测试 数据集 kaggle titanic 先做简单的特征提取,得到训练数据集。
| 0
|
118,030
| 9,968,483,883
|
IssuesEvent
|
2019-07-08 15:41:03
|
golang/go
|
https://api.github.com/repos/golang/go
|
opened
|
crypto/tls: TestVerifyHostnameResumed consistently failing in longtest builder
|
Soon Testing release-blocker
|
`crypto/tls.TestVerifyHostnameResumed` is consistently failing in the `linux-amd64-longtest` builder.
First failure was at CL 184099, which appears to be unrelated. That suggests some sort of change in a non-hermetic dependency, or perhaps in the builder itself.
Example: https://build.golang.org/log/67d2b12bec6bf70eb818bb3246aee32990ecd9e6
```
--- FAIL: TestVerifyHostnameResumed (0.10s)
--- FAIL: TestVerifyHostnameResumed/TLSv12 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
--- FAIL: TestVerifyHostnameResumed/TLSv13 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
FAIL
FAIL crypto/tls 0.828s
```
CC @agl
|
1.0
|
crypto/tls: TestVerifyHostnameResumed consistently failing in longtest builder - `crypto/tls.TestVerifyHostnameResumed` is consistently failing in the `linux-amd64-longtest` builder.
First failure was at CL 184099, which appears to be unrelated. That suggests some sort of change in a non-hermetic dependency, or perhaps in the builder itself.
Example: https://build.golang.org/log/67d2b12bec6bf70eb818bb3246aee32990ecd9e6
```
--- FAIL: TestVerifyHostnameResumed (0.10s)
--- FAIL: TestVerifyHostnameResumed/TLSv12 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
--- FAIL: TestVerifyHostnameResumed/TLSv13 (0.05s)
tls_test.go:383: Subsequent connection unexpectedly didn't resume
FAIL
FAIL crypto/tls 0.828s
```
CC @agl
|
non_defect
|
crypto tls testverifyhostnameresumed consistently failing in longtest builder crypto tls testverifyhostnameresumed is consistently failing in the linux longtest builder first failure was at cl which appears to be unrelated that suggests some sort of change in a non hermetic dependency or perhaps in the builder itself example fail testverifyhostnameresumed fail testverifyhostnameresumed tls test go subsequent connection unexpectedly didn t resume fail testverifyhostnameresumed tls test go subsequent connection unexpectedly didn t resume fail fail crypto tls cc agl
| 0
|
13,581
| 2,770,884,468
|
IssuesEvent
|
2015-05-01 17:43:59
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
Old values of Postprocessors aren't right in first timestep
|
C: MOOSE P: normal T: defect
|
If you execute a Postprocessor on `initial` you would expect that the _old_ value of that Postprocessor during the first timestep would be the value you got during `initial`... but that's currently not the case.
I have a fix for this.
|
1.0
|
Old values of Postprocessors aren't right in first timestep - If you execute a Postprocessor on `initial` you would expect that the _old_ value of that Postprocessor during the first timestep would be the value you got during `initial`... but that's currently not the case.
I have a fix for this.
|
defect
|
old values of postprocessors aren t right in first timestep if you execute a postprocessor on initial you would expect that the old value of that postprocessor during the first timestep would be the value you got during initial but that s currently not the case i have a fix for this
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.