Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
73,172 | 24,482,443,672 | IssuesEvent | 2022-10-09 02:00:15 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | BUG: Upgrading from Matplotlib 3.5.3 to 3.6.0 Breaks voronoi_plot_2d() | defect scipy.spatial | ### Describe your issue.
Upgrading from Matplotlib 3.5.3 to 3.6.0 Breaks voronoi_plot_2d()
### Reproducing Code Example
```python
rng = np.random.default_rng()
points = rng.random((10,2))
vor = Voronoi(points)
fig = voronoi_plot_2d(vor)
fig = voronoi_plot_2d(vor, show_vertices=False, line_colors='orange',
line_width=2, line_alpha=0.6, point_size=2)
plt.show()
```
### Error message
```shell
Traceback (most recent call last):
File "C:\Users\cpgui\AppData\Local\Programs\Python\Python310\lib\code.py", line 90, in runcode
exec(code, self.locals)
File "<input>", line 1, in <module>
File "C:\Program Files\JetBrains\PyCharm 2022.2\plugins\python\helpers\pydev\_pydev_bundle\pydev_umd.py", line 198, in runfile
pydev_imports.execfile(filename, global_vars, local_vars) # execute the script
File "C:\Program Files\JetBrains\PyCharm 2022.2\plugins\python\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile
exec(compile(contents+"\n", file, 'exec'), glob, loc)
File "C:/Users/cpgui/PycharmProjects/Toyota-Research/main.py", line 34, in <module>
fig = voronoi_plot_2d(vor)
File "<decorator-gen-10>", line 2, in voronoi_plot_2d
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\scipy\spatial\_plotutils.py", line 12, in _held_figure
fig = plt.figure()
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\_api\deprecation.py", line 454, in wrapper
return func(*args, **kwargs)
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 771, in figure
manager = new_figure_manager(
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 346, in new_figure_manager
_warn_if_gui_out_of_main_thread()
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 336, in _warn_if_gui_out_of_main_thread
if (_get_required_interactive_framework(_get_backend_mod()) and
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 206, in _get_backend_mod
switch_backend(dict.__getitem__(rcParams, "backend"))
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 266, in switch_backend
canvas_class = backend_mod.FigureCanvas
AttributeError: module 'backend_interagg' has no attribute 'FigureCanvas'. Did you mean: 'FigureCanvasAgg'?
```
### SciPy/NumPy/Python version information
scipy 1.9.1, numpy 1.23.3, python 3.10 | 1.0 | BUG: Upgrading from Matplotlib 3.5.3 to 3.6.0 Breaks voronoi_plot_2d() - ### Describe your issue.
Upgrading from Matplotlib 3.5.3 to 3.6.0 Breaks voronoi_plot_2d()
### Reproducing Code Example
```python
rng = np.random.default_rng()
points = rng.random((10,2))
vor = Voronoi(points)
fig = voronoi_plot_2d(vor)
fig = voronoi_plot_2d(vor, show_vertices=False, line_colors='orange',
line_width=2, line_alpha=0.6, point_size=2)
plt.show()
```
### Error message
```shell
Traceback (most recent call last):
File "C:\Users\cpgui\AppData\Local\Programs\Python\Python310\lib\code.py", line 90, in runcode
exec(code, self.locals)
File "<input>", line 1, in <module>
File "C:\Program Files\JetBrains\PyCharm 2022.2\plugins\python\helpers\pydev\_pydev_bundle\pydev_umd.py", line 198, in runfile
pydev_imports.execfile(filename, global_vars, local_vars) # execute the script
File "C:\Program Files\JetBrains\PyCharm 2022.2\plugins\python\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile
exec(compile(contents+"\n", file, 'exec'), glob, loc)
File "C:/Users/cpgui/PycharmProjects/Toyota-Research/main.py", line 34, in <module>
fig = voronoi_plot_2d(vor)
File "<decorator-gen-10>", line 2, in voronoi_plot_2d
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\scipy\spatial\_plotutils.py", line 12, in _held_figure
fig = plt.figure()
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\_api\deprecation.py", line 454, in wrapper
return func(*args, **kwargs)
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 771, in figure
manager = new_figure_manager(
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 346, in new_figure_manager
_warn_if_gui_out_of_main_thread()
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 336, in _warn_if_gui_out_of_main_thread
if (_get_required_interactive_framework(_get_backend_mod()) and
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 206, in _get_backend_mod
switch_backend(dict.__getitem__(rcParams, "backend"))
File "C:\Users\cpgui\PycharmProjects\Toyota-Research\venv\lib\site-packages\matplotlib\pyplot.py", line 266, in switch_backend
canvas_class = backend_mod.FigureCanvas
AttributeError: module 'backend_interagg' has no attribute 'FigureCanvas'. Did you mean: 'FigureCanvasAgg'?
```
### SciPy/NumPy/Python version information
scipy 1.9.1, numpy 1.23.3, python 3.10 | defect | bug upgrading from matplotlib to breaks voronoi plot describe your issue upgrading from matplotlib to breaks voronoi plot reproducing code example python rng np random default rng points rng random vor voronoi points fig voronoi plot vor fig voronoi plot vor show vertices false line colors orange line width line alpha point size plt show error message shell traceback most recent call last file c users cpgui appdata local programs python lib code py line in runcode exec code self locals file line in file c program files jetbrains pycharm plugins python helpers pydev pydev bundle pydev umd py line in runfile pydev imports execfile filename global vars local vars execute the script file c program files jetbrains pycharm plugins python helpers pydev pydev imps pydev execfile py line in execfile exec compile contents n file exec glob loc file c users cpgui pycharmprojects toyota research main py line in fig voronoi plot vor file line in voronoi plot file c users cpgui pycharmprojects toyota research venv lib site packages scipy spatial plotutils py line in held figure fig plt figure file c users cpgui pycharmprojects toyota research venv lib site packages matplotlib api deprecation py line in wrapper return func args kwargs file c users cpgui pycharmprojects toyota research venv lib site packages matplotlib pyplot py line in figure manager new figure manager file c users cpgui pycharmprojects toyota research venv lib site packages matplotlib pyplot py line in new figure manager warn if gui out of main thread file c users cpgui pycharmprojects toyota research venv lib site packages matplotlib pyplot py line in warn if gui out of main thread if get required interactive framework get backend mod and file c users cpgui pycharmprojects toyota research venv lib site packages matplotlib pyplot py line in get backend mod switch backend dict getitem rcparams backend file c users cpgui pycharmprojects toyota research venv lib site packages matplotlib pyplot py line in switch backend canvas class backend mod figurecanvas attributeerror module backend interagg has no attribute figurecanvas did you mean figurecanvasagg scipy numpy python version information scipy numpy python | 1 |
125,696 | 4,963,665,581 | IssuesEvent | 2016-12-03 10:54:21 | gre/gl-react | https://api.github.com/repos/gre/gl-react | closed | Configure the linear interpolation of any sampler2D uniform | feature priority:medium | linear interpolation can be defined on almost everything: textures, content and fbos.
so we should be able to define a `disableLinearInterpolation: true` prop on a GL.Uniform or attached in a `{value, opts: { disableLinearInterpolation: true }}` object, like it's already possible for ndarray in DOM impl: https://github.com/ProjectSeptemberInc/gl-react-dom/blob/601180581a461760f541ee459552df9b8d2c52ef/Examples/Simple/Colorify.js#L36 / https://github.com/ProjectSeptemberInc/gl-react-dom/blob/6e4b9f4a8eb0045977c643a4920651082cb36fcb/Examples/Tests/index.js#L102-L111
use case:
you have a HelloGL 2x2 and you want to scale up to 200x200 but without linear interpolation so you preserve the 4 pixels in the display
| 1.0 | Configure the linear interpolation of any sampler2D uniform - linear interpolation can be defined on almost everything: textures, content and fbos.
so we should be able to define a `disableLinearInterpolation: true` prop on a GL.Uniform or attached in a `{value, opts: { disableLinearInterpolation: true }}` object, like it's already possible for ndarray in DOM impl: https://github.com/ProjectSeptemberInc/gl-react-dom/blob/601180581a461760f541ee459552df9b8d2c52ef/Examples/Simple/Colorify.js#L36 / https://github.com/ProjectSeptemberInc/gl-react-dom/blob/6e4b9f4a8eb0045977c643a4920651082cb36fcb/Examples/Tests/index.js#L102-L111
use case:
you have a HelloGL 2x2 and you want to scale up to 200x200 but without linear interpolation so you preserve the 4 pixels in the display
| non_defect | configure the linear interpolation of any uniform linear interpolation can be defined on almost everything textures content and fbos so we should be able to define a disablelinearinterpolation true prop on a gl uniform or attached in a value opts disablelinearinterpolation true object like it s already possible for ndarray in dom impl use case you have a hellogl and you want to scale up to but without linear interpolation so you preserve the pixels in the display | 0 |
41,961 | 10,727,953,552 | IssuesEvent | 2019-10-28 12:57:10 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Dock: problems when page has scrollbar | defect | Hi, I noticed some erroneous behaviour with the dock component on the latest 6.2 RC1
Problem 1:
When reloading the page, as scroll status is not the top (e.g. in the middle of the page) the dock does not animate correctly.

Problem 2:
When using halign="right" the dock moves into the scrollbar

Code to reproduce:
```xml
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:h="http://java.sun.com/jsf/html"
xmlns:p="http://primefaces.org/ui"
xmlns:f="http://xmlns.jcp.org/jsf/core">
<h:head>
</h:head>
<h:body>
<h:form>
<div
style="height: 20000px; background: #e0efe5; border: 1px solid teal;">
Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam
nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam
erat, sed diam voluptua. At vero eos et accusam et justo duo dolores
et ea rebum. Stet clita kasd gubergren, no sea takimata sanctus est
Lorem ipsum dolor sit amet. Lorem ipsum dolor sit amet, consetetur
sadipscing elitr, sed diam nonumy eirmod tempor invidunt ut labore et
dolore magna aliquyam erat, sed diam voluptua. At vero eos et accusam
et justo duo dolores et ea rebum. Stet clita kasd gubergren, no sea
takimata sanctus est Lorem ipsum dolor sit amet.</div>
<p:dock position="bottom">
<p:menuitem value="123"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
<p:menuitem value="456"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
<p:menuitem value="789"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
<p:menuitem value="abc"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
</p:dock>
</h:form>
</h:body>
</html>
```
| 1.0 | Dock: problems when page has scrollbar - Hi, I noticed some erroneous behaviour with the dock component on the latest 6.2 RC1
Problem 1:
When reloading the page, as scroll status is not the top (e.g. in the middle of the page) the dock does not animate correctly.

Problem 2:
When using halign="right" the dock moves into the scrollbar

Code to reproduce:
```xml
<html xmlns="http://www.w3.org/1999/xhtml"
xmlns:h="http://java.sun.com/jsf/html"
xmlns:p="http://primefaces.org/ui"
xmlns:f="http://xmlns.jcp.org/jsf/core">
<h:head>
</h:head>
<h:body>
<h:form>
<div
style="height: 20000px; background: #e0efe5; border: 1px solid teal;">
Lorem ipsum dolor sit amet, consetetur sadipscing elitr, sed diam
nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam
erat, sed diam voluptua. At vero eos et accusam et justo duo dolores
et ea rebum. Stet clita kasd gubergren, no sea takimata sanctus est
Lorem ipsum dolor sit amet. Lorem ipsum dolor sit amet, consetetur
sadipscing elitr, sed diam nonumy eirmod tempor invidunt ut labore et
dolore magna aliquyam erat, sed diam voluptua. At vero eos et accusam
et justo duo dolores et ea rebum. Stet clita kasd gubergren, no sea
takimata sanctus est Lorem ipsum dolor sit amet.</div>
<p:dock position="bottom">
<p:menuitem value="123"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
<p:menuitem value="456"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
<p:menuitem value="789"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
<p:menuitem value="abc"
icon="https://www.primefaces.org/showcase/resources/demo/images/dock/home.png"
onclick="return false;" />
</p:dock>
</h:form>
</h:body>
</html>
```
| defect | dock problems when page has scrollbar hi i noticed some erroneous behaviour with the dock component on the latest problem when reloading the page as scroll status is not the top e g in the middle of the page the dock does not animate correctly problem when using halign right the dock moves into the scrollbar code to reproduce xml html xmlns xmlns h xmlns p xmlns f div style height background border solid teal lorem ipsum dolor sit amet consetetur sadipscing elitr sed diam nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam erat sed diam voluptua at vero eos et accusam et justo duo dolores et ea rebum stet clita kasd gubergren no sea takimata sanctus est lorem ipsum dolor sit amet lorem ipsum dolor sit amet consetetur sadipscing elitr sed diam nonumy eirmod tempor invidunt ut labore et dolore magna aliquyam erat sed diam voluptua at vero eos et accusam et justo duo dolores et ea rebum stet clita kasd gubergren no sea takimata sanctus est lorem ipsum dolor sit amet p menuitem value icon onclick return false p menuitem value icon onclick return false p menuitem value icon onclick return false p menuitem value abc icon onclick return false | 1 |
84,038 | 10,467,336,648 | IssuesEvent | 2019-09-22 04:04:00 | Original-heapsters/gitdrnk | https://api.github.com/repos/Original-heapsters/gitdrnk | closed | Add a user to a game if they join and havent been there before | API Backend DBDesign | user email needs to be appended to the game they join(if they havent before) | 1.0 | Add a user to a game if they join and havent been there before - user email needs to be appended to the game they join(if they havent before) | non_defect | add a user to a game if they join and havent been there before user email needs to be appended to the game they join if they havent before | 0 |
104,081 | 4,194,991,296 | IssuesEvent | 2016-06-25 12:41:47 | fac-freelancers/website | https://api.github.com/repos/fac-freelancers/website | opened | Give the icons in the howWeWork section an animation | priority-2 T30min | They just jump suddenly to a larger size. They should smoothly enlarge and shrink on click | 1.0 | Give the icons in the howWeWork section an animation - They just jump suddenly to a larger size. They should smoothly enlarge and shrink on click | non_defect | give the icons in the howwework section an animation they just jump suddenly to a larger size they should smoothly enlarge and shrink on click | 0 |
53,302 | 3,038,120,174 | IssuesEvent | 2015-08-06 20:43:06 | sweet4lorie/TestRepo | https://api.github.com/repos/sweet4lorie/TestRepo | closed | Importing pymel outside of Maya for the benefit of context-specific auto-completion in a python text editor. | bug imported Milestone-0.7.x Priority-Medium | _From [olegalex...@gmail.com](https://code.google.com/u/117772517933350215565/) on November 19, 2007 12:03:29_
When attempting to import pymel outside of maya on Windows XP:
\>>> import pymel
Traceback (most recent call last):
File "<interactive input>", line 1, in \<module>
File "C:\Program Files\Autodesk\Maya2008\Python\lib\site-
packages\pymel\__init__.py", line 341, in \<module>
del(eval)
NameError: name 'eval' is not defined
_Original issue: http://code.google.com/p/pymel/issues/detail?id=1_ | 1.0 | Importing pymel outside of Maya for the benefit of context-specific auto-completion in a python text editor. - _From [olegalex...@gmail.com](https://code.google.com/u/117772517933350215565/) on November 19, 2007 12:03:29_
When attempting to import pymel outside of maya on Windows XP:
\>>> import pymel
Traceback (most recent call last):
File "<interactive input>", line 1, in \<module>
File "C:\Program Files\Autodesk\Maya2008\Python\lib\site-
packages\pymel\__init__.py", line 341, in \<module>
del(eval)
NameError: name 'eval' is not defined
_Original issue: http://code.google.com/p/pymel/issues/detail?id=1_ | non_defect | importing pymel outside of maya for the benefit of context specific auto completion in a python text editor from on november when attempting to import pymel outside of maya on windows xp import pymel traceback most recent call last file line in file c program files autodesk python lib site packages pymel init py line in del eval nameerror name eval is not defined original issue | 0 |
57,843 | 16,101,205,748 | IssuesEvent | 2021-04-27 09:31:02 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | enclosure path not updated when it changes between reboots | Status: Triage Needed Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04.2 LTS
Linux Kernel |5.4.0-72-generic
Architecture |x86_64
ZFS Version |0.8.3-1ubuntu12.8 and 2.1.0-rc4
SPL Version |0.8.3-1ubuntu12.8 and 2.1.0-rc4
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
If enclosure path changes between reboots, the old enclosure path is still left in the configuration even though it should be updated on import, AIUI. Even after an explicit `zpool export` followed by `zpool import` of the pool the old/wrong enclosure path is still present. It should be noted that the config does update other properties, like path (there seems to be some regression with our vdev_id.conf so the disk numbers are different with 2.1.0-rc4).
```
# zdb -C blob3|tail -n 15
children[11]:
type: 'disk'
id: 11
guid: 15843257147329148638
path: '/dev/disk/by-vdev/enc10d11-part1'
devid: 'scsi-35000cca23b221f98-part1'
phys_path: 'pci-0000:20:00.0-sas-exp0x5001438018375120-phy11-lun-0'
vdev_enc_sysfs_path: '/sys/class/enclosure/2:0:7:0/11'
whole_disk: 1
DTL: 512
create_txg: 4
com.delphix:vdev_zap_leaf: 141
# ls -la /dev/disk/by-vdev/enc10d11
lrwxrwxrwx 1 root root 10 Apr 27 10:59 /dev/disk/by-vdev/enc10d11 -> ../../sdai
# ls -la /sys/block/sdai/device/enclosure_device*
lrwxrwxrwx 1 root root 0 Apr 27 11:01 /sys/block/sdai/device/enclosure_device:11 -> ../../../../port-2:0:12/end_device-2:0:12/target2:0:16/2:0:16:0/enclosure/2:0:16:0/11/
# ls /sys/class/enclosure/
1:0:12:0@ 2:0:16:0@
```
As can be seen above, enclosure 2:0:7:0 isn't present anymore, but 2:0:16:0 now exists.
### Describe how to reproduce the problem
Our enclosures are primarily HP D2700 and home-built ones based on HP DL180G6.
The enclosure device number seems to move depending on how many disks are inserted when plugged in. So for example if you have 3 disks in the enclosure, plug it in, and then fill it with disks, and then reboot the machine, you will find the enclosure on a different ID.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
| 1.0 | enclosure path not updated when it changes between reboots - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Ubuntu
Distribution Version | 20.04.2 LTS
Linux Kernel |5.4.0-72-generic
Architecture |x86_64
ZFS Version |0.8.3-1ubuntu12.8 and 2.1.0-rc4
SPL Version |0.8.3-1ubuntu12.8 and 2.1.0-rc4
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
If enclosure path changes between reboots, the old enclosure path is still left in the configuration even though it should be updated on import, AIUI. Even after an explicit `zpool export` followed by `zpool import` of the pool the old/wrong enclosure path is still present. It should be noted that the config does update other properties, like path (there seems to be some regression with our vdev_id.conf so the disk numbers are different with 2.1.0-rc4).
```
# zdb -C blob3|tail -n 15
children[11]:
type: 'disk'
id: 11
guid: 15843257147329148638
path: '/dev/disk/by-vdev/enc10d11-part1'
devid: 'scsi-35000cca23b221f98-part1'
phys_path: 'pci-0000:20:00.0-sas-exp0x5001438018375120-phy11-lun-0'
vdev_enc_sysfs_path: '/sys/class/enclosure/2:0:7:0/11'
whole_disk: 1
DTL: 512
create_txg: 4
com.delphix:vdev_zap_leaf: 141
# ls -la /dev/disk/by-vdev/enc10d11
lrwxrwxrwx 1 root root 10 Apr 27 10:59 /dev/disk/by-vdev/enc10d11 -> ../../sdai
# ls -la /sys/block/sdai/device/enclosure_device*
lrwxrwxrwx 1 root root 0 Apr 27 11:01 /sys/block/sdai/device/enclosure_device:11 -> ../../../../port-2:0:12/end_device-2:0:12/target2:0:16/2:0:16:0/enclosure/2:0:16:0/11/
# ls /sys/class/enclosure/
1:0:12:0@ 2:0:16:0@
```
As can be seen above, enclosure 2:0:7:0 isn't present anymore, but 2:0:16:0 now exists.
### Describe how to reproduce the problem
Our enclosures are primarily HP D2700 and home-built ones based on HP DL180G6.
The enclosure device number seems to move depending on how many disks are inserted when plugged in. So for example if you have 3 disks in the enclosure, plug it in, and then fill it with disks, and then reboot the machine, you will find the enclosure on a different ID.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
| defect | enclosure path not updated when it changes between reboots thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu distribution version lts linux kernel generic architecture zfs version and spl version and commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing if enclosure path changes between reboots the old enclosure path is still left in the configuration even though it should be updated on import aiui even after an explicit zpool export followed by zpool import of the pool the old wrong enclosure path is still present it should be noted that the config does update other properties like path there seems to be some regression with our vdev id conf so the disk numbers are different with zdb c tail n children type disk id guid path dev disk by vdev devid scsi phys path pci sas lun vdev enc sysfs path sys class enclosure whole disk dtl create txg com delphix vdev zap leaf ls la dev disk by vdev lrwxrwxrwx root root apr dev disk by vdev sdai ls la sys block sdai device enclosure device lrwxrwxrwx root root apr sys block sdai device enclosure device port end device enclosure ls sys class enclosure as can be seen above enclosure isn t present anymore but now exists describe how to reproduce the problem our enclosures are primarily hp and home built ones based on hp the enclosure device number seems to move depending on how many disks are inserted when plugged in so for example if you have disks in the enclosure plug it in and then fill it with disks and then reboot the machine you will find the enclosure on a different id include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with | 1 |
564,481 | 16,726,766,583 | IssuesEvent | 2021-06-10 13:50:06 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Users can not update VM config | Category: Core & System Community Priority: Normal Status: Accepted Type: Bug | **Description**
When a regular user tries to update the VM's config (for example, a context var), it gets the following error:
```
one.vm.updateconf result FAILURE [one.vm.updateconf] User [2] : Template includes a restricted attribute DISK.
```
The specific setting from oned.conf creating the issue is:
```
VM_RESTRICTED_ATTR = "DISK/ORIGINAL_SIZE"
```
The weird part is that this setting exists also in older versions (for example, 5.10.5), but this issue is non existent on that version.
**To Reproduce**
Create a user (not belonging to oneadmin group), allow this user to use a specific vm (use permission, change owner, whatever). Try to update the VM's context.
**Details**
- Affected components: core
- Version: 5.12.3
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| 1.0 | Users can not update VM config - **Description**
When a regular user tries to update the VM's config (for example, a context var), it gets the following error:
```
one.vm.updateconf result FAILURE [one.vm.updateconf] User [2] : Template includes a restricted attribute DISK.
```
The specific setting from oned.conf creating the issue is:
```
VM_RESTRICTED_ATTR = "DISK/ORIGINAL_SIZE"
```
The weird part is that this setting exists also in older versions (for example, 5.10.5), but this issue is non existent on that version.
**To Reproduce**
Create a user (not belonging to oneadmin group), allow this user to use a specific vm (use permission, change owner, whatever). Try to update the VM's context.
**Details**
- Affected components: core
- Version: 5.12.3
**Additional context**
Add any other context about the problem here.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
| non_defect | users can not update vm config description when a regular user tries to update the vm s config for example a context var it gets the following error one vm updateconf result failure user template includes a restricted attribute disk the specific setting from oned conf creating the issue is vm restricted attr disk original size the weird part is that this setting exists also in older versions for example but this issue is non existent on that version to reproduce create a user not belonging to oneadmin group allow this user to use a specific vm use permission change owner whatever try to update the vm s context details affected components core version additional context add any other context about the problem here progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches | 0 |
102,823 | 12,825,728,502 | IssuesEvent | 2020-07-06 15:23:53 | Opentrons/opentrons | https://api.github.com/repos/Opentrons/opentrons | opened | TC PD UX: Add tooltips to profile buttons | :spider: SPDDRS design protocol designer | ## Background
As a user I would like to know the difference between the + STEP and + CYCLE buttons
## Acceptance criteria
- [ ] Add tooltip to + STEP button
- [ ] Add tooltip to + CYCLE button
- [ ] Add tooltip to + STEP button within a cycle
## Copy
tbd | 2.0 | TC PD UX: Add tooltips to profile buttons - ## Background
As a user I would like to know the difference between the + STEP and + CYCLE buttons
## Acceptance criteria
- [ ] Add tooltip to + STEP button
- [ ] Add tooltip to + CYCLE button
- [ ] Add tooltip to + STEP button within a cycle
## Copy
tbd | non_defect | tc pd ux add tooltips to profile buttons background as a user i would like to know the difference between the step and cycle buttons acceptance criteria add tooltip to step button add tooltip to cycle button add tooltip to step button within a cycle copy tbd | 0 |
67,510 | 20,972,894,683 | IssuesEvent | 2022-03-28 13:01:11 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | p-autoComplete: Dropdown stays open when using iOS Safari next/prev keyboard buttons | defect | **I'm submitting a ...**
[x] bug report
**Current behavior**
Go to page with multiple input fields and an autocomplete, press safari next button on ipad keyboard. The focus jumps to next input field but the dropdown is not closed (example in attached image)
**Expected behavior**
After pressing the safari next button, the dropdown should close.
**Minimal reproduction of the problem with instructions**
1. Focus autocomplete field
2. Type a letter so suggestions are shown
3. Press next button on safari keyboard to next input

**Please tell us about your environment:**
Ipad air 2 - iOS 12 - Safari
* **Angular version:** 6.X
* **PrimeNG version:** 6.1.4
* **Browser:** iOS 12 Safari
| 1.0 | p-autoComplete: Dropdown stays open when using iOS Safari next/prev keyboard buttons - **I'm submitting a ...**
[x] bug report
**Current behavior**
Go to page with multiple input fields and an autocomplete, press safari next button on ipad keyboard. The focus jumps to next input field but the dropdown is not closed (example in attached image)
**Expected behavior**
After pressing the safari next button, the dropdown should close.
**Minimal reproduction of the problem with instructions**
1. Focus autocomplete field
2. Type a letter so suggestions are shown
3. Press next button on safari keyboard to next input

**Please tell us about your environment:**
Ipad air 2 - iOS 12 - Safari
* **Angular version:** 6.X
* **PrimeNG version:** 6.1.4
* **Browser:** iOS 12 Safari
| defect | p autocomplete dropdown stays open when using ios safari next prev keyboard buttons i m submitting a bug report current behavior go to page with multiple input fields and an autocomplete press safari next button on ipad keyboard the focus jumps to next input field but the dropdown is not closed example in attached image expected behavior after pressing the safari next button the dropdown should close minimal reproduction of the problem with instructions focus autocomplete field type a letter so suggestions are shown press next button on safari keyboard to next input please tell us about your environment ipad air ios safari angular version x primeng version browser ios safari | 1 |
34,980 | 6,398,168,044 | IssuesEvent | 2017-08-04 19:53:41 | wizeline/wizelink-back | https://api.github.com/repos/wizeline/wizelink-back | closed | Design and document architecture for Story 1 (Save Links) | documentation p0 | The resulting document will be augmented as more stories are included, and will probably grow organically. It's just important to get it started so that there's a place to document architectural decisions as we move along. | 1.0 | Design and document architecture for Story 1 (Save Links) - The resulting document will be augmented as more stories are included, and will probably grow organically. It's just important to get it started so that there's a place to document architectural decisions as we move along. | non_defect | design and document architecture for story save links the resulting document will be augmented as more stories are included and will probably grow organically it s just important to get it started so that there s a place to document architectural decisions as we move along | 0 |
22,823 | 3,703,133,875 | IssuesEvent | 2016-02-29 19:18:36 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Investigate enabling field length guard for non-final fields. | accepted area-vm priority-unassigned Type-Defect | Field guard length checking is only done for final fields. Investigate relaxing this. | 1.0 | Investigate enabling field length guard for non-final fields. - Field guard length checking is only done for final fields. Investigate relaxing this. | defect | investigate enabling field length guard for non final fields field guard length checking is only done for final fields investigate relaxing this | 1 |
67,770 | 28,045,622,753 | IssuesEvent | 2023-03-28 22:31:39 | cncf/cnf-testsuite | https://api.github.com/repos/cncf/cnf-testsuite | closed | [Workload] Check if processes use a sig term | 8 pts workload microservice sprint23.07 v0.41.3 | ## Title: [Workload] Microservices test: sig_term_handled
**Is your workload test idea related to a problem? Please describe.**
- SIGTERM
Problem 1: How the Linux kernel handles signals
The Linux kernel handles signals differently for the process that has PID 1 than it does for other processes. Signal handlers aren't automatically registered for this process, meaning that signals such as SIGTERM or SIGINT will have no effect by default. By default, you must kill processes by using SIGKILL, preventing any graceful shutdown. Depending on your app, using SIGKILL can result in user-facing errors, interrupted writes (for data stores), or unwanted alerts in your monitoring system.
https://cloud.google.com/architecture/best-practices-for-building-containers#problem_1_how_the_linux_kernel_handles_signals
**Describe the solution you'd like**
- Execute a termination and see if SIGTERM signal is passed into the child processes using strace.
**Test Category Name**
- Microservices
**Type of test (static or runtime)**
- runtime
---
### Documentation tasks:
- [ ] Update [installation instructions](https://github.com/cncf/cnf-testsuite/blob/main/install.md) if needed
- [ ] Update [Test Categories md](https://github.com/cncf/cnf-testsuite/blob/main/TEST-CATEGORIES.md) if needed
- [ ] Update [USAGE md](https://github.com/cncf/cnf-testsuite/blob/main/USAGE.md) if needed
- [ ] How to run
- [ ] Description and details
- [ ] What the best practice is
- [ ] Why are we testing this
- [ ] Remediation steps if test does not pass
### QA tasks
Dev Review:
- [ ] walk through A/C
- [ ] do you get the expected result?
- [ ] if yes,
- [ ] move to `Needs Peer Review` column
- [ ] create Pull Request and follow check list
- [ ] Assign 1 or more people for peer review
- [ ] if no, document what additional tasks will be needed
Peer review:
- [ ] walk through A/C
- [ ] do you get the expected result?
- [ ] if yes,
- [ ] move to `Reviewer Approved` column
- [ ] Approve pull request
- [ ] if no,
- [ ] document what did not go as expected, including error messages and screenshots (if possible)
- [ ] Add comment to pull request
- [ ] request changes to pull request
| 1.0 | [Workload] Check if processes use a sig term - ## Title: [Workload] Microservices test: sig_term_handled
**Is your workload test idea related to a problem? Please describe.**
- SIGTERM
Problem 1: How the Linux kernel handles signals
The Linux kernel handles signals differently for the process that has PID 1 than it does for other processes. Signal handlers aren't automatically registered for this process, meaning that signals such as SIGTERM or SIGINT will have no effect by default. By default, you must kill processes by using SIGKILL, preventing any graceful shutdown. Depending on your app, using SIGKILL can result in user-facing errors, interrupted writes (for data stores), or unwanted alerts in your monitoring system.
https://cloud.google.com/architecture/best-practices-for-building-containers#problem_1_how_the_linux_kernel_handles_signals
**Describe the solution you'd like**
- Execute a termination and see if SIGTERM signal is passed into the child processes using strace.
**Test Category Name**
- Microservices
**Type of test (static or runtime)**
- runtime
---
### Documentation tasks:
- [ ] Update [installation instructions](https://github.com/cncf/cnf-testsuite/blob/main/install.md) if needed
- [ ] Update [Test Categories md](https://github.com/cncf/cnf-testsuite/blob/main/TEST-CATEGORIES.md) if needed
- [ ] Update [USAGE md](https://github.com/cncf/cnf-testsuite/blob/main/USAGE.md) if needed
- [ ] How to run
- [ ] Description and details
- [ ] What the best practice is
- [ ] Why are we testing this
- [ ] Remediation steps if test does not pass
### QA tasks
Dev Review:
- [ ] walk through A/C
- [ ] do you get the expected result?
- [ ] if yes,
- [ ] move to `Needs Peer Review` column
- [ ] create Pull Request and follow check list
- [ ] Assign 1 or more people for peer review
- [ ] if no, document what additional tasks will be needed
Peer review:
- [ ] walk through A/C
- [ ] do you get the expected result?
- [ ] if yes,
- [ ] move to `Reviewer Approved` column
- [ ] Approve pull request
- [ ] if no,
- [ ] document what did not go as expected, including error messages and screenshots (if possible)
- [ ] Add comment to pull request
- [ ] request changes to pull request
| non_defect | check if processes use a sig term title microservices test sig term handled is your workload test idea related to a problem please describe sigterm problem how the linux kernel handles signals the linux kernel handles signals differently for the process that has pid than it does for other processes signal handlers aren t automatically registered for this process meaning that signals such as sigterm or sigint will have no effect by default by default you must kill processes by using sigkill preventing any graceful shutdown depending on your app using sigkill can result in user facing errors interrupted writes for data stores or unwanted alerts in your monitoring system describe the solution you d like execute a termination and see if sigterm signal is passed into the child processes using strace test category name microservices type of test static or runtime runtime documentation tasks update if needed update if needed update if needed how to run description and details what the best practice is why are we testing this remediation steps if test does not pass qa tasks dev review walk through a c do you get the expected result if yes move to needs peer review column create pull request and follow check list assign or more people for peer review if no document what additional tasks will be needed peer review walk through a c do you get the expected result if yes move to reviewer approved column approve pull request if no document what did not go as expected including error messages and screenshots if possible add comment to pull request request changes to pull request | 0 |
188,355 | 22,046,321,550 | IssuesEvent | 2022-05-30 02:24:49 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | closed | CVE-2020-13143 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed | security vulnerability | ## CVE-2020-13143 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/gadget/configfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/gadget/configfs.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
gadget_dev_desc_UDC_store in drivers/usb/gadget/configfs.c in the Linux kernel 3.16 through 5.6.13 relies on kstrdup without considering the possibility of an internal '\0' value, which allows attackers to trigger an out-of-bounds read, aka CID-15753588bcd4.
<p>Publish Date: 2020-05-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13143>CVE-2020-13143</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-13143">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-13143</a></p>
<p>Release Date: 2020-05-18</p>
<p>Fix Resolution: v5.7-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-13143 (Medium) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2020-13143 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/gadget/configfs.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/gadget/configfs.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
gadget_dev_desc_UDC_store in drivers/usb/gadget/configfs.c in the Linux kernel 3.16 through 5.6.13 relies on kstrdup without considering the possibility of an internal '\0' value, which allows attackers to trigger an out-of-bounds read, aka CID-15753588bcd4.
<p>Publish Date: 2020-05-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13143>CVE-2020-13143</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-13143">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-13143</a></p>
<p>Release Date: 2020-05-18</p>
<p>Fix Resolution: v5.7-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in linux stable autoclosed cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers usb gadget configfs c drivers usb gadget configfs c vulnerability details gadget dev desc udc store in drivers usb gadget configfs c in the linux kernel through relies on kstrdup without considering the possibility of an internal value which allows attackers to trigger an out of bounds read aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
58,458 | 16,542,772,922 | IssuesEvent | 2021-05-27 19:06:38 | google/guava | https://api.github.com/repos/google/guava | closed | Null map field reference in ImmutableMultimap after deserialization | P3 package=collect status=triaged type=defect | When attempting to use Guava version 19.0, 20.0, and 21.0 `ImmutableListMultimap` in Apache Spark 2.1.0, I'm getting NPEs in unexpected places, leading me to believe that the transient `map` field reference is not being set when deserialized, e.g.
```
java.lang.NullPointerException
at com.google.common.collect.ImmutableMultimap.containsKey
(ImmutableMultimap.java:478)
...
java.lang.NullPointerException
at com.google.common.collect.ImmutableListMultimap.get
(ImmutableListMultimap.java:298)
at com.google.common.collect.ImmutableListMultimap.get
(ImmutableListMultimap.java:44)
``` | 1.0 | Null map field reference in ImmutableMultimap after deserialization - When attempting to use Guava version 19.0, 20.0, and 21.0 `ImmutableListMultimap` in Apache Spark 2.1.0, I'm getting NPEs in unexpected places, leading me to believe that the transient `map` field reference is not being set when deserialized, e.g.
```
java.lang.NullPointerException
at com.google.common.collect.ImmutableMultimap.containsKey
(ImmutableMultimap.java:478)
...
java.lang.NullPointerException
at com.google.common.collect.ImmutableListMultimap.get
(ImmutableListMultimap.java:298)
at com.google.common.collect.ImmutableListMultimap.get
(ImmutableListMultimap.java:44)
``` | defect | null map field reference in immutablemultimap after deserialization when attempting to use guava version and immutablelistmultimap in apache spark i m getting npes in unexpected places leading me to believe that the transient map field reference is not being set when deserialized e g java lang nullpointerexception at com google common collect immutablemultimap containskey immutablemultimap java java lang nullpointerexception at com google common collect immutablelistmultimap get immutablelistmultimap java at com google common collect immutablelistmultimap get immutablelistmultimap java | 1 |
199 | 2,521,958,151 | IssuesEvent | 2015-01-19 18:10:43 | numenta/nupic | https://api.github.com/repos/numenta/nupic | closed | Move from cmake to setuptools for building and preparation for distribution | build deployment P3 super | In order to properly distribute NuPIC as a pip package, we need to transfer from a cmake build to a complete setuptools build. This incorporates several related subtasks that are involved in this cleanup, listed below.
* * *
- [x] [Put all functionally of "CMakeLists.txt" into "setup.py"](https://github.com/numenta/nupic/issues/1573)
- [X] [Refactor bindings to adhere to setuptools standards](https://github.com/numenta/nupic/issues/1616)
- [x] [Use python test commands in README and travis (remove cmake commands)](https://github.com/numenta/nupic/issues/1618)
| 1.0 | Move from cmake to setuptools for building and preparation for distribution - In order to properly distribute NuPIC as a pip package, we need to transfer from a cmake build to a complete setuptools build. This incorporates several related subtasks that are involved in this cleanup, listed below.
* * *
- [x] [Put all functionally of "CMakeLists.txt" into "setup.py"](https://github.com/numenta/nupic/issues/1573)
- [X] [Refactor bindings to adhere to setuptools standards](https://github.com/numenta/nupic/issues/1616)
- [x] [Use python test commands in README and travis (remove cmake commands)](https://github.com/numenta/nupic/issues/1618)
| non_defect | move from cmake to setuptools for building and preparation for distribution in order to properly distribute nupic as a pip package we need to transfer from a cmake build to a complete setuptools build this incorporates several related subtasks that are involved in this cleanup listed below | 0 |
59,329 | 17,023,086,529 | IssuesEvent | 2021-07-03 00:19:23 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Feature requests and ideas from the wiki should be changed into trac tickets | Component: admin Priority: minor Resolution: invalid Type: defect | **[Submitted to the original trac issue database at 12.57pm, Tuesday, 29th November 2005]**
We should move these pages to Trac:
http://www.openstreetmap.org/wiki/index.php/Roadmap_For_Version_1.0
http://www.openstreetmap.org/wiki/index.php/Ideas | 1.0 | Feature requests and ideas from the wiki should be changed into trac tickets - **[Submitted to the original trac issue database at 12.57pm, Tuesday, 29th November 2005]**
We should move these pages to Trac:
http://www.openstreetmap.org/wiki/index.php/Roadmap_For_Version_1.0
http://www.openstreetmap.org/wiki/index.php/Ideas | defect | feature requests and ideas from the wiki should be changed into trac tickets we should move these pages to trac | 1 |
4,545 | 4,427,233,279 | IssuesEvent | 2016-08-16 20:47:23 | grpc/grpc-java | https://api.github.com/repos/grpc/grpc-java | opened | Excess garbage in Metadata | performance | Metadata today is a Hashmap of ArrayLists, and each arraylist has an object array. This creates a lot of garbage since Metadata objects are short lived.
Some ideas on how to improve the situation:
- Recycle the objects
- Store headers as a flat array of an initial size, and swap to using a full map if slow. (and avoid Strings when possible) | True | Excess garbage in Metadata - Metadata today is a Hashmap of ArrayLists, and each arraylist has an object array. This creates a lot of garbage since Metadata objects are short lived.
Some ideas on how to improve the situation:
- Recycle the objects
- Store headers as a flat array of an initial size, and swap to using a full map if slow. (and avoid Strings when possible) | non_defect | excess garbage in metadata metadata today is a hashmap of arraylists and each arraylist has an object array this creates a lot of garbage since metadata objects are short lived some ideas on how to improve the situation recycle the objects store headers as a flat array of an initial size and swap to using a full map if slow and avoid strings when possible | 0 |
70,248 | 23,072,934,344 | IssuesEvent | 2022-07-25 19:58:00 | zed-industries/feedback | https://api.github.com/repos/zed-industries/feedback | closed | Can not insert diacritics | defect | I apologize if there’s already an issue open about this! I couldn’t find it, although #199 seems related.
**Describe the bug**
I can not add diacritics to any character in the editor using my Spanish keyboard layout.
**To reproduce**
Press the acute accent key followed by a vowel, like A. Observe that an alert sound is played, and then ‘a’, lacking the diacritical mark, appears.
**Expected behavior**
á is inserted.
**Environment:**
```
Zed 0.47.1 – /Volumes/Zed/Zed.app
macOS 12.4
architecture arm64
```
| 1.0 | Can not insert diacritics - I apologize if there’s already an issue open about this! I couldn’t find it, although #199 seems related.
**Describe the bug**
I can not add diacritics to any character in the editor using my Spanish keyboard layout.
**To reproduce**
Press the acute accent key followed by a vowel, like A. Observe that an alert sound is played, and then ‘a’, lacking the diacritical mark, appears.
**Expected behavior**
á is inserted.
**Environment:**
```
Zed 0.47.1 – /Volumes/Zed/Zed.app
macOS 12.4
architecture arm64
```
| defect | can not insert diacritics i apologize if there’s already an issue open about this i couldn’t find it although seems related describe the bug i can not add diacritics to any character in the editor using my spanish keyboard layout to reproduce press the acute accent key followed by a vowel like a observe that an alert sound is played and then ‘a’ lacking the diacritical mark appears expected behavior á is inserted environment zed – volumes zed zed app macos architecture | 1 |
16,624 | 4,074,260,797 | IssuesEvent | 2016-05-28 09:48:00 | petabyte-research/redflags | https://api.github.com/repos/petabyte-research/redflags | opened | Documentation: add to github | documentation | Commit http://docs.redflags.eu/ to Github (a catalogue? another branch?) so corrections can be done by anybody | 1.0 | Documentation: add to github - Commit http://docs.redflags.eu/ to Github (a catalogue? another branch?) so corrections can be done by anybody | non_defect | documentation add to github commit to github a catalogue another branch so corrections can be done by anybody | 0 |
16,774 | 2,945,068,549 | IssuesEvent | 2015-07-03 10:07:17 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | PF Mobile Tabview bug (rendering three times after switching tab) | 5.1.21 5.2.8 defect mobile | Hi, i am having exactly the same issue described here. (The shown code can be used for testing)
http://forum.primefaces.org/viewtopic.php?f=3&t=41479
PrimeFaces Version: 5.2.7 + Mojarra 2.2.11
Thanks, hope this can be fixed in a future patch, i am using a very ugly workaround up to now | 1.0 | PF Mobile Tabview bug (rendering three times after switching tab) - Hi, i am having exactly the same issue described here. (The shown code can be used for testing)
http://forum.primefaces.org/viewtopic.php?f=3&t=41479
PrimeFaces Version: 5.2.7 + Mojarra 2.2.11
Thanks, hope this can be fixed in a future patch, i am using a very ugly workaround up to now | defect | pf mobile tabview bug rendering three times after switching tab hi i am having exactly the same issue described here the shown code can be used for testing primefaces version mojarra thanks hope this can be fixed in a future patch i am using a very ugly workaround up to now | 1 |
10,223 | 2,618,942,904 | IssuesEvent | 2015-03-03 00:04:51 | marmarek/test | https://api.github.com/repos/marmarek/test | closed | Remove unused menus from Kickoff | C: desktop-linux P: minor R: duplicate T: defect | **Reported by joanna on 19 May 40420130 03:06 UTC**
* Computer/Places
* Recently Used/Documents | 1.0 | Remove unused menus from Kickoff - **Reported by joanna on 19 May 40420130 03:06 UTC**
* Computer/Places
* Recently Used/Documents | defect | remove unused menus from kickoff reported by joanna on may utc computer places recently used documents | 1 |
690,084 | 23,645,242,830 | IssuesEvent | 2022-08-25 21:18:07 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | closed | Add profile support to `skaffold verify` | kind/feature-request priority/p1 2.0.0 area/verify | Currently `skaffold verify` does not support profiles (no support for the `-p` flag, etc). This should be added to support various important use cases for testing across different envs/services | 1.0 | Add profile support to `skaffold verify` - Currently `skaffold verify` does not support profiles (no support for the `-p` flag, etc). This should be added to support various important use cases for testing across different envs/services | non_defect | add profile support to skaffold verify currently skaffold verify does not support profiles no support for the p flag etc this should be added to support various important use cases for testing across different envs services | 0 |
48,401 | 13,068,508,173 | IssuesEvent | 2020-07-31 03:48:04 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | [payload-parsing] unit test fails (Trac #2337) | Migrated from Trac combo core defect | on cobalt with py2-v3.1.1
```text
Start 358: payload-parsing::test
358/479 Test #357: payload-parsing::test ..........................................***Failed 3.20 sec
Running all tests:
bad_decode.cxx...
size_zero_test..............................................FATAL (payload-parsing): 0 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])
ok
uint64_t_read_past_end......................................FATAL (payload-parsing): -10 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
uint64_t_test...............................................0
FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
uint64_t_test_from_middle...................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
bitbuffertest.cxx...
remainingBits............................................... ok
test_read_in_middle.........................................
100=4 -4
100000=32 -32
00010010001=145 145
00000000001=1 1
000000=0 0
111=7 -1
01=1 1
00=0 0
1=1 -1
11=3 -1
00=0 0
1=1 -1
01=1 1
620
ok
test_read_in_middle_oneread.................................
64
100=4 -4
ok
the_test....................................................
100=4 -4
100000=32 -32
00010010001=145 145
00000000001=1 1
000000=0 0
111=7 -1
01=1 1
00=0 0
1=1 -1
11=3 -1
00=0 0
1=1 -1
01=1 1
ok
twos_complement.............................................0000=0 0
0001=1 1
0010=2 2
0011=3 3
0100=4 4
0101=5 5
0110=6 6
0111=7 7
1000=8 -8
1001=9 -7
1010=10 -6
1011=11 -5
1100=12 -4
1101=13 -3
1110=14 -2
1111=15 -1
ok
deltaTest.cxx...
Delta.......................................................FATAL (payload-parsing): expected type 13 and got 0 (deltaTest.cxx:231 in void DeltaDecode(int, std::string))
cannot read event size from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat
cannot read event from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat
UNCAUGHT:expected type 13 and got 0 (in void DeltaDecode(int, std::string))
Version5Waveforms...........................................FATAL (payload-parsing): expected type 21 and got 0 (deltaTest.cxx:313 in void V5Decode(int, std::string))
cannot read event size from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat
cannot read event from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat
UNCAUGHT:expected type 21 and got 0 (in void V5Decode(int, std::string))
endianness.cxx...
decode16.................................................... ok
decode32.................................................... ok
decode64.................................................... ok
decode_uint16_t............................................. ok
decode_uint16_t_end......................................... ok
decode_uint16_t_enoughdtata.................................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])
ok
decode_uint16_t_enoughdtata_BIG.............................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])
ok
decode_uint16_t_middle...................................... ok
decode_uint32_t............................................. ok
decode_uint32_t_end......................................... ok
decode_uint32_t_enoughdtata.................................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])
ok
decode_uint32_t_enoughdtata_BIG.............................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])
ok
decode_uint32_t_middle...................................... ok
decode_uint64_t............................................. ok
decode_uint64_t_end......................................... ok
decode_uint64_t_enoughdtata.................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
decode_uint64_t_enoughdtata_BIG.............................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
decode_uint64_t_middle...................................... ok
endian_swap_16.............................................. ok
endian_swap_32.............................................. ok
endian_swap_64.............................................. ok
signed_vs_unsigned.......................................... ok
read_type_3_hit.cxx...
read_type_3_payloads........................................ FAIL
/scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx:14: FAIL
File: /scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx
Line: 14
Predicate: ifs.good()
Message: unspecified
read_v2_file.cxx...
craps_out_with_not_type_13..................................FATAL (payload-parsing): Asked to decode a payload of type 13 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 13u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<13u>::type])
ok
endian_swapping.............................................Hi, Peter
ok
v2file......................................................EventID: 1
RunID: 87531
N inice hit DOM's: 15
N icetop hit DOM's: 0
...
--------------------------------------------------
2006
107380177990659673
107380177990837934
1168
3
28
0
ok
read_v5_file.cxx...
fails_when_not_type_21......................................FATAL (payload-parsing): Asked to decode a payload of type 21 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 21u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<21u>::type])
ok
v5file......................................................FATAL (I3PayloadParsingEventDecoder): expected type 13, 19, 20, 21, 22 and got 0 (I3PayloadParsingEventDecoder.cxx:77 in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)
cannot read event size from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat
cannot read event from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat
UNCAUGHT:expected type 13, 19, 20, 21, 22 and got 0 (in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)
===================================================================
Pass: 35
Fail: 4
***** THESE TESTS FAIL *****
deltaTest.cxx/Delta
deltaTest.cxx/Version5Waveforms
read_type_3_hit.cxx/read_type_3_payloads
read_v5_file.cxx/v5file
```
Migrated from https://code.icecube.wisc.edu/ticket/2337
```json
{
"status": "closed",
"changetime": "2019-09-04T13:00:01",
"description": "on cobalt with py2-v3.1.1\n{{{\n Start 358: payload-parsing::test\n358/479 Test #358: payload-parsing::test ..........................................***Failed 3.20 sec\nRunning all tests:\nbad_decode.cxx...\n size_zero_test..............................................FATAL (payload-parsing): 0 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])\n ok\n uint64_t_read_past_end......................................FATAL (payload-parsing): -10 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n uint64_t_test...............................................0\nFATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n uint64_t_test_from_middle...................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\nbitbuffertest.cxx...\n remainingBits............................................... ok\n test_read_in_middle.........................................\n\n\n100=4 -4\n100000=32 -32\n00010010001=145 145\n00000000001=1 1\n000000=0 0\n111=7 -1\n01=1 1\n00=0 0\n1=1 -1\n11=3 -1\n00=0 0\n1=1 -1\n01=1 1\n620\n ok\n test_read_in_middle_oneread.................................\n\n\n64\n100=4 -4\n ok\n the_test....................................................\n\n\n100=4 -4\n100000=32 -32\n00010010001=145 145\n00000000001=1 1\n000000=0 0\n111=7 -1\n01=1 1\n00=0 0\n1=1 -1\n11=3 -1\n00=0 0\n1=1 -1\n01=1 1\n ok\n twos_complement.............................................0000=0 0\n0001=1 1\n0010=2 2\n0011=3 3\n0100=4 4\n0101=5 5\n0110=6 6\n0111=7 7\n1000=8 -8\n1001=9 -7\n1010=10 -6\n1011=11 -5\n1100=12 -4\n1101=13 -3\n1110=14 -2\n1111=15 -1\n ok\ndeltaTest.cxx...\n Delta.......................................................FATAL (payload-parsing): expected type 13 and got 0 (deltaTest.cxx:231 in void DeltaDecode(int, std::string))\ncannot read event size from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat\ncannot read event from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat\n UNCAUGHT:expected type 13 and got 0 (in void DeltaDecode(int, std::string))\n Version5Waveforms...........................................FATAL (payload-parsing): expected type 21 and got 0 (deltaTest.cxx:313 in void V5Decode(int, std::string))\ncannot read event size from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat\ncannot read event from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat\n UNCAUGHT:expected type 21 and got 0 (in void V5Decode(int, std::string))\nendianness.cxx...\n decode16.................................................... ok\n decode32.................................................... ok\n decode64.................................................... ok\n decode_uint16_t............................................. ok\n decode_uint16_t_end......................................... ok\n decode_uint16_t_enoughdtata.................................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])\n ok\n decode_uint16_t_enoughdtata_BIG.............................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])\n ok\n decode_uint16_t_middle...................................... ok\n decode_uint32_t............................................. ok\n decode_uint32_t_end......................................... ok\n decode_uint32_t_enoughdtata.................................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])\n ok\n decode_uint32_t_enoughdtata_BIG.............................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])\n ok\n decode_uint32_t_middle...................................... ok\n decode_uint64_t............................................. ok\n decode_uint64_t_end......................................... ok\n decode_uint64_t_enoughdtata.................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n decode_uint64_t_enoughdtata_BIG.............................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n decode_uint64_t_middle...................................... ok\n endian_swap_16.............................................. ok\n endian_swap_32.............................................. ok\n endian_swap_64.............................................. ok\n signed_vs_unsigned.......................................... ok\nread_type_3_hit.cxx...\n read_type_3_payloads........................................ FAIL\n/scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx:14: FAIL\n\n\tFile: /scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx\n\tLine: 14\n\tPredicate: ifs.good()\n\tMessage: unspecified\n\n\nread_v2_file.cxx...\n craps_out_with_not_type_13..................................FATAL (payload-parsing): Asked to decode a payload of type 13 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 13u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<13u>::type])\n ok\n endian_swapping.............................................Hi, Peter\n ok\n v2file......................................................EventID: 1\nRunID: 87531\nN inice hit DOM's: 15\nN icetop hit DOM's: 0\n\n...\n\n--------------------------------------------------\n2006\n107380177990659673\n107380177990837934\n1168\n3\n28\n0\n ok\nread_v5_file.cxx...\n fails_when_not_type_21......................................FATAL (payload-parsing): Asked to decode a payload of type 21 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 21u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<21u>::type])\n ok\n v5file......................................................FATAL (I3PayloadParsingEventDecoder): expected type 13, 19, 20, 21, 22 and got 0 (I3PayloadParsingEventDecoder.cxx:77 in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)\ncannot read event size from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat\ncannot read event from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat\n UNCAUGHT:expected type 13, 19, 20, 21, 22 and got 0 (in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)\n===================================================================\nPass: 35\nFail: 4\n ***** THESE TESTS FAIL *****\n deltaTest.cxx/Delta\n deltaTest.cxx/Version5Waveforms\n read_type_3_hit.cxx/read_type_3_payloads\n read_v5_file.cxx/v5file\n\n}}}\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1567602001168896",
"component": "combo core",
"summary": "[payload-parsing] unit test fails",
"priority": "normal",
"keywords": "",
"time": "2019-06-27T20:35:54",
"milestone": "Autumnal Equinox 2019",
"owner": "",
"type": "defect"
}
```
| 1.0 | [payload-parsing] unit test fails (Trac #2337) - on cobalt with py2-v3.1.1
```text
Start 358: payload-parsing::test
358/479 Test #357: payload-parsing::test ..........................................***Failed 3.20 sec
Running all tests:
bad_decode.cxx...
size_zero_test..............................................FATAL (payload-parsing): 0 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])
ok
uint64_t_read_past_end......................................FATAL (payload-parsing): -10 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
uint64_t_test...............................................0
FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
uint64_t_test_from_middle...................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
bitbuffertest.cxx...
remainingBits............................................... ok
test_read_in_middle.........................................
100=4 -4
100000=32 -32
00010010001=145 145
00000000001=1 1
000000=0 0
111=7 -1
01=1 1
00=0 0
1=1 -1
11=3 -1
00=0 0
1=1 -1
01=1 1
620
ok
test_read_in_middle_oneread.................................
64
100=4 -4
ok
the_test....................................................
100=4 -4
100000=32 -32
00010010001=145 145
00000000001=1 1
000000=0 0
111=7 -1
01=1 1
00=0 0
1=1 -1
11=3 -1
00=0 0
1=1 -1
01=1 1
ok
twos_complement.............................................0000=0 0
0001=1 1
0010=2 2
0011=3 3
0100=4 4
0101=5 5
0110=6 6
0111=7 7
1000=8 -8
1001=9 -7
1010=10 -6
1011=11 -5
1100=12 -4
1101=13 -3
1110=14 -2
1111=15 -1
ok
deltaTest.cxx...
Delta.......................................................FATAL (payload-parsing): expected type 13 and got 0 (deltaTest.cxx:231 in void DeltaDecode(int, std::string))
cannot read event size from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat
cannot read event from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat
UNCAUGHT:expected type 13 and got 0 (in void DeltaDecode(int, std::string))
Version5Waveforms...........................................FATAL (payload-parsing): expected type 21 and got 0 (deltaTest.cxx:313 in void V5Decode(int, std::string))
cannot read event size from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat
cannot read event from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat
UNCAUGHT:expected type 21 and got 0 (in void V5Decode(int, std::string))
endianness.cxx...
decode16.................................................... ok
decode32.................................................... ok
decode64.................................................... ok
decode_uint16_t............................................. ok
decode_uint16_t_end......................................... ok
decode_uint16_t_enoughdtata.................................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])
ok
decode_uint16_t_enoughdtata_BIG.............................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])
ok
decode_uint16_t_middle...................................... ok
decode_uint32_t............................................. ok
decode_uint32_t_end......................................... ok
decode_uint32_t_enoughdtata.................................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])
ok
decode_uint32_t_enoughdtata_BIG.............................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])
ok
decode_uint32_t_middle...................................... ok
decode_uint64_t............................................. ok
decode_uint64_t_end......................................... ok
decode_uint64_t_enoughdtata.................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
decode_uint64_t_enoughdtata_BIG.............................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])
ok
decode_uint64_t_middle...................................... ok
endian_swap_16.............................................. ok
endian_swap_32.............................................. ok
endian_swap_64.............................................. ok
signed_vs_unsigned.......................................... ok
read_type_3_hit.cxx...
read_type_3_payloads........................................ FAIL
/scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx:14: FAIL
File: /scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx
Line: 14
Predicate: ifs.good()
Message: unspecified
read_v2_file.cxx...
craps_out_with_not_type_13..................................FATAL (payload-parsing): Asked to decode a payload of type 13 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 13u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<13u>::type])
ok
endian_swapping.............................................Hi, Peter
ok
v2file......................................................EventID: 1
RunID: 87531
N inice hit DOM's: 15
N icetop hit DOM's: 0
...
--------------------------------------------------
2006
107380177990659673
107380177990837934
1168
3
28
0
ok
read_v5_file.cxx...
fails_when_not_type_21......................................FATAL (payload-parsing): Asked to decode a payload of type 21 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 21u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<21u>::type])
ok
v5file......................................................FATAL (I3PayloadParsingEventDecoder): expected type 13, 19, 20, 21, 22 and got 0 (I3PayloadParsingEventDecoder.cxx:77 in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)
cannot read event size from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat
cannot read event from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat
UNCAUGHT:expected type 13, 19, 20, 21, 22 and got 0 (in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)
===================================================================
Pass: 35
Fail: 4
***** THESE TESTS FAIL *****
deltaTest.cxx/Delta
deltaTest.cxx/Version5Waveforms
read_type_3_hit.cxx/read_type_3_payloads
read_v5_file.cxx/v5file
```
Migrated from https://code.icecube.wisc.edu/ticket/2337
```json
{
"status": "closed",
"changetime": "2019-09-04T13:00:01",
"description": "on cobalt with py2-v3.1.1\n{{{\n Start 358: payload-parsing::test\n358/479 Test #358: payload-parsing::test ..........................................***Failed 3.20 sec\nRunning all tests:\nbad_decode.cxx...\n size_zero_test..............................................FATAL (payload-parsing): 0 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])\n ok\n uint64_t_read_past_end......................................FATAL (payload-parsing): -10 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n uint64_t_test...............................................0\nFATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n uint64_t_test_from_middle...................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\nbitbuffertest.cxx...\n remainingBits............................................... ok\n test_read_in_middle.........................................\n\n\n100=4 -4\n100000=32 -32\n00010010001=145 145\n00000000001=1 1\n000000=0 0\n111=7 -1\n01=1 1\n00=0 0\n1=1 -1\n11=3 -1\n00=0 0\n1=1 -1\n01=1 1\n620\n ok\n test_read_in_middle_oneread.................................\n\n\n64\n100=4 -4\n ok\n the_test....................................................\n\n\n100=4 -4\n100000=32 -32\n00010010001=145 145\n00000000001=1 1\n000000=0 0\n111=7 -1\n01=1 1\n00=0 0\n1=1 -1\n11=3 -1\n00=0 0\n1=1 -1\n01=1 1\n ok\n twos_complement.............................................0000=0 0\n0001=1 1\n0010=2 2\n0011=3 3\n0100=4 4\n0101=5 5\n0110=6 6\n0111=7 7\n1000=8 -8\n1001=9 -7\n1010=10 -6\n1011=11 -5\n1100=12 -4\n1101=13 -3\n1110=14 -2\n1111=15 -1\n ok\ndeltaTest.cxx...\n Delta.......................................................FATAL (payload-parsing): expected type 13 and got 0 (deltaTest.cxx:231 in void DeltaDecode(int, std::string))\ncannot read event size from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat\ncannot read event from file /scratch/kmeagher/testdata/trunk//DeltaWaveforms/input/physics_912_0_0_914.dat\n UNCAUGHT:expected type 13 and got 0 (in void DeltaDecode(int, std::string))\n Version5Waveforms...........................................FATAL (payload-parsing): expected type 21 and got 0 (deltaTest.cxx:313 in void V5Decode(int, std::string))\ncannot read event size from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat\ncannot read event from file /scratch/kmeagher/testdata/trunk//V5Waveforms/physics_22964_0_0_2053.dat\n UNCAUGHT:expected type 21 and got 0 (in void V5Decode(int, std::string))\nendianness.cxx...\n decode16.................................................... ok\n decode32.................................................... ok\n decode64.................................................... ok\n decode_uint16_t............................................. ok\n decode_uint16_t_end......................................... ok\n decode_uint16_t_enoughdtata.................................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])\n ok\n decode_uint16_t_enoughdtata_BIG.............................FATAL (payload-parsing): 1 bytes in the buffer and that's not enough to get a type of size 2 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = short unsigned int; size_t = long unsigned int])\n ok\n decode_uint16_t_middle...................................... ok\n decode_uint32_t............................................. ok\n decode_uint32_t_end......................................... ok\n decode_uint32_t_enoughdtata.................................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])\n ok\n decode_uint32_t_enoughdtata_BIG.............................FATAL (payload-parsing): 3 bytes in the buffer and that's not enough to get a type of size 4 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = unsigned int; size_t = long unsigned int])\n ok\n decode_uint32_t_middle...................................... ok\n decode_uint64_t............................................. ok\n decode_uint64_t_end......................................... ok\n decode_uint64_t_enoughdtata.................................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n decode_uint64_t_enoughdtata_BIG.............................FATAL (payload-parsing): 7 bytes in the buffer and that's not enough to get a type of size 8 (utility.h:46 in T payload_parsing::decode(payload_parsing::Endian, size_t, const std::vector<char>&) [with T = long unsigned int; size_t = long unsigned int])\n ok\n decode_uint64_t_middle...................................... ok\n endian_swap_16.............................................. ok\n endian_swap_32.............................................. ok\n endian_swap_64.............................................. ok\n signed_vs_unsigned.......................................... ok\nread_type_3_hit.cxx...\n read_type_3_payloads........................................ FAIL\n/scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx:14: FAIL\n\n\tFile: /scratch/kmeagher/combo/src/payload-parsing/private/test/read_type_3_hit.cxx\n\tLine: 14\n\tPredicate: ifs.good()\n\tMessage: unspecified\n\n\nread_v2_file.cxx...\n craps_out_with_not_type_13..................................FATAL (payload-parsing): Asked to decode a payload of type 13 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 13u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<13u>::type])\n ok\n endian_swapping.............................................Hi, Peter\n ok\n v2file......................................................EventID: 1\nRunID: 87531\nN inice hit DOM's: 15\nN icetop hit DOM's: 0\n\n...\n\n--------------------------------------------------\n2006\n107380177990659673\n107380177990837934\n1168\n3\n28\n0\n ok\nread_v5_file.cxx...\n fails_when_not_type_21......................................FATAL (payload-parsing): Asked to decode a payload of type 21 and found payload type 50 instead (decode.h:294 in void payload_parsing::decode_payload(typename payload_parsing::DecodeTarget<payloadType>::type&, const payload_parsing::DecodeConfiguration&, std::vector<char>, unsigned int) [with unsigned int payloadType = 21u; typename payload_parsing::DecodeTarget<payloadType>::type = payload_parsing::DecodeTarget<21u>::type])\n ok\n v5file......................................................FATAL (I3PayloadParsingEventDecoder): expected type 13, 19, 20, 21, 22 and got 0 (I3PayloadParsingEventDecoder.cxx:77 in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)\ncannot read event size from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat\ncannot read event from file /scratch/kmeagher/testdata/trunk//payload_testdata/physics-v5.dat\n UNCAUGHT:expected type 13, 19, 20, 21, 22 and got 0 (in virtual I3Time I3PayloadParsingEventDecoder::FillEvent(I3Frame&, const std::vector<char>&) const)\n===================================================================\nPass: 35\nFail: 4\n ***** THESE TESTS FAIL *****\n deltaTest.cxx/Delta\n deltaTest.cxx/Version5Waveforms\n read_type_3_hit.cxx/read_type_3_payloads\n read_v5_file.cxx/v5file\n\n}}}\n",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"_ts": "1567602001168896",
"component": "combo core",
"summary": "[payload-parsing] unit test fails",
"priority": "normal",
"keywords": "",
"time": "2019-06-27T20:35:54",
"milestone": "Autumnal Equinox 2019",
"owner": "",
"type": "defect"
}
```
| defect | unit test fails trac on cobalt with text start payload parsing test test payload parsing test failed sec running all tests bad decode cxx size zero test fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok t read past end fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok t test fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok t test from middle fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok bitbuffertest cxx remainingbits ok test read in middle ok test read in middle oneread ok the test ok twos complement ok deltatest cxx delta fatal payload parsing expected type and got deltatest cxx in void deltadecode int std string cannot read event size from file scratch kmeagher testdata trunk deltawaveforms input physics dat cannot read event from file scratch kmeagher testdata trunk deltawaveforms input physics dat uncaught expected type and got in void deltadecode int std string fatal payload parsing expected type and got deltatest cxx in void int std string cannot read event size from file scratch kmeagher testdata trunk physics dat cannot read event from file scratch kmeagher testdata trunk physics dat uncaught expected type and got in void int std string endianness cxx ok ok ok decode t ok decode t end ok decode t enoughdtata fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok decode t enoughdtata big fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok decode t middle ok decode t ok decode t end ok decode t enoughdtata fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok decode t enoughdtata big fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok decode t middle ok decode t ok decode t end ok decode t enoughdtata fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok decode t enoughdtata big fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector ok decode t middle ok endian swap ok endian swap ok endian swap ok signed vs unsigned ok read type hit cxx read type payloads fail scratch kmeagher combo src payload parsing private test read type hit cxx fail file scratch kmeagher combo src payload parsing private test read type hit cxx line predicate ifs good message unspecified read file cxx craps out with not type fatal payload parsing asked to decode a payload of type and found payload type instead decode h in void payload parsing decode payload typename payload parsing decodetarget type const payload parsing decodeconfiguration std vector unsigned int ok endian swapping hi peter ok eventid runid n inice hit dom s n icetop hit dom s ok read file cxx fails when not type fatal payload parsing asked to decode a payload of type and found payload type instead decode h in void payload parsing decode payload typename payload parsing decodetarget type const payload parsing decodeconfiguration std vector unsigned int ok fatal expected type and got cxx in virtual fillevent const std vector const cannot read event size from file scratch kmeagher testdata trunk payload testdata physics dat cannot read event from file scratch kmeagher testdata trunk payload testdata physics dat uncaught expected type and got in virtual fillevent const std vector const pass fail these tests fail deltatest cxx delta deltatest cxx read type hit cxx read type payloads read file cxx migrated from json status closed changetime description on cobalt with n n start payload parsing test test payload parsing test failed sec nrunning all tests nbad decode cxx n size zero test fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n t read past end fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n t test nfatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n t test from middle fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok nbitbuffertest cxx n remainingbits ok n test read in middle n n n ok n test read in middle oneread n n n ok n the test n n n ok n twos complement n ok ndeltatest cxx n delta fatal payload parsing expected type and got deltatest cxx in void deltadecode int std string ncannot read event size from file scratch kmeagher testdata trunk deltawaveforms input physics dat ncannot read event from file scratch kmeagher testdata trunk deltawaveforms input physics dat n uncaught expected type and got in void deltadecode int std string n fatal payload parsing expected type and got deltatest cxx in void int std string ncannot read event size from file scratch kmeagher testdata trunk physics dat ncannot read event from file scratch kmeagher testdata trunk physics dat n uncaught expected type and got in void int std string nendianness cxx n ok n ok n ok n decode t ok n decode t end ok n decode t enoughdtata fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n decode t enoughdtata big fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n decode t middle ok n decode t ok n decode t end ok n decode t enoughdtata fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n decode t enoughdtata big fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n decode t middle ok n decode t ok n decode t end ok n decode t enoughdtata fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n decode t enoughdtata big fatal payload parsing bytes in the buffer and that s not enough to get a type of size utility h in t payload parsing decode payload parsing endian size t const std vector n ok n decode t middle ok n endian swap ok n endian swap ok n endian swap ok n signed vs unsigned ok nread type hit cxx n read type payloads fail n scratch kmeagher combo src payload parsing private test read type hit cxx fail n n tfile scratch kmeagher combo src payload parsing private test read type hit cxx n tline n tpredicate ifs good n tmessage unspecified n n nread file cxx n craps out with not type fatal payload parsing asked to decode a payload of type and found payload type instead decode h in void payload parsing decode payload typename payload parsing decodetarget type const payload parsing decodeconfiguration std vector unsigned int n ok n endian swapping hi peter n ok n eventid nrunid nn inice hit dom s nn icetop hit dom s n n n n n ok nread file cxx n fails when not type fatal payload parsing asked to decode a payload of type and found payload type instead decode h in void payload parsing decode payload typename payload parsing decodetarget type const payload parsing decodeconfiguration std vector unsigned int n ok n fatal expected type and got cxx in virtual fillevent const std vector const ncannot read event size from file scratch kmeagher testdata trunk payload testdata physics dat ncannot read event from file scratch kmeagher testdata trunk payload testdata physics dat n uncaught expected type and got in virtual fillevent const std vector const n npass nfail n these tests fail n deltatest cxx delta n deltatest cxx n read type hit cxx read type payloads n read file cxx n n n reporter kjmeagher cc resolution fixed ts component combo core summary unit test fails priority normal keywords time milestone autumnal equinox owner type defect | 1 |
43,545 | 11,745,779,729 | IssuesEvent | 2020-03-12 10:26:15 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | Missing selection module in leaflet.py | Component-Analysis defect | ## Expected behavior ##
<!-- A clear and concise description of what you want to do and what you think should happen. (Code to reproduce the behavior can be added below). -->
write_selection works
## Actual behavior ##
<!-- What happened instead. Add as much detail as you can. Include (copy and paste) stack traces and any output. -->
``` python
NameError: name 'selections' is not defined
```
## Code to reproduce the behavior ##
<!-- Show us how to reproduce the failure. If you can, use trajectory files from the test data. Use the code snipped below as a starting point. -->
``` python
from MDAnalysis.analysis.leaflet import LeafletFinder, optimize_cutoff
from MDAnalysisTests.datafiles import Martini_membrane_gro
LIPID_HEAD_STRING = "name PO4"
universe = MDAnalysis.Universe(Martini_membrane_gro)
lipid_heads = universe.select_atoms(LIPID_HEAD_STRING)
lfls_ag = LeafletFinder(universe, lipid_heads, pbc=False)
lfls_ag.write_selection('test.vmd')
```
## Current version of MDAnalysis ##
- Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) '0.20.2-dev0'
- Which version of Python (`python -V`)? 3.7
- Which operating system? Ubuntu 18
| 1.0 | Missing selection module in leaflet.py - ## Expected behavior ##
<!-- A clear and concise description of what you want to do and what you think should happen. (Code to reproduce the behavior can be added below). -->
write_selection works
## Actual behavior ##
<!-- What happened instead. Add as much detail as you can. Include (copy and paste) stack traces and any output. -->
``` python
NameError: name 'selections' is not defined
```
## Code to reproduce the behavior ##
<!-- Show us how to reproduce the failure. If you can, use trajectory files from the test data. Use the code snipped below as a starting point. -->
``` python
from MDAnalysis.analysis.leaflet import LeafletFinder, optimize_cutoff
from MDAnalysisTests.datafiles import Martini_membrane_gro
LIPID_HEAD_STRING = "name PO4"
universe = MDAnalysis.Universe(Martini_membrane_gro)
lipid_heads = universe.select_atoms(LIPID_HEAD_STRING)
lfls_ag = LeafletFinder(universe, lipid_heads, pbc=False)
lfls_ag.write_selection('test.vmd')
```
## Current version of MDAnalysis ##
- Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) '0.20.2-dev0'
- Which version of Python (`python -V`)? 3.7
- Which operating system? Ubuntu 18
| defect | missing selection module in leaflet py expected behavior write selection works actual behavior python nameerror name selections is not defined code to reproduce the behavior python from mdanalysis analysis leaflet import leafletfinder optimize cutoff from mdanalysistests datafiles import martini membrane gro lipid head string name universe mdanalysis universe martini membrane gro lipid heads universe select atoms lipid head string lfls ag leafletfinder universe lipid heads pbc false lfls ag write selection test vmd current version of mdanalysis which version are you using run python c import mdanalysis as mda print mda version which version of python python v which operating system ubuntu | 1 |
17,260 | 2,993,522,348 | IssuesEvent | 2015-07-22 04:38:56 | RRUZ/delphi-ide-theme-editor | https://api.github.com/repos/RRUZ/delphi-ide-theme-editor | closed | Enabling VCL style XE8 in DIC - does not let IDE start | auto-migrated Priority-Medium Type-Defect | ```
see attached pic
```
Original issue reported on code.google.com by `pario...@gmail.com` on 23 Apr 2015 at 12:39
Attachments:
* [2015-04-23_5-35-36.jpg](https://storage.googleapis.com/google-code-attachments/delphi-ide-theme-editor/issue-92/comment-0/2015-04-23_5-35-36.jpg)
| 1.0 | Enabling VCL style XE8 in DIC - does not let IDE start - ```
see attached pic
```
Original issue reported on code.google.com by `pario...@gmail.com` on 23 Apr 2015 at 12:39
Attachments:
* [2015-04-23_5-35-36.jpg](https://storage.googleapis.com/google-code-attachments/delphi-ide-theme-editor/issue-92/comment-0/2015-04-23_5-35-36.jpg)
| defect | enabling vcl style in dic does not let ide start see attached pic original issue reported on code google com by pario gmail com on apr at attachments | 1 |
16,250 | 2,882,323,248 | IssuesEvent | 2015-06-11 03:33:16 | CompEvol/beast2 | https://api.github.com/repos/CompEvol/beast2 | closed | TreeParser fails to correctly parse node metadata | defect MEDIUM priority | TreeParser cannot handle metadata that include vectors such as those produced by TreeAnnotator. | 1.0 | TreeParser fails to correctly parse node metadata - TreeParser cannot handle metadata that include vectors such as those produced by TreeAnnotator. | defect | treeparser fails to correctly parse node metadata treeparser cannot handle metadata that include vectors such as those produced by treeannotator | 1 |
126,234 | 4,974,665,065 | IssuesEvent | 2016-12-06 07:43:26 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | [Request] socket.io v1.7.0 and v1.7.1 | High Priority in progress Library - Request to Add/Update | **Library name:** socket.io
**Git repository url:** https://github.com/socketio/socket.io
**npm package url(optional):** https://www.npmjs.com/package/socket.io
**License(s):**
**Official homepage:** http://socket.io/
**Wanna say something? Leave message here:**
=====================
Notes from cdnjs maintainer:
Please read the README.md and CONTRIBUTING.md document first.
v1.7.0 and v1.7.1
thx~
You are welcome to add a library via sending pull request,
it'll be faster than just opening a request issue,
and please don't forget to read the guidelines for contributing, thanks!! | 1.0 | [Request] socket.io v1.7.0 and v1.7.1 - **Library name:** socket.io
**Git repository url:** https://github.com/socketio/socket.io
**npm package url(optional):** https://www.npmjs.com/package/socket.io
**License(s):**
**Official homepage:** http://socket.io/
**Wanna say something? Leave message here:**
=====================
Notes from cdnjs maintainer:
Please read the README.md and CONTRIBUTING.md document first.
v1.7.0 and v1.7.1
thx~
You are welcome to add a library via sending pull request,
it'll be faster than just opening a request issue,
and please don't forget to read the guidelines for contributing, thanks!! | non_defect | socket io and library name socket io git repository url npm package url optional license s official homepage wanna say something leave message here notes from cdnjs maintainer please read the readme md and contributing md document first and thx you are welcome to add a library via sending pull request it ll be faster than just opening a request issue and please don t forget to read the guidelines for contributing thanks | 0 |
695,900 | 23,875,800,243 | IssuesEvent | 2022-09-07 18:55:42 | operator-framework/rukpak | https://api.github.com/repos/operator-framework/rukpak | closed | Unpacker pod is constantly being re-created | priority/critical-urgent | Our diff logic that decides whether or not to replace an unpack pod is not working correctly. When rukpak is deployed in a cluster with a pod mutation webhook that affects the unpack pod, we currently see those unrelated pod mutations as diffs and decide we need to re-create the pod, which results in those same mutations, which results in our re-spinning the pod... and the cycle continues forever. | 1.0 | Unpacker pod is constantly being re-created - Our diff logic that decides whether or not to replace an unpack pod is not working correctly. When rukpak is deployed in a cluster with a pod mutation webhook that affects the unpack pod, we currently see those unrelated pod mutations as diffs and decide we need to re-create the pod, which results in those same mutations, which results in our re-spinning the pod... and the cycle continues forever. | non_defect | unpacker pod is constantly being re created our diff logic that decides whether or not to replace an unpack pod is not working correctly when rukpak is deployed in a cluster with a pod mutation webhook that affects the unpack pod we currently see those unrelated pod mutations as diffs and decide we need to re create the pod which results in those same mutations which results in our re spinning the pod and the cycle continues forever | 0 |
13,273 | 2,744,102,758 | IssuesEvent | 2015-04-22 03:41:32 | reingart/prueba | https://api.github.com/repos/reingart/prueba | opened | Error al cargar archivos JSON | auto-migrated Priority-Medium Type-Defect | ```
¿Que pasos reproducirán el problema?
1. Al generar el instalador PyRece
2.
3.
¿Cual es la salida esperada? ¿Que es lo que ve en cambio?
El software deberia poder cargar facturas en formato JSON, sin embargo tira un
error de que no hay una variable global llamada json.
Aparentemente el ejecutable no esta pudiendo cargar el modulo simplejson.
Simplejson esta instalado.
¿Que versión del producto están usando? ¿En que sistema operativo?
1.24e en W7 de 32 bits.
```
Original issue reported on code.google.com by `elrul...@gmail.com` on 22 Aug 2011 at 6:33 | 1.0 | Error al cargar archivos JSON - ```
¿Que pasos reproducirán el problema?
1. Al generar el instalador PyRece
2.
3.
¿Cual es la salida esperada? ¿Que es lo que ve en cambio?
El software deberia poder cargar facturas en formato JSON, sin embargo tira un
error de que no hay una variable global llamada json.
Aparentemente el ejecutable no esta pudiendo cargar el modulo simplejson.
Simplejson esta instalado.
¿Que versión del producto están usando? ¿En que sistema operativo?
1.24e en W7 de 32 bits.
```
Original issue reported on code.google.com by `elrul...@gmail.com` on 22 Aug 2011 at 6:33 | defect | error al cargar archivos json ¿que pasos reproducirán el problema al generar el instalador pyrece ¿cual es la salida esperada ¿que es lo que ve en cambio el software deberia poder cargar facturas en formato json sin embargo tira un error de que no hay una variable global llamada json aparentemente el ejecutable no esta pudiendo cargar el modulo simplejson simplejson esta instalado ¿que versión del producto están usando ¿en que sistema operativo en de bits original issue reported on code google com by elrul gmail com on aug at | 1 |
79,992 | 29,820,350,887 | IssuesEvent | 2023-06-17 01:32:33 | microsoft/TypeScript | https://api.github.com/repos/microsoft/TypeScript | closed | Cannot pass result of `typeof ... == 'function'` of `instanceof Function` as function parameter | Not a Defect | # Bug Report
### 🔎 Search Terms
typeof function, instanceof Function, narrowing, Function
<!--
What search terms did you use when trying to find an existing bug report?
List them here so people in the future can find this one more easily.
-->
### 🕗 Version & Regression Information
- This is the behavior in every version I tried, and I reviewed the FAQ for entries about
### ⏯ Playground Link
[Playground link with relevant code](https://www.typescriptlang.org/play?#code/GYVwdgxgLglg9mABAdwE4EMAOiAUwwBcuAdKeqgOYDOR6YAngNoC6AlIgLwB8i4A1mDjIw7AN4BfALAAoGaEiwEiKKnoB1DNjyFeYAUJFEAbnBgATRKJmJEMYLij1MAUzj38nDogDk86PDBvdjQsbVYZKVlpIA)
### 💻 Code
```ts
function wrap (fn: (...args: any[]) => unknown) {}
function tryWrap (fn: unknown): void {
if (typeof fn == 'function') wrap(fn)
} // The error is over here ↑↑
```
### 🙁 Actual behavior
Error: `Argument of type 'Function' is not assignable to parameter of type '(...args: any[]) => unknown'.
Type 'Function' provides no match for the signature '(...args: any[]): unknown'.(2345)`
### 🙂 Expected behavior
No error
### Notes
The reason for having this problem is because `typeof ... == 'function'` is narrowing to `Function`. And here we have a few problems:
* Assuming that `Function` [interface is basically an implementation detail of the compiler](https://github.com/microsoft/TypeScript/issues/41225#issuecomment-716627722)
* `Function` is [too generic](https://github.com/microsoft/TypeScript/issues/34540#issuecomment-543358887)
The behavior for `instanceof Function` is the same.
So, the solution would be to narrow it to a different type, because implementation details should not be exposed as a result of basic ES operations, or fix the `Function` type | 1.0 | Cannot pass result of `typeof ... == 'function'` of `instanceof Function` as function parameter - # Bug Report
### 🔎 Search Terms
typeof function, instanceof Function, narrowing, Function
<!--
What search terms did you use when trying to find an existing bug report?
List them here so people in the future can find this one more easily.
-->
### 🕗 Version & Regression Information
- This is the behavior in every version I tried, and I reviewed the FAQ for entries about
### ⏯ Playground Link
[Playground link with relevant code](https://www.typescriptlang.org/play?#code/GYVwdgxgLglg9mABAdwE4EMAOiAUwwBcuAdKeqgOYDOR6YAngNoC6AlIgLwB8i4A1mDjIw7AN4BfALAAoGaEiwEiKKnoB1DNjyFeYAUJFEAbnBgATRKJmJEMYLij1MAUzj38nDogDk86PDBvdjQsbVYZKVlpIA)
### 💻 Code
```ts
function wrap (fn: (...args: any[]) => unknown) {}
function tryWrap (fn: unknown): void {
if (typeof fn == 'function') wrap(fn)
} // The error is over here ↑↑
```
### 🙁 Actual behavior
Error: `Argument of type 'Function' is not assignable to parameter of type '(...args: any[]) => unknown'.
Type 'Function' provides no match for the signature '(...args: any[]): unknown'.(2345)`
### 🙂 Expected behavior
No error
### Notes
The reason for having this problem is because `typeof ... == 'function'` is narrowing to `Function`. And here we have a few problems:
* Assuming that `Function` [interface is basically an implementation detail of the compiler](https://github.com/microsoft/TypeScript/issues/41225#issuecomment-716627722)
* `Function` is [too generic](https://github.com/microsoft/TypeScript/issues/34540#issuecomment-543358887)
The behavior for `instanceof Function` is the same.
So, the solution would be to narrow it to a different type, because implementation details should not be exposed as a result of basic ES operations, or fix the `Function` type | defect | cannot pass result of typeof function of instanceof function as function parameter bug report 🔎 search terms typeof function instanceof function narrowing function what search terms did you use when trying to find an existing bug report list them here so people in the future can find this one more easily 🕗 version regression information this is the behavior in every version i tried and i reviewed the faq for entries about ⏯ playground link 💻 code ts function wrap fn args any unknown function trywrap fn unknown void if typeof fn function wrap fn the error is over here ↑↑ 🙁 actual behavior error argument of type function is not assignable to parameter of type args any unknown type function provides no match for the signature args any unknown 🙂 expected behavior no error notes the reason for having this problem is because typeof function is narrowing to function and here we have a few problems assuming that function function is the behavior for instanceof function is the same so the solution would be to narrow it to a different type because implementation details should not be exposed as a result of basic es operations or fix the function type | 1 |
19,781 | 3,783,530,202 | IssuesEvent | 2016-03-19 06:13:44 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | "load.go" is failing in large clusters | area/performance area/test priority/P1 team/control-plane | "load.go" is currently regularly failing in 1000-node clusters
Generally it is failing like that:
creating rc load-test-small-rc-1703
Expected error:
<*errors.errorString | 0xc20e439b00>: {
s: "Only 0 pods started out of 5",
}
Only 0 pods started out of 5
not to have occurred
@gmarek @davidopp | 1.0 | "load.go" is failing in large clusters - "load.go" is currently regularly failing in 1000-node clusters
Generally it is failing like that:
creating rc load-test-small-rc-1703
Expected error:
<*errors.errorString | 0xc20e439b00>: {
s: "Only 0 pods started out of 5",
}
Only 0 pods started out of 5
not to have occurred
@gmarek @davidopp | non_defect | load go is failing in large clusters load go is currently regularly failing in node clusters generally it is failing like that creating rc load test small rc expected error s only pods started out of only pods started out of not to have occurred gmarek davidopp | 0 |
289,116 | 8,854,984,967 | IssuesEvent | 2019-01-09 04:00:27 | visit-dav/issues-test | https://api.github.com/repos/visit-dav/issues-test | closed | VisIt hangs during re-execution prompted by pick. | bug crash likelihood medium priority reviewed severity high wrong results | This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.siloAdd PC Plot of dAdd CoordSwap operator, swap x and y coordsDraw.Create a scalar expression dp,Open Pick window, set variable to dp.Apply Do a Zone PickUsing Navigation, change the view either by zooming or panning.Do another Zone Pick.Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException)viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready.+ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: This source should not load balance the data.Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027+ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2169
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt hangs during re-execution prompted by pick.
Assigned to: Kathleen Biagas
Category: -
Target version: 2.9.1
Author: Kathleen Biagas
Start: 03/03/2015
Due date:
% Done: 100%
Estimated time:
Created: 03/03/2015 04:55 pm
Updated: 03/20/2015 05:44 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.8.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.siloAdd PC Plot of dAdd CoordSwap operator, swap x and y coordsDraw.Create a scalar expression dp,Open Pick window, set variable to dp.Apply Do a Zone PickUsing Navigation, change the view either by zooming or panning.Do another Zone Pick.Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException)viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready.+ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: This source should not load balance the data.Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027+ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
Comments:
Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
| 1.0 | VisIt hangs during re-execution prompted by pick. - This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.siloAdd PC Plot of dAdd CoordSwap operator, swap x and y coordsDraw.Create a scalar expression dp,Open Pick window, set variable to dp.Apply Do a Zone PickUsing Navigation, change the view either by zooming or panning.Do another Zone Pick.Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException)viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready.+ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: This source should not load balance the data.Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027+ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2169
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: VisIt hangs during re-execution prompted by pick.
Assigned to: Kathleen Biagas
Category: -
Target version: 2.9.1
Author: Kathleen Biagas
Start: 03/03/2015
Due date:
% Done: 100%
Estimated time:
Created: 03/03/2015 04:55 pm
Updated: 03/20/2015 05:44 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.8.2
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
This is a bug Bruce Hammel claims to have been experiencing with VisIt for years, so I've set the priority as high, now that I have reproducible steps. This seems only to occur with 2 nodes. Multiple processors on single node does not replicate.Also seems only to occur in conjunction with CoordSwap operator, and with Pick var being set to an expression (and not the pipeline var) On surface: (ensure parallel engine with 2 nodes) Open multi_curv2d.siloAdd PC Plot of dAdd CoordSwap operator, swap x and y coordsDraw.Create a scalar expression dp,Open Pick window, set variable to dp.Apply Do a Zone PickUsing Navigation, change the view either by zooming or panning.Do another Zone Pick.Engine will hang, must cancel the engine_par job in order to interact with VisIt again. Information window shows Pick wanting to re-execute, and a merge exception: VisIt does not have all the information it needs to perform a pick. Please wait while the necessary information is calculated. All current pick selections have been cached and will be performed when calculations are complete. VisIt will notify you when it is fully ready for more picks. Shortly thereafter, the following occured... Pseudocolor: (InvalidMergeException)viewer: Cannot merge datasets because of an incompatible field 1 and 2. Pick mode now fully ready.+ This shows error seeming to come from the viewer, but if you run with -debug 5, then 2 processors' log files will show this error: This source should not load balance the data.Exception: (InvalidMergeException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/avt/Pipeline/Data/avtDataAttributes.C, line 1360: Cannot merge datasets because of an incompatible field 1 and 2.catch(VisItException) /usr/tmp/brugger/aztec/visitbuild/visit2.8.2/src/engine/main/Executors.h:1027+ This is consistent with what I saw with Bruce's real data, running on Muir in pdebug with 10 nodes and 120 processors, but also 2 nodes and 24 processors.
Comments:
Turns out that Pick was sending the secondary variable request to only 1 processor, which caused an 'Invalid Merge' exception during pipeline re-execution.I modified Pick to request SecondaryVars of all processors.SVN update 25986 (2.9RC), 25988 (trunk)M /src/avt/Queries/Pick/avtPickQuery.C
| non_defect | visit hangs during re execution prompted by pick this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi siloadd pc plot of dadd coordswap operator swap x and y coordsdraw create a scalar expression dp open pick window set variable to dp apply do a zone pickusing navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject visit hangs during re execution prompted by pick assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description this is a bug bruce hammel claims to have been experiencing with visit for years so i ve set the priority as high now that i have reproducible steps this seems only to occur with nodes multiple processors on single node does not replicate also seems only to occur in conjunction with coordswap operator and with pick var being set to an expression and not the pipeline var on surface ensure parallel engine with nodes open multi siloadd pc plot of dadd coordswap operator swap x and y coordsdraw create a scalar expression dp open pick window set variable to dp apply do a zone pickusing navigation change the view either by zooming or panning do another zone pick engine will hang must cancel the engine par job in order to interact with visit again information window shows pick wanting to re execute and a merge exception visit does not have all the information it needs to perform a pick please wait while the necessary information is calculated all current pick selections have been cached and will be performed when calculations are complete visit will notify you when it is fully ready for more picks shortly thereafter the following occured pseudocolor invalidmergeexception viewer cannot merge datasets because of an incompatible field and pick mode now fully ready this shows error seeming to come from the viewer but if you run with debug then processors log files will show this error this source should not load balance the data exception invalidmergeexception usr tmp brugger aztec visitbuild src avt pipeline data avtdataattributes c line cannot merge datasets because of an incompatible field and catch visitexception usr tmp brugger aztec visitbuild src engine main executors h this is consistent with what i saw with bruce s real data running on muir in pdebug with nodes and processors but also nodes and processors comments turns out that pick was sending the secondary variable request to only processor which caused an invalid merge exception during pipeline re execution i modified pick to request secondaryvars of all processors svn update trunk m src avt queries pick avtpickquery c | 0 |
757,592 | 26,519,595,922 | IssuesEvent | 2023-01-19 00:40:41 | SETI/pds-oops | https://api.github.com/repos/SETI/pds-oops | opened | Need method for creating new bodies derived from old bodies | A-Enhancement B-OOPS Priority 5 Minor Effort 2 Medium | For example, I need to create a "Titan+Atmosphere" body that has all of the same attributes as Titan (especially its orbit), but has a slightly larger radius. Currently we can use `copy()` on the body object and register a new body, but this may mess up some internal concepts (how many moons does Saturn have now?).
| 1.0 | Need method for creating new bodies derived from old bodies - For example, I need to create a "Titan+Atmosphere" body that has all of the same attributes as Titan (especially its orbit), but has a slightly larger radius. Currently we can use `copy()` on the body object and register a new body, but this may mess up some internal concepts (how many moons does Saturn have now?).
| non_defect | need method for creating new bodies derived from old bodies for example i need to create a titan atmosphere body that has all of the same attributes as titan especially its orbit but has a slightly larger radius currently we can use copy on the body object and register a new body but this may mess up some internal concepts how many moons does saturn have now | 0 |
8,752 | 2,611,542,847 | IssuesEvent | 2015-02-27 06:11:33 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | opened | Script parameters are not available in onPreviewInit, but should | auto-migrated Priority-Medium Type-Defect | ```
Well, it appears to me that it is not possible to access the script parameters
within onPreviewInit, which is a bummer, since that means while you could
configure Lua-generated maps at will, you will not be able to show a correct
preview for parameters other than the default. So there is a very good reason
to make the parameters available to onPreviewInit.
You can test this for yourself, but this slightly tricky:
First, create a multiplayer script called “ScriptParamTest.lua” containing
this test code:
-- snip --
function onPreviewInit()
if ScriptParam == nil then
print("__LUATEST__: ScriptParam == nil")
else
print("__LUATEST__: ScriptParam ~= nil")
end
end
-- snip --
Also create ScriptParamTest.cfg with two lines containing “Default” each so
you can edit the schemes.
Then create a simple test game scheme where you write anything into the script
parameter field. I don’t know if this step is important.
Lastly, start Hedgewars but make sure you write the Standard error stream
(stderr) into a file. I had to use print because WriteLnToConsole does not work
inside onPreviewInit. Then simply start a game using this script and your test
game scheme and leave Hedgewars as soon as the game has been loaded. Now look
into the file you used for logging and search for the string “__LUATEST__”.
This line will tell you wheather ScriptParam equalled nil.
For me, ScriptParam equalled nil. To make sure I made no stupid flaws in my
tests, I also quickly checked the contents of this variable in onParameters,
and of course in this function this variable contained the actual string.
I tested this on ra88647ead05c under GNU/Linux.
```
Original issue reported on code.google.com by `almikes@aol.com` on 19 Dec 2014 at 8:20 | 1.0 | Script parameters are not available in onPreviewInit, but should - ```
Well, it appears to me that it is not possible to access the script parameters
within onPreviewInit, which is a bummer, since that means while you could
configure Lua-generated maps at will, you will not be able to show a correct
preview for parameters other than the default. So there is a very good reason
to make the parameters available to onPreviewInit.
You can test this for yourself, but this slightly tricky:
First, create a multiplayer script called “ScriptParamTest.lua” containing
this test code:
-- snip --
function onPreviewInit()
if ScriptParam == nil then
print("__LUATEST__: ScriptParam == nil")
else
print("__LUATEST__: ScriptParam ~= nil")
end
end
-- snip --
Also create ScriptParamTest.cfg with two lines containing “Default” each so
you can edit the schemes.
Then create a simple test game scheme where you write anything into the script
parameter field. I don’t know if this step is important.
Lastly, start Hedgewars but make sure you write the Standard error stream
(stderr) into a file. I had to use print because WriteLnToConsole does not work
inside onPreviewInit. Then simply start a game using this script and your test
game scheme and leave Hedgewars as soon as the game has been loaded. Now look
into the file you used for logging and search for the string “__LUATEST__”.
This line will tell you wheather ScriptParam equalled nil.
For me, ScriptParam equalled nil. To make sure I made no stupid flaws in my
tests, I also quickly checked the contents of this variable in onParameters,
and of course in this function this variable contained the actual string.
I tested this on ra88647ead05c under GNU/Linux.
```
Original issue reported on code.google.com by `almikes@aol.com` on 19 Dec 2014 at 8:20 | defect | script parameters are not available in onpreviewinit but should well it appears to me that it is not possible to access the script parameters within onpreviewinit which is a bummer since that means while you could configure lua generated maps at will you will not be able to show a correct preview for parameters other than the default so there is a very good reason to make the parameters available to onpreviewinit you can test this for yourself but this slightly tricky first create a multiplayer script called “scriptparamtest lua” containing this test code snip function onpreviewinit if scriptparam nil then print luatest scriptparam nil else print luatest scriptparam nil end end snip also create scriptparamtest cfg with two lines containing “default” each so you can edit the schemes then create a simple test game scheme where you write anything into the script parameter field i don’t know if this step is important lastly start hedgewars but make sure you write the standard error stream stderr into a file i had to use print because writelntoconsole does not work inside onpreviewinit then simply start a game using this script and your test game scheme and leave hedgewars as soon as the game has been loaded now look into the file you used for logging and search for the string “ luatest ” this line will tell you wheather scriptparam equalled nil for me scriptparam equalled nil to make sure i made no stupid flaws in my tests i also quickly checked the contents of this variable in onparameters and of course in this function this variable contained the actual string i tested this on under gnu linux original issue reported on code google com by almikes aol com on dec at | 1 |
69,210 | 22,282,349,423 | IssuesEvent | 2022-06-11 04:07:57 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Message bubble highlight and the new message composer not aligned | T-Defect | ### Steps to reproduce
1. Enable bubble message layout
1. Open a room
2. Send a message
### Outcome
#### What did you expect?
The width of the message bubble highlight and the new message composer should be same.
#### What happened instead?
It is different and edges are not aligned.


It is reproduced on the modern layout as well.

### Operating system
Debian
### Browser information
Firefox ESR 99
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Message bubble highlight and the new message composer not aligned - ### Steps to reproduce
1. Enable bubble message layout
1. Open a room
2. Send a message
### Outcome
#### What did you expect?
The width of the message bubble highlight and the new message composer should be same.
#### What happened instead?
It is different and edges are not aligned.


It is reproduced on the modern layout as well.

### Operating system
Debian
### Browser information
Firefox ESR 99
### URL for webapp
localhost
### Application version
develop branch
### Homeserver
_No response_
### Will you send logs?
No | defect | message bubble highlight and the new message composer not aligned steps to reproduce enable bubble message layout open a room send a message outcome what did you expect the width of the message bubble highlight and the new message composer should be same what happened instead it is different and edges are not aligned it is reproduced on the modern layout as well operating system debian browser information firefox esr url for webapp localhost application version develop branch homeserver no response will you send logs no | 1 |
11,085 | 2,632,746,943 | IssuesEvent | 2015-03-08 13:29:03 | simonsteele/pn | https://api.github.com/repos/simonsteele/pn | closed | Autocomplete is interfering with trying to type | Priority-Medium Type-Defect | Original [issue 20](https://code.google.com/p/pnotepad/issues/detail?id=20) created by simonsteele on 2008-03-22T18:06:01.000Z:
<b>What steps will reproduce the problem?</b>
1. Run Windows Vista. (maybe. my desktop doesn't have this problem, it's
running XP.)
2. Search. (Ctrl-F)
3. Try to type anything into the search field.
<b>What is the expected output? What do you see instead?</b>
I expect to be able to type out a word. Instead, it attempts to
autocomplete JUST using the letter I type as the first letter. For example,
I'd like to search for "register". I type R, and it doesn't type anything,
it just opens the autocomplete menu and shows me previous queries I've run
that start with r. If I select "r" from that submenu, move the pointer
after it, and press "e", it lets me complete it normally (sometimes). Other
times, it won't let me type AT ALL unless I choose something from
autocomplete, type in the middle of the word, then delete everything on
each side of what I want. Or, I can type out the word I want in the
document itself, highlight it, and then it will show up when I hit ctrl-F.
In general, it is absurdly difficult for me to actually search.
<b>What version of the product are you using? On what operating system?</b>
v2.0.6.1-ella, though I update frequently hoping this is fixed, but it's
not. I've used several versions. I run Windows Vista on this laptop.
<b>Please provide any additional information below.</b>
I uploaded a video demonstrating my problem to YouTube:
http://youtube.com/watch?v=h0qCT6PcBao
| 1.0 | Autocomplete is interfering with trying to type - Original [issue 20](https://code.google.com/p/pnotepad/issues/detail?id=20) created by simonsteele on 2008-03-22T18:06:01.000Z:
<b>What steps will reproduce the problem?</b>
1. Run Windows Vista. (maybe. my desktop doesn't have this problem, it's
running XP.)
2. Search. (Ctrl-F)
3. Try to type anything into the search field.
<b>What is the expected output? What do you see instead?</b>
I expect to be able to type out a word. Instead, it attempts to
autocomplete JUST using the letter I type as the first letter. For example,
I'd like to search for "register". I type R, and it doesn't type anything,
it just opens the autocomplete menu and shows me previous queries I've run
that start with r. If I select "r" from that submenu, move the pointer
after it, and press "e", it lets me complete it normally (sometimes). Other
times, it won't let me type AT ALL unless I choose something from
autocomplete, type in the middle of the word, then delete everything on
each side of what I want. Or, I can type out the word I want in the
document itself, highlight it, and then it will show up when I hit ctrl-F.
In general, it is absurdly difficult for me to actually search.
<b>What version of the product are you using? On what operating system?</b>
v2.0.6.1-ella, though I update frequently hoping this is fixed, but it's
not. I've used several versions. I run Windows Vista on this laptop.
<b>Please provide any additional information below.</b>
I uploaded a video demonstrating my problem to YouTube:
http://youtube.com/watch?v=h0qCT6PcBao
| defect | autocomplete is interfering with trying to type original created by simonsteele on what steps will reproduce the problem run windows vista maybe my desktop doesn t have this problem it s running xp search ctrl f try to type anything into the search field what is the expected output what do you see instead i expect to be able to type out a word instead it attempts to autocomplete just using the letter i type as the first letter for example i d like to search for quot register quot i type r and it doesn t type anything it just opens the autocomplete menu and shows me previous queries i ve run that start with r if i select quot r quot from that submenu move the pointer after it and press quot e quot it lets me complete it normally sometimes other times it won t let me type at all unless i choose something from autocomplete type in the middle of the word then delete everything on each side of what i want or i can type out the word i want in the document itself highlight it and then it will show up when i hit ctrl f in general it is absurdly difficult for me to actually search what version of the product are you using on what operating system ella though i update frequently hoping this is fixed but it s not i ve used several versions i run windows vista on this laptop please provide any additional information below i uploaded a video demonstrating my problem to youtube | 1 |
441,068 | 12,707,380,597 | IssuesEvent | 2020-06-23 08:49:28 | justalemon/SimpleSync | https://api.github.com/repos/justalemon/SimpleSync | closed | sometimes all textures are blinking | priority: p3 low status: needs info type: bug | sometimes all textures are blinking, if you stay at one place, everything is fine. debug shows only time, no errors at all | 1.0 | sometimes all textures are blinking - sometimes all textures are blinking, if you stay at one place, everything is fine. debug shows only time, no errors at all | non_defect | sometimes all textures are blinking sometimes all textures are blinking if you stay at one place everything is fine debug shows only time no errors at all | 0 |
159,513 | 20,068,292,730 | IssuesEvent | 2022-02-04 01:08:16 | scottstientjes/icu | https://api.github.com/repos/scottstientjes/icu | opened | CVE-2021-36373 (Medium) detected in ant-1.10.6.jar | security vulnerability | ## CVE-2021-36373 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.10.6.jar</b></p></summary>
<p>master POM</p>
<p>Library home page: <a href="http://ant.apache.org/">http://ant.apache.org/</a></p>
<p>Path to dependency file: /tools/cldr/cldr-to-icu/pom.xml</p>
<p>Path to vulnerable library: /canner/.m2/repository/org/apache/ant/ant/1.10.6/ant-1.10.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.10.6.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When reading a specially crafted TAR archive an Apache Ant build can be made to allocate large amounts of memory that finally leads to an out of memory error, even for small inputs. This can be used to disrupt builds using Apache Ant. Apache Ant prior to 1.9.16 and 1.10.11 were affected.
<p>Publish Date: 2021-07-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-36373>CVE-2021-36373</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-36373">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-36373</a></p>
<p>Release Date: 2021-07-14</p>
<p>Fix Resolution: org.apache.ant:ant:1.9.16,1.10.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-36373 (Medium) detected in ant-1.10.6.jar - ## CVE-2021-36373 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.10.6.jar</b></p></summary>
<p>master POM</p>
<p>Library home page: <a href="http://ant.apache.org/">http://ant.apache.org/</a></p>
<p>Path to dependency file: /tools/cldr/cldr-to-icu/pom.xml</p>
<p>Path to vulnerable library: /canner/.m2/repository/org/apache/ant/ant/1.10.6/ant-1.10.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.10.6.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When reading a specially crafted TAR archive an Apache Ant build can be made to allocate large amounts of memory that finally leads to an out of memory error, even for small inputs. This can be used to disrupt builds using Apache Ant. Apache Ant prior to 1.9.16 and 1.10.11 were affected.
<p>Publish Date: 2021-07-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-36373>CVE-2021-36373</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-36373">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-36373</a></p>
<p>Release Date: 2021-07-14</p>
<p>Fix Resolution: org.apache.ant:ant:1.9.16,1.10.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in ant jar cve medium severity vulnerability vulnerable library ant jar master pom library home page a href path to dependency file tools cldr cldr to icu pom xml path to vulnerable library canner repository org apache ant ant ant jar dependency hierarchy x ant jar vulnerable library vulnerability details when reading a specially crafted tar archive an apache ant build can be made to allocate large amounts of memory that finally leads to an out of memory error even for small inputs this can be used to disrupt builds using apache ant apache ant prior to and were affected publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache ant ant step up your open source security game with whitesource | 0 |
17,512 | 3,011,002,652 | IssuesEvent | 2015-07-28 15:47:20 | patric-r/jvmtop | https://api.github.com/repos/patric-r/jvmtop | closed | ERROR: Connection refused/access denied | auto-migrated Priority-Medium Type-Defect | ```
# cat /etc/SuSE-release
SUSE Linux Enterprise Server 11 (x86_64)
VERSION = 11
PATCHLEVEL = 0
# java -version
java version "1.6.0_45"
Java(TM) SE Runtime Environment (build 1.6.0_45-b06)
Java HotSpot(TM) 64-Bit Server VM (build 20.45-b01, mixed mode)
# echo $JAVA_HOME
/home/app_env_1/app/Java/jdk1.6.0_45/
# ./jvmtop.sh --once --verbose
Sep 30, 2014 12:28:59 PM com.jvmtop.JvmTop main
FINE: Verbosity mode.
Sep 30, 2014 12:29:00 PM com.jvmtop.monitor.VMInfo attachToVM
FINE: connection refused (PID=25393)
java.rmi.ConnectException: Connection refused to host: 10.249.140.239; nested
exception is:
java.net.ConnectException: Connection refused
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:601)
at sun.rmi.transport.tcp.TCPChannel.createConnection(TCPChannel.java:198)
at sun.rmi.transport.tcp.TCPChannel.newConnection(TCPChannel.java:184)
at sun.rmi.server.UnicastRef.invoke(UnicastRef.java:110)
at javax.management.remote.rmi.RMIServerImpl_Stub.newClient(Unknown Source)
at javax.management.remote.rmi.RMIConnector.getConnection(RMIConnector.java:2327)
at javax.management.remote.rmi.RMIConnector.connect(RMIConnector.java:277)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:248)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:207)
at com.jvmtop.openjdk.tools.ProxyClient.tryConnect(Unknown Source)
at com.jvmtop.openjdk.tools.ProxyClient.connect(Unknown Source)
at com.jvmtop.monitor.VMInfo.attachToVM(Unknown Source)
at com.jvmtop.monitor.VMInfo.processNewVM(Unknown Source)
at com.jvmtop.view.VMOverviewView.scanForNewVMs(Unknown Source)
at com.jvmtop.view.VMOverviewView.printView(Unknown Source)
at com.jvmtop.JvmTop.run(Unknown Source)
at com.jvmtop.JvmTop.main(Unknown Source)
Caused by: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method)
at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:351)
at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:213)
at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:200)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
at java.net.Socket.connect(Socket.java:529)
at java.net.Socket.connect(Socket.java:478)
at java.net.Socket.<init>(Socket.java:375)
at java.net.Socket.<init>(Socket.java:189)
at sun.rmi.transport.proxy.RMIDirectSocketFactory.createSocket(RMIDirectSocketFactory.java:22)
at sun.rmi.transport.proxy.RMIMasterSocketFactory.createSocket(RMIMasterSocketFactory.java:128)
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:595)
... 16 more
JvmTop 0.8.0 alpha - 12:28:59, amd64, 2 cpus, Linux 2.6.27.19, load avg 0.03
http://code.google.com/p/jvmtop
PID MAIN-CLASS HPCUR HPMAX NHCUR NHMAX CPU GC VM USERNAME #T DL
14392 m.jvmtop.JvmTop 12m 1747m 10m 130m 57.69% 0.00% S6U45 root 13
25393 artup.Bootstrap [ERROR: Connection refused/access denied]
# ps aux | grep tomcat | grep -v grep
1001 25393 1.2 16.1 2801376 1299808 pts/9 Sl Sep15 276:12
/home/app_env_1/APP/Java/jdk1.6.0_45/jre/bin/java
-Djava.util.logging.config.file=/home/app_env_1/APP/app-manager/conf/server/logg
ing.properties
-Djava.util.logging.manager=org.apache.juli.ClassLoaderLogManager
-Dapp.conf=/home/app_env_1/APP/app-manager/conf -server -Xms1G -Xmx1G
-XX:+UseCodeCacheFlushing -XX:ReservedCodeCacheSize=64M
-XX:+HeapDumpOnOutOfMemoryError -XX:MaxPermSize=1024M
-Djavax.net.debug=ssl,handshake -Dcom.sun.management.jmxremote
-Dcom.sun.management.jmxremote.port=8004
-Dcom.sun.management.jmxremote.authenticate=false
-Dcom.sun.management.jmxremote.ssl=false
-Djava.rmi.server.hostname=10.249.140.239 -Denv=KIEV_DEV
-Doutput.dir=/home/app_env_1/APP/logs
-Dapp.app=/home/app_env_1/APP/app-application
-Dapp.conf=/home/app_env_1/app/limits-manager/conf -Dapp.version=4.7.0.qa.99
-Djava.endorsed.dirs=/home/app_env_1/app/app-server/common/endorsed -classpath
/home/app_env_1/APP/app-server/bin/bootstrap.jar:/home/app_env_1/APP/app-server/
bin/tomcat-juli.jar -Dcatalina.base=/home/app_env_1/APP/app-server
-Dcatalina.home=/home/app_env_1/APP/app-server
-Djava.io.tmpdir=/home/app_env_1/APP/app-server/temp
org.apache.catalina.startup.Bootstrap -config
/home/app_env_1/APP/app-manager/conf/server/server.xml start
Tomcat version 5.5.23 (yes, I know it is too old, but we must use it).
jvmtop 0.8.0
Same error if run jvmtop under Tomcat's user:
$ id
uid=1001(app_env_1)
$ ./jvmtop.sh --once --verbose
Sep 30, 2014 12:36:38 PM com.jvmtop.JvmTop main
FINE: Verbosity mode.
Sep 30, 2014 12:36:39 PM com.jvmtop.monitor.VMInfo attachToVM
FINE: connection refused (PID=25393)
java.rmi.ConnectException: Connection refused to host: 10.249.140.239; nested
exception is:
java.net.ConnectException: Connection refused
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:601)
at sun.rmi.transport.tcp.TCPChannel.createConnection(TCPChannel.java:198)
at sun.rmi.transport.tcp.TCPChannel.newConnection(TCPChannel.java:184)
at sun.rmi.server.UnicastRef.invoke(UnicastRef.java:110)
at javax.management.remote.rmi.RMIServerImpl_Stub.newClient(Unknown Source)
at javax.management.remote.rmi.RMIConnector.getConnection(RMIConnector.java:2327)
at javax.management.remote.rmi.RMIConnector.connect(RMIConnector.java:277)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:248)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:207)
at com.jvmtop.openjdk.tools.ProxyClient.tryConnect(Unknown Source)
at com.jvmtop.openjdk.tools.ProxyClient.connect(Unknown Source)
at com.jvmtop.monitor.VMInfo.attachToVM(Unknown Source)
at com.jvmtop.monitor.VMInfo.processNewVM(Unknown Source)
at com.jvmtop.view.VMOverviewView.scanForNewVMs(Unknown Source)
at com.jvmtop.view.VMOverviewView.printView(Unknown Source)
at com.jvmtop.JvmTop.run(Unknown Source)
at com.jvmtop.JvmTop.main(Unknown Source)
Caused by: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method)
at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:351)
at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:213)
at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:200)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
at java.net.Socket.connect(Socket.java:529)
at java.net.Socket.connect(Socket.java:478)
at java.net.Socket.<init>(Socket.java:375)
at java.net.Socket.<init>(Socket.java:189)
at sun.rmi.transport.proxy.RMIDirectSocketFactory.createSocket(RMIDirectSocketFactory.java:22)
at sun.rmi.transport.proxy.RMIMasterSocketFactory.createSocket(RMIMasterSocketFactory.java:128)
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:595)
... 16 more
JvmTop 0.8.0 alpha - 12:36:38, amd64, 2 cpus, Linux 2.6.27.19, load avg 0.03
http://code.google.com/p/jvmtop
PID MAIN-CLASS HPCUR HPMAX NHCUR NHMAX CPU GC VM USERNAME #T DL
14520 m.jvmtop.JvmTop 7m 1747m 10m 130m 80.65% 0.00% S6U45 lmt_env_ 13
25393 artup.Bootstrap [ERROR: Connection refused/access denied]
User's Java:
$ echo $JAVA_HOME
/home/app_env_1/LMS/Java/jdk1.6.0_45/
root's Java:
# echo $JAVA_HOME
/home/app_env_1/LMS/Java/jdk1.6.0_45/
```
Original issue reported on code.google.com by `setev...@gmail.com` on 30 Sep 2014 at 9:40 | 1.0 | ERROR: Connection refused/access denied - ```
# cat /etc/SuSE-release
SUSE Linux Enterprise Server 11 (x86_64)
VERSION = 11
PATCHLEVEL = 0
# java -version
java version "1.6.0_45"
Java(TM) SE Runtime Environment (build 1.6.0_45-b06)
Java HotSpot(TM) 64-Bit Server VM (build 20.45-b01, mixed mode)
# echo $JAVA_HOME
/home/app_env_1/app/Java/jdk1.6.0_45/
# ./jvmtop.sh --once --verbose
Sep 30, 2014 12:28:59 PM com.jvmtop.JvmTop main
FINE: Verbosity mode.
Sep 30, 2014 12:29:00 PM com.jvmtop.monitor.VMInfo attachToVM
FINE: connection refused (PID=25393)
java.rmi.ConnectException: Connection refused to host: 10.249.140.239; nested
exception is:
java.net.ConnectException: Connection refused
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:601)
at sun.rmi.transport.tcp.TCPChannel.createConnection(TCPChannel.java:198)
at sun.rmi.transport.tcp.TCPChannel.newConnection(TCPChannel.java:184)
at sun.rmi.server.UnicastRef.invoke(UnicastRef.java:110)
at javax.management.remote.rmi.RMIServerImpl_Stub.newClient(Unknown Source)
at javax.management.remote.rmi.RMIConnector.getConnection(RMIConnector.java:2327)
at javax.management.remote.rmi.RMIConnector.connect(RMIConnector.java:277)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:248)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:207)
at com.jvmtop.openjdk.tools.ProxyClient.tryConnect(Unknown Source)
at com.jvmtop.openjdk.tools.ProxyClient.connect(Unknown Source)
at com.jvmtop.monitor.VMInfo.attachToVM(Unknown Source)
at com.jvmtop.monitor.VMInfo.processNewVM(Unknown Source)
at com.jvmtop.view.VMOverviewView.scanForNewVMs(Unknown Source)
at com.jvmtop.view.VMOverviewView.printView(Unknown Source)
at com.jvmtop.JvmTop.run(Unknown Source)
at com.jvmtop.JvmTop.main(Unknown Source)
Caused by: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method)
at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:351)
at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:213)
at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:200)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
at java.net.Socket.connect(Socket.java:529)
at java.net.Socket.connect(Socket.java:478)
at java.net.Socket.<init>(Socket.java:375)
at java.net.Socket.<init>(Socket.java:189)
at sun.rmi.transport.proxy.RMIDirectSocketFactory.createSocket(RMIDirectSocketFactory.java:22)
at sun.rmi.transport.proxy.RMIMasterSocketFactory.createSocket(RMIMasterSocketFactory.java:128)
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:595)
... 16 more
JvmTop 0.8.0 alpha - 12:28:59, amd64, 2 cpus, Linux 2.6.27.19, load avg 0.03
http://code.google.com/p/jvmtop
PID MAIN-CLASS HPCUR HPMAX NHCUR NHMAX CPU GC VM USERNAME #T DL
14392 m.jvmtop.JvmTop 12m 1747m 10m 130m 57.69% 0.00% S6U45 root 13
25393 artup.Bootstrap [ERROR: Connection refused/access denied]
# ps aux | grep tomcat | grep -v grep
1001 25393 1.2 16.1 2801376 1299808 pts/9 Sl Sep15 276:12
/home/app_env_1/APP/Java/jdk1.6.0_45/jre/bin/java
-Djava.util.logging.config.file=/home/app_env_1/APP/app-manager/conf/server/logg
ing.properties
-Djava.util.logging.manager=org.apache.juli.ClassLoaderLogManager
-Dapp.conf=/home/app_env_1/APP/app-manager/conf -server -Xms1G -Xmx1G
-XX:+UseCodeCacheFlushing -XX:ReservedCodeCacheSize=64M
-XX:+HeapDumpOnOutOfMemoryError -XX:MaxPermSize=1024M
-Djavax.net.debug=ssl,handshake -Dcom.sun.management.jmxremote
-Dcom.sun.management.jmxremote.port=8004
-Dcom.sun.management.jmxremote.authenticate=false
-Dcom.sun.management.jmxremote.ssl=false
-Djava.rmi.server.hostname=10.249.140.239 -Denv=KIEV_DEV
-Doutput.dir=/home/app_env_1/APP/logs
-Dapp.app=/home/app_env_1/APP/app-application
-Dapp.conf=/home/app_env_1/app/limits-manager/conf -Dapp.version=4.7.0.qa.99
-Djava.endorsed.dirs=/home/app_env_1/app/app-server/common/endorsed -classpath
/home/app_env_1/APP/app-server/bin/bootstrap.jar:/home/app_env_1/APP/app-server/
bin/tomcat-juli.jar -Dcatalina.base=/home/app_env_1/APP/app-server
-Dcatalina.home=/home/app_env_1/APP/app-server
-Djava.io.tmpdir=/home/app_env_1/APP/app-server/temp
org.apache.catalina.startup.Bootstrap -config
/home/app_env_1/APP/app-manager/conf/server/server.xml start
Tomcat version 5.5.23 (yes, I know it is too old, but we must use it).
jvmtop 0.8.0
Same error if run jvmtop under Tomcat's user:
$ id
uid=1001(app_env_1)
$ ./jvmtop.sh --once --verbose
Sep 30, 2014 12:36:38 PM com.jvmtop.JvmTop main
FINE: Verbosity mode.
Sep 30, 2014 12:36:39 PM com.jvmtop.monitor.VMInfo attachToVM
FINE: connection refused (PID=25393)
java.rmi.ConnectException: Connection refused to host: 10.249.140.239; nested
exception is:
java.net.ConnectException: Connection refused
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:601)
at sun.rmi.transport.tcp.TCPChannel.createConnection(TCPChannel.java:198)
at sun.rmi.transport.tcp.TCPChannel.newConnection(TCPChannel.java:184)
at sun.rmi.server.UnicastRef.invoke(UnicastRef.java:110)
at javax.management.remote.rmi.RMIServerImpl_Stub.newClient(Unknown Source)
at javax.management.remote.rmi.RMIConnector.getConnection(RMIConnector.java:2327)
at javax.management.remote.rmi.RMIConnector.connect(RMIConnector.java:277)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:248)
at javax.management.remote.JMXConnectorFactory.connect(JMXConnectorFactory.java:207)
at com.jvmtop.openjdk.tools.ProxyClient.tryConnect(Unknown Source)
at com.jvmtop.openjdk.tools.ProxyClient.connect(Unknown Source)
at com.jvmtop.monitor.VMInfo.attachToVM(Unknown Source)
at com.jvmtop.monitor.VMInfo.processNewVM(Unknown Source)
at com.jvmtop.view.VMOverviewView.scanForNewVMs(Unknown Source)
at com.jvmtop.view.VMOverviewView.printView(Unknown Source)
at com.jvmtop.JvmTop.run(Unknown Source)
at com.jvmtop.JvmTop.main(Unknown Source)
Caused by: java.net.ConnectException: Connection refused
at java.net.PlainSocketImpl.socketConnect(Native Method)
at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:351)
at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:213)
at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:200)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
at java.net.Socket.connect(Socket.java:529)
at java.net.Socket.connect(Socket.java:478)
at java.net.Socket.<init>(Socket.java:375)
at java.net.Socket.<init>(Socket.java:189)
at sun.rmi.transport.proxy.RMIDirectSocketFactory.createSocket(RMIDirectSocketFactory.java:22)
at sun.rmi.transport.proxy.RMIMasterSocketFactory.createSocket(RMIMasterSocketFactory.java:128)
at sun.rmi.transport.tcp.TCPEndpoint.newSocket(TCPEndpoint.java:595)
... 16 more
JvmTop 0.8.0 alpha - 12:36:38, amd64, 2 cpus, Linux 2.6.27.19, load avg 0.03
http://code.google.com/p/jvmtop
PID MAIN-CLASS HPCUR HPMAX NHCUR NHMAX CPU GC VM USERNAME #T DL
14520 m.jvmtop.JvmTop 7m 1747m 10m 130m 80.65% 0.00% S6U45 lmt_env_ 13
25393 artup.Bootstrap [ERROR: Connection refused/access denied]
User's Java:
$ echo $JAVA_HOME
/home/app_env_1/LMS/Java/jdk1.6.0_45/
root's Java:
# echo $JAVA_HOME
/home/app_env_1/LMS/Java/jdk1.6.0_45/
```
Original issue reported on code.google.com by `setev...@gmail.com` on 30 Sep 2014 at 9:40 | defect | error connection refused access denied cat etc suse release suse linux enterprise server version patchlevel java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode echo java home home app env app java jvmtop sh once verbose sep pm com jvmtop jvmtop main fine verbosity mode sep pm com jvmtop monitor vminfo attachtovm fine connection refused pid java rmi connectexception connection refused to host nested exception is java net connectexception connection refused at sun rmi transport tcp tcpendpoint newsocket tcpendpoint java at sun rmi transport tcp tcpchannel createconnection tcpchannel java at sun rmi transport tcp tcpchannel newconnection tcpchannel java at sun rmi server unicastref invoke unicastref java at javax management remote rmi rmiserverimpl stub newclient unknown source at javax management remote rmi rmiconnector getconnection rmiconnector java at javax management remote rmi rmiconnector connect rmiconnector java at javax management remote jmxconnectorfactory connect jmxconnectorfactory java at javax management remote jmxconnectorfactory connect jmxconnectorfactory java at com jvmtop openjdk tools proxyclient tryconnect unknown source at com jvmtop openjdk tools proxyclient connect unknown source at com jvmtop monitor vminfo attachtovm unknown source at com jvmtop monitor vminfo processnewvm unknown source at com jvmtop view vmoverviewview scanfornewvms unknown source at com jvmtop view vmoverviewview printview unknown source at com jvmtop jvmtop run unknown source at com jvmtop jvmtop main unknown source caused by java net connectexception connection refused at java net plainsocketimpl socketconnect native method at java net plainsocketimpl doconnect plainsocketimpl java at java net plainsocketimpl connecttoaddress plainsocketimpl java at java net plainsocketimpl connect plainsocketimpl java at java net sockssocketimpl connect sockssocketimpl java at java net socket connect socket java at java net socket connect socket java at java net socket socket java at java net socket socket java at sun rmi transport proxy rmidirectsocketfactory createsocket rmidirectsocketfactory java at sun rmi transport proxy rmimastersocketfactory createsocket rmimastersocketfactory java at sun rmi transport tcp tcpendpoint newsocket tcpendpoint java more jvmtop alpha cpus linux load avg pid main class hpcur hpmax nhcur nhmax cpu gc vm username t dl m jvmtop jvmtop root artup bootstrap ps aux grep tomcat grep v grep pts sl home app env app java jre bin java djava util logging config file home app env app app manager conf server logg ing properties djava util logging manager org apache juli classloaderlogmanager dapp conf home app env app app manager conf server xx usecodecacheflushing xx reservedcodecachesize xx heapdumponoutofmemoryerror xx maxpermsize djavax net debug ssl handshake dcom sun management jmxremote dcom sun management jmxremote port dcom sun management jmxremote authenticate false dcom sun management jmxremote ssl false djava rmi server hostname denv kiev dev doutput dir home app env app logs dapp app home app env app app application dapp conf home app env app limits manager conf dapp version qa djava endorsed dirs home app env app app server common endorsed classpath home app env app app server bin bootstrap jar home app env app app server bin tomcat juli jar dcatalina base home app env app app server dcatalina home home app env app app server djava io tmpdir home app env app app server temp org apache catalina startup bootstrap config home app env app app manager conf server server xml start tomcat version yes i know it is too old but we must use it jvmtop same error if run jvmtop under tomcat s user id uid app env jvmtop sh once verbose sep pm com jvmtop jvmtop main fine verbosity mode sep pm com jvmtop monitor vminfo attachtovm fine connection refused pid java rmi connectexception connection refused to host nested exception is java net connectexception connection refused at sun rmi transport tcp tcpendpoint newsocket tcpendpoint java at sun rmi transport tcp tcpchannel createconnection tcpchannel java at sun rmi transport tcp tcpchannel newconnection tcpchannel java at sun rmi server unicastref invoke unicastref java at javax management remote rmi rmiserverimpl stub newclient unknown source at javax management remote rmi rmiconnector getconnection rmiconnector java at javax management remote rmi rmiconnector connect rmiconnector java at javax management remote jmxconnectorfactory connect jmxconnectorfactory java at javax management remote jmxconnectorfactory connect jmxconnectorfactory java at com jvmtop openjdk tools proxyclient tryconnect unknown source at com jvmtop openjdk tools proxyclient connect unknown source at com jvmtop monitor vminfo attachtovm unknown source at com jvmtop monitor vminfo processnewvm unknown source at com jvmtop view vmoverviewview scanfornewvms unknown source at com jvmtop view vmoverviewview printview unknown source at com jvmtop jvmtop run unknown source at com jvmtop jvmtop main unknown source caused by java net connectexception connection refused at java net plainsocketimpl socketconnect native method at java net plainsocketimpl doconnect plainsocketimpl java at java net plainsocketimpl connecttoaddress plainsocketimpl java at java net plainsocketimpl connect plainsocketimpl java at java net sockssocketimpl connect sockssocketimpl java at java net socket connect socket java at java net socket connect socket java at java net socket socket java at java net socket socket java at sun rmi transport proxy rmidirectsocketfactory createsocket rmidirectsocketfactory java at sun rmi transport proxy rmimastersocketfactory createsocket rmimastersocketfactory java at sun rmi transport tcp tcpendpoint newsocket tcpendpoint java more jvmtop alpha cpus linux load avg pid main class hpcur hpmax nhcur nhmax cpu gc vm username t dl m jvmtop jvmtop lmt env artup bootstrap user s java echo java home home app env lms java root s java echo java home home app env lms java original issue reported on code google com by setev gmail com on sep at | 1 |
46,027 | 13,055,841,083 | IssuesEvent | 2020-07-30 02:53:43 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | steamshovel screenshots and VirtualBox (Trac #465) | Incomplete Migration Migrated from Trac defect steamshovel | Migrated from https://code.icecube.wisc.edu/ticket/465
```json
{
"status": "closed",
"changetime": "2013-07-08T19:14:28",
"description": "When trying to test screenshot functionality on a Debian Testing VM in VirtualBox, I get this error. Something appears to be wrong with the VirtualBox graphics driver, since this is a common problem for multiple projects.\n\n\n{{{\nIn [3]: window.gl.screenshot(1024,768,'test.png')\nOpenGL Warning: glXChooseFBConfig returning NULL, due to attrib=0x8010, next=0x4\nQGLPixelBuffer: Unable to find a context/format match - giving up.\nlibpng warning: Image width is zero in IHDR\nlibpng warning: Image height is zero in IHDR\nlibpng error: Invalid IHDR data\nOpenGL Warning: glXDestroyPbuffer not implemented by Chromium\n\nIn [4]: window.gl.screenshot(1024,768,'test.jpg')\nOpenGL Warning: glXChooseFBConfig returning NULL, due to attrib=0x8010, next=0x4\nQGLPixelBuffer: Unable to find a context/format match - giving up.\nEmpty JPEG image (DNL not supported)\nOpenGL Warning: glXDestroyPbuffer not implemented by Chromium\n}}}\n",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1373310868000000",
"component": "steamshovel",
"summary": "steamshovel screenshots and VirtualBox",
"priority": "minor",
"keywords": "",
"time": "2013-06-12T17:26:05",
"milestone": "",
"owner": "sjackso",
"type": "defect"
}
```
| 1.0 | steamshovel screenshots and VirtualBox (Trac #465) - Migrated from https://code.icecube.wisc.edu/ticket/465
```json
{
"status": "closed",
"changetime": "2013-07-08T19:14:28",
"description": "When trying to test screenshot functionality on a Debian Testing VM in VirtualBox, I get this error. Something appears to be wrong with the VirtualBox graphics driver, since this is a common problem for multiple projects.\n\n\n{{{\nIn [3]: window.gl.screenshot(1024,768,'test.png')\nOpenGL Warning: glXChooseFBConfig returning NULL, due to attrib=0x8010, next=0x4\nQGLPixelBuffer: Unable to find a context/format match - giving up.\nlibpng warning: Image width is zero in IHDR\nlibpng warning: Image height is zero in IHDR\nlibpng error: Invalid IHDR data\nOpenGL Warning: glXDestroyPbuffer not implemented by Chromium\n\nIn [4]: window.gl.screenshot(1024,768,'test.jpg')\nOpenGL Warning: glXChooseFBConfig returning NULL, due to attrib=0x8010, next=0x4\nQGLPixelBuffer: Unable to find a context/format match - giving up.\nEmpty JPEG image (DNL not supported)\nOpenGL Warning: glXDestroyPbuffer not implemented by Chromium\n}}}\n",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1373310868000000",
"component": "steamshovel",
"summary": "steamshovel screenshots and VirtualBox",
"priority": "minor",
"keywords": "",
"time": "2013-06-12T17:26:05",
"milestone": "",
"owner": "sjackso",
"type": "defect"
}
```
| defect | steamshovel screenshots and virtualbox trac migrated from json status closed changetime description when trying to test screenshot functionality on a debian testing vm in virtualbox i get this error something appears to be wrong with the virtualbox graphics driver since this is a common problem for multiple projects n n n nin window gl screenshot test png nopengl warning glxchoosefbconfig returning null due to attrib next nqglpixelbuffer unable to find a context format match giving up nlibpng warning image width is zero in ihdr nlibpng warning image height is zero in ihdr nlibpng error invalid ihdr data nopengl warning glxdestroypbuffer not implemented by chromium n nin window gl screenshot test jpg nopengl warning glxchoosefbconfig returning null due to attrib next nqglpixelbuffer unable to find a context format match giving up nempty jpeg image dnl not supported nopengl warning glxdestroypbuffer not implemented by chromium n n reporter david schultz cc resolution fixed ts component steamshovel summary steamshovel screenshots and virtualbox priority minor keywords time milestone owner sjackso type defect | 1 |
10,647 | 2,622,178,781 | IssuesEvent | 2015-03-04 00:17:51 | byzhang/leveldb | https://api.github.com/repos/byzhang/leveldb | closed | memory allocated for default env is not deallocated | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. create db
2. delete db
3. run the program using valgrind.
Options options_;
options_.create_if_missing = true;
options_.block_cache = cache_; // uses the default
options_.write_buffer_size = 0;
Status s = DB::Open(options_, dbRoot, &db_);
delete db_;
What is the expected output? What do you see instead?
Expected no memory leak.
==605== 512 bytes in 1 blocks are still reachable in loss record 11 of 13
==605== at 0x4C274B4: operator new(unsigned long) (vg_replace_malloc.c:287)
==605== by 0x79836C2: leveldb::InitDefaultEnv() (in
/usr/local/maui/lib/libleveldb.so.1)
==605== by 0x8968B52: pthread_once (in /lib64/libpthread-2.11.1.so)
==605== by 0x7982A26: leveldb::Env::Default() (in
/usr/local/maui/lib/libleveldb.so.1)
==605== by 0x7985FFC: leveldb::Options::Options() (in
/usr/local/maui/lib/libleveldb.so.1)
==605== by 0x4EAFBB1: KVStoreStage::KVStoreStage(char const*)
(kvstorestage.cxx:44)
==605== by 0x4EAFCFF: KVStoreStage::makeStage(std::string const&,
Properties*) (kvstorestage.cxx:252)
==605== by 0x5658525: ClassFactory<Stage>::makeInstance(std::string const&,
Properties*) (classfactory.hxx:150)
==605== by 0x5656945: SedaConfig::instantiate() (sedaconfig.cxx:334)
==605== by 0x5658388: SedaConfig::init() (sedaconfig.cxx:215)
==605== by 0x464259: main (kvs.cxx:120)
env.h
class Env {
public:
Env() { }
virtual ~Env();
// Return a default environment suitable for the current operating
// system. Sophisticated users may wish to provide their own Env
// implementation instead of relying on this default environment.
//
// The result of Default() belongs to leveldb and must never be deleted.
static Env* Default();
What version of the product are you using? On what operating system?
1.5.0, OS Suse Linux 11
Please provide any additional information below.
```
Original issue reported on code.google.com by `gsai2...@gmail.com` on 27 Jul 2012 at 12:38 | 1.0 | memory allocated for default env is not deallocated - ```
What steps will reproduce the problem?
1. create db
2. delete db
3. run the program using valgrind.
Options options_;
options_.create_if_missing = true;
options_.block_cache = cache_; // uses the default
options_.write_buffer_size = 0;
Status s = DB::Open(options_, dbRoot, &db_);
delete db_;
What is the expected output? What do you see instead?
Expected no memory leak.
==605== 512 bytes in 1 blocks are still reachable in loss record 11 of 13
==605== at 0x4C274B4: operator new(unsigned long) (vg_replace_malloc.c:287)
==605== by 0x79836C2: leveldb::InitDefaultEnv() (in
/usr/local/maui/lib/libleveldb.so.1)
==605== by 0x8968B52: pthread_once (in /lib64/libpthread-2.11.1.so)
==605== by 0x7982A26: leveldb::Env::Default() (in
/usr/local/maui/lib/libleveldb.so.1)
==605== by 0x7985FFC: leveldb::Options::Options() (in
/usr/local/maui/lib/libleveldb.so.1)
==605== by 0x4EAFBB1: KVStoreStage::KVStoreStage(char const*)
(kvstorestage.cxx:44)
==605== by 0x4EAFCFF: KVStoreStage::makeStage(std::string const&,
Properties*) (kvstorestage.cxx:252)
==605== by 0x5658525: ClassFactory<Stage>::makeInstance(std::string const&,
Properties*) (classfactory.hxx:150)
==605== by 0x5656945: SedaConfig::instantiate() (sedaconfig.cxx:334)
==605== by 0x5658388: SedaConfig::init() (sedaconfig.cxx:215)
==605== by 0x464259: main (kvs.cxx:120)
env.h
class Env {
public:
Env() { }
virtual ~Env();
// Return a default environment suitable for the current operating
// system. Sophisticated users may wish to provide their own Env
// implementation instead of relying on this default environment.
//
// The result of Default() belongs to leveldb and must never be deleted.
static Env* Default();
What version of the product are you using? On what operating system?
1.5.0, OS Suse Linux 11
Please provide any additional information below.
```
Original issue reported on code.google.com by `gsai2...@gmail.com` on 27 Jul 2012 at 12:38 | defect | memory allocated for default env is not deallocated what steps will reproduce the problem create db delete db run the program using valgrind options options options create if missing true options block cache cache uses the default options write buffer size status s db open options dbroot db delete db what is the expected output what do you see instead expected no memory leak bytes in blocks are still reachable in loss record of at operator new unsigned long vg replace malloc c by leveldb initdefaultenv in usr local maui lib libleveldb so by pthread once in libpthread so by leveldb env default in usr local maui lib libleveldb so by leveldb options options in usr local maui lib libleveldb so by kvstorestage kvstorestage char const kvstorestage cxx by kvstorestage makestage std string const properties kvstorestage cxx by classfactory makeinstance std string const properties classfactory hxx by sedaconfig instantiate sedaconfig cxx by sedaconfig init sedaconfig cxx by main kvs cxx env h class env public env virtual env return a default environment suitable for the current operating system sophisticated users may wish to provide their own env implementation instead of relying on this default environment the result of default belongs to leveldb and must never be deleted static env default what version of the product are you using on what operating system os suse linux please provide any additional information below original issue reported on code google com by gmail com on jul at | 1 |
115,473 | 9,796,482,136 | IssuesEvent | 2019-06-11 07:45:07 | microsoft/azure-pipelines-tasks | https://api.github.com/repos/microsoft/azure-pipelines-tasks | closed | VSTest Task Fails with System.TimeoutException: The HTTP request timed out after 00:01:40 | Area: Test | **Setup:**
TFS Update 3 on-premises
vsts agent 2.140.2
VsTest installed via Test Platform Installer (15.9.1 preview)
Visual Studio Test task Version 2.*
- Select tests using Test Plan
- multi-agent parallelism
**Issue:**
I am seeing test runs aborting in the middle, I believe due to the following error:
_2018-10-25T05:58:16.5424818Z ##[error]Failure attempting to call the restapis. Method: System.Threading.Tasks.Task`1[Microsoft.TeamFoundation.Test.WebApi.TestExecutionServiceCommand] <GetCommand>b__0(), Exception: System.AggregateException: One or more errors occurred. ---> System.TimeoutException: The HTTP request timed out after 00:01:40. ---> System.Threading.Tasks.TaskCanceledException: A task was canceled._
We see this on multiple test runs quite often - our setup is new and have seen this each time we try to run all of our tests (over 1200 test points across multiple environments). I cannot yet tell a pattern of when it happens. This particular test run was using 7 test agents. Two of the test agents completed their tests, one of them hadn't started running tests yet, and out of the remaining 4 agents, 3 of them are all similar logs with other agent appearing different, so I suspect it was that different agent that encountered the issue that caused the problem. Here are the logs around the failing section for that agent:
[HttpTimeoutLogs.txt](https://github.com/Microsoft/azure-pipelines-tasks/files/2515984/HttpTimeoutLogs.txt)
Please let me know if you need any other information. | 1.0 | VSTest Task Fails with System.TimeoutException: The HTTP request timed out after 00:01:40 - **Setup:**
TFS Update 3 on-premises
vsts agent 2.140.2
VsTest installed via Test Platform Installer (15.9.1 preview)
Visual Studio Test task Version 2.*
- Select tests using Test Plan
- multi-agent parallelism
**Issue:**
I am seeing test runs aborting in the middle, I believe due to the following error:
_2018-10-25T05:58:16.5424818Z ##[error]Failure attempting to call the restapis. Method: System.Threading.Tasks.Task`1[Microsoft.TeamFoundation.Test.WebApi.TestExecutionServiceCommand] <GetCommand>b__0(), Exception: System.AggregateException: One or more errors occurred. ---> System.TimeoutException: The HTTP request timed out after 00:01:40. ---> System.Threading.Tasks.TaskCanceledException: A task was canceled._
We see this on multiple test runs quite often - our setup is new and have seen this each time we try to run all of our tests (over 1200 test points across multiple environments). I cannot yet tell a pattern of when it happens. This particular test run was using 7 test agents. Two of the test agents completed their tests, one of them hadn't started running tests yet, and out of the remaining 4 agents, 3 of them are all similar logs with other agent appearing different, so I suspect it was that different agent that encountered the issue that caused the problem. Here are the logs around the failing section for that agent:
[HttpTimeoutLogs.txt](https://github.com/Microsoft/azure-pipelines-tasks/files/2515984/HttpTimeoutLogs.txt)
Please let me know if you need any other information. | non_defect | vstest task fails with system timeoutexception the http request timed out after setup tfs update on premises vsts agent vstest installed via test platform installer preview visual studio test task version select tests using test plan multi agent parallelism issue i am seeing test runs aborting in the middle i believe due to the following error failure attempting to call the restapis method system threading tasks task b exception system aggregateexception one or more errors occurred system timeoutexception the http request timed out after system threading tasks taskcanceledexception a task was canceled we see this on multiple test runs quite often our setup is new and have seen this each time we try to run all of our tests over test points across multiple environments i cannot yet tell a pattern of when it happens this particular test run was using test agents two of the test agents completed their tests one of them hadn t started running tests yet and out of the remaining agents of them are all similar logs with other agent appearing different so i suspect it was that different agent that encountered the issue that caused the problem here are the logs around the failing section for that agent please let me know if you need any other information | 0 |
173,627 | 27,500,843,192 | IssuesEvent | 2023-03-05 17:24:46 | hashicorp/terraform | https://api.github.com/repos/hashicorp/terraform | closed | provider cache becomes ineffective with 1.4.0-alpha release | bug new working as designed v1.4 | ### Terraform Version
```shell
Terraform v1.4.0-alpha20221109
on linux_amd64
```
### Terraform Configuration Files
Enable the global provider cache by setting `TF_PLUGIN_CACHE_DIR`.
### Debug Output
Not really applicable.
### Expected Behavior
The global provider cache should be used when initializing a terraform repository. We use terragrunt to initialize and run hundreds of terraform repositories simultaneously and rely on the global cache to optimize this process. We ensure that the provider cache is fully populated first because of the lack of concurrency support, but once it has been populated, we rely on terraform being able to utilize this to initialize by simply reading from the cache.
This is what 1.3.x shows and is what is expected:
```
Initializing provider plugins...
- Finding latest version of opsgenie/opsgenie...
- Finding latest version of jfrog/artifactory...
- Finding latest version of sumologic/sumologic...
- Finding latest version of hashicorp/template...
- Finding latest version of hashicorp/aws...
- Finding latest version of hashicorp/random...
- Finding latest version of hashicorp/vault...
- Finding latest version of hashicorp/external...
- Finding latest version of hashicorp/kubernetes...
- Finding latest version of cloudflare/cloudflare...
- Finding latest version of hashicorp/cloudinit...
- Finding latest version of okta/okta...
- Finding latest version of hashicorp/helm...
- Using cloudflare/cloudflare v3.27.0 from the shared cache directory
- Using okta/okta v3.38.0 from the shared cache directory
- Using hashicorp/helm v2.7.1 from the shared cache directory
- Using hashicorp/template v2.2.0 from the shared cache directory
- Using hashicorp/random v3.4.3 from the shared cache directory
- Using sumologic/sumologic v2.19.2 from the shared cache directory
- Using hashicorp/aws v4.38.0 from the shared cache directory
- Using hashicorp/vault v3.10.0 from the shared cache directory
- Using hashicorp/external v2.2.3 from the shared cache directory
- Using hashicorp/kubernetes v2.15.0 from the shared cache directory
- Using hashicorp/cloudinit v2.2.0 from the shared cache directory
- Using opsgenie/opsgenie v0.6.10 from the shared cache directory
- Using jfrog/artifactory v6.19.0 from the shared cache directory
```
### Actual Behavior
Terraform tries to re-populate the provider in the global cache, with the exact same file. Because this process is not concurrency-safe, this causes problems when we use terragrunt and we have multiple terraform initializations running simultaneously. It leads to checksum errors and failures.
With 1.4.0-alpha, initialization of every repository shows:
```
Initializing provider plugins...
- Finding latest version of hashicorp/external...
- Finding latest version of jfrog/artifactory...
- Finding latest version of sumologic/sumologic...
- Finding latest version of okta/okta...
- Finding latest version of hashicorp/cloudinit...
- Finding latest version of hashicorp/helm...
- Finding latest version of hashicorp/aws...
- Finding latest version of opsgenie/opsgenie...
- Finding latest version of hashicorp/kubernetes...
- Finding latest version of hashicorp/vault...
- Finding latest version of hashicorp/random...
- Finding latest version of cloudflare/cloudflare...
- Finding latest version of hashicorp/template...
- Installing hashicorp/random v3.4.3...
- Installed hashicorp/random v3.4.3 (unauthenticated)
- Installing okta/okta v3.38.0...
- Installed okta/okta v3.38.0 (unauthenticated)
- Installing hashicorp/helm v2.7.1...
- Installed hashicorp/helm v2.7.1 (unauthenticated)
- Installing hashicorp/aws v4.38.0...
- Installed hashicorp/aws v4.38.0 (unauthenticated)
- Installing hashicorp/vault v3.10.0...
- Installed hashicorp/vault v3.10.0 (unauthenticated)
- Installing opsgenie/opsgenie v0.6.10...
- Installed opsgenie/opsgenie v0.6.10 (unauthenticated)
- Installing hashicorp/kubernetes v2.15.0...
- Installed hashicorp/kubernetes v2.15.0 (unauthenticated)
- Installing cloudflare/cloudflare v3.27.0...
- Installed cloudflare/cloudflare v3.27.0 (unauthenticated)
- Installing hashicorp/template v2.2.0...
- Installed hashicorp/template v2.2.0 (unauthenticated)
- Installing hashicorp/external v2.2.3...
- Installed hashicorp/external v2.2.3 (unauthenticated)
- Installing jfrog/artifactory v6.19.0...
- Installed jfrog/artifactory v6.19.0 (unauthenticated)
- Installing sumologic/sumologic v2.19.2...
- Installed sumologic/sumologic v2.19.2 (unauthenticated)
- Installing hashicorp/cloudinit v2.2.0...
- Installed hashicorp/cloudinit v2.2.0 (unauthenticated)
```
### Steps to Reproduce
Initialize two separate terraform repositories with the global provider cache enabled. Both will install the same providers into the cache rather than re-using them.
### Additional Context
_No response_
### References
- #32129 | 1.0 | provider cache becomes ineffective with 1.4.0-alpha release - ### Terraform Version
```shell
Terraform v1.4.0-alpha20221109
on linux_amd64
```
### Terraform Configuration Files
Enable the global provider cache by setting `TF_PLUGIN_CACHE_DIR`.
### Debug Output
Not really applicable.
### Expected Behavior
The global provider cache should be used when initializing a terraform repository. We use terragrunt to initialize and run hundreds of terraform repositories simultaneously and rely on the global cache to optimize this process. We ensure that the provider cache is fully populated first because of the lack of concurrency support, but once it has been populated, we rely on terraform being able to utilize this to initialize by simply reading from the cache.
This is what 1.3.x shows and is what is expected:
```
Initializing provider plugins...
- Finding latest version of opsgenie/opsgenie...
- Finding latest version of jfrog/artifactory...
- Finding latest version of sumologic/sumologic...
- Finding latest version of hashicorp/template...
- Finding latest version of hashicorp/aws...
- Finding latest version of hashicorp/random...
- Finding latest version of hashicorp/vault...
- Finding latest version of hashicorp/external...
- Finding latest version of hashicorp/kubernetes...
- Finding latest version of cloudflare/cloudflare...
- Finding latest version of hashicorp/cloudinit...
- Finding latest version of okta/okta...
- Finding latest version of hashicorp/helm...
- Using cloudflare/cloudflare v3.27.0 from the shared cache directory
- Using okta/okta v3.38.0 from the shared cache directory
- Using hashicorp/helm v2.7.1 from the shared cache directory
- Using hashicorp/template v2.2.0 from the shared cache directory
- Using hashicorp/random v3.4.3 from the shared cache directory
- Using sumologic/sumologic v2.19.2 from the shared cache directory
- Using hashicorp/aws v4.38.0 from the shared cache directory
- Using hashicorp/vault v3.10.0 from the shared cache directory
- Using hashicorp/external v2.2.3 from the shared cache directory
- Using hashicorp/kubernetes v2.15.0 from the shared cache directory
- Using hashicorp/cloudinit v2.2.0 from the shared cache directory
- Using opsgenie/opsgenie v0.6.10 from the shared cache directory
- Using jfrog/artifactory v6.19.0 from the shared cache directory
```
### Actual Behavior
Terraform tries to re-populate the provider in the global cache, with the exact same file. Because this process is not concurrency-safe, this causes problems when we use terragrunt and we have multiple terraform initializations running simultaneously. It leads to checksum errors and failures.
With 1.4.0-alpha, initialization of every repository shows:
```
Initializing provider plugins...
- Finding latest version of hashicorp/external...
- Finding latest version of jfrog/artifactory...
- Finding latest version of sumologic/sumologic...
- Finding latest version of okta/okta...
- Finding latest version of hashicorp/cloudinit...
- Finding latest version of hashicorp/helm...
- Finding latest version of hashicorp/aws...
- Finding latest version of opsgenie/opsgenie...
- Finding latest version of hashicorp/kubernetes...
- Finding latest version of hashicorp/vault...
- Finding latest version of hashicorp/random...
- Finding latest version of cloudflare/cloudflare...
- Finding latest version of hashicorp/template...
- Installing hashicorp/random v3.4.3...
- Installed hashicorp/random v3.4.3 (unauthenticated)
- Installing okta/okta v3.38.0...
- Installed okta/okta v3.38.0 (unauthenticated)
- Installing hashicorp/helm v2.7.1...
- Installed hashicorp/helm v2.7.1 (unauthenticated)
- Installing hashicorp/aws v4.38.0...
- Installed hashicorp/aws v4.38.0 (unauthenticated)
- Installing hashicorp/vault v3.10.0...
- Installed hashicorp/vault v3.10.0 (unauthenticated)
- Installing opsgenie/opsgenie v0.6.10...
- Installed opsgenie/opsgenie v0.6.10 (unauthenticated)
- Installing hashicorp/kubernetes v2.15.0...
- Installed hashicorp/kubernetes v2.15.0 (unauthenticated)
- Installing cloudflare/cloudflare v3.27.0...
- Installed cloudflare/cloudflare v3.27.0 (unauthenticated)
- Installing hashicorp/template v2.2.0...
- Installed hashicorp/template v2.2.0 (unauthenticated)
- Installing hashicorp/external v2.2.3...
- Installed hashicorp/external v2.2.3 (unauthenticated)
- Installing jfrog/artifactory v6.19.0...
- Installed jfrog/artifactory v6.19.0 (unauthenticated)
- Installing sumologic/sumologic v2.19.2...
- Installed sumologic/sumologic v2.19.2 (unauthenticated)
- Installing hashicorp/cloudinit v2.2.0...
- Installed hashicorp/cloudinit v2.2.0 (unauthenticated)
```
### Steps to Reproduce
Initialize two separate terraform repositories with the global provider cache enabled. Both will install the same providers into the cache rather than re-using them.
### Additional Context
_No response_
### References
- #32129 | non_defect | provider cache becomes ineffective with alpha release terraform version shell terraform on linux terraform configuration files enable the global provider cache by setting tf plugin cache dir debug output not really applicable expected behavior the global provider cache should be used when initializing a terraform repository we use terragrunt to initialize and run hundreds of terraform repositories simultaneously and rely on the global cache to optimize this process we ensure that the provider cache is fully populated first because of the lack of concurrency support but once it has been populated we rely on terraform being able to utilize this to initialize by simply reading from the cache this is what x shows and is what is expected initializing provider plugins finding latest version of opsgenie opsgenie finding latest version of jfrog artifactory finding latest version of sumologic sumologic finding latest version of hashicorp template finding latest version of hashicorp aws finding latest version of hashicorp random finding latest version of hashicorp vault finding latest version of hashicorp external finding latest version of hashicorp kubernetes finding latest version of cloudflare cloudflare finding latest version of hashicorp cloudinit finding latest version of okta okta finding latest version of hashicorp helm using cloudflare cloudflare from the shared cache directory using okta okta from the shared cache directory using hashicorp helm from the shared cache directory using hashicorp template from the shared cache directory using hashicorp random from the shared cache directory using sumologic sumologic from the shared cache directory using hashicorp aws from the shared cache directory using hashicorp vault from the shared cache directory using hashicorp external from the shared cache directory using hashicorp kubernetes from the shared cache directory using hashicorp cloudinit from the shared cache directory using opsgenie opsgenie from the shared cache directory using jfrog artifactory from the shared cache directory actual behavior terraform tries to re populate the provider in the global cache with the exact same file because this process is not concurrency safe this causes problems when we use terragrunt and we have multiple terraform initializations running simultaneously it leads to checksum errors and failures with alpha initialization of every repository shows initializing provider plugins finding latest version of hashicorp external finding latest version of jfrog artifactory finding latest version of sumologic sumologic finding latest version of okta okta finding latest version of hashicorp cloudinit finding latest version of hashicorp helm finding latest version of hashicorp aws finding latest version of opsgenie opsgenie finding latest version of hashicorp kubernetes finding latest version of hashicorp vault finding latest version of hashicorp random finding latest version of cloudflare cloudflare finding latest version of hashicorp template installing hashicorp random installed hashicorp random unauthenticated installing okta okta installed okta okta unauthenticated installing hashicorp helm installed hashicorp helm unauthenticated installing hashicorp aws installed hashicorp aws unauthenticated installing hashicorp vault installed hashicorp vault unauthenticated installing opsgenie opsgenie installed opsgenie opsgenie unauthenticated installing hashicorp kubernetes installed hashicorp kubernetes unauthenticated installing cloudflare cloudflare installed cloudflare cloudflare unauthenticated installing hashicorp template installed hashicorp template unauthenticated installing hashicorp external installed hashicorp external unauthenticated installing jfrog artifactory installed jfrog artifactory unauthenticated installing sumologic sumologic installed sumologic sumologic unauthenticated installing hashicorp cloudinit installed hashicorp cloudinit unauthenticated steps to reproduce initialize two separate terraform repositories with the global provider cache enabled both will install the same providers into the cache rather than re using them additional context no response references | 0 |
48,629 | 5,965,346,422 | IssuesEvent | 2017-05-30 11:20:28 | LDMW/app | https://api.github.com/repos/LDMW/app | closed | 'Success' message styling | please-test | Success message should appear as below:
- [x] Success message replaces/appears in text box
- [x] Tick should be in top right corner
<img width="486" alt="899884e0-3401-11e7-821f-5378e58f81c3" src="https://cloud.githubusercontent.com/assets/25007700/26154717/85209836-3b08-11e7-8289-b51309887402.png">
| 1.0 | 'Success' message styling - Success message should appear as below:
- [x] Success message replaces/appears in text box
- [x] Tick should be in top right corner
<img width="486" alt="899884e0-3401-11e7-821f-5378e58f81c3" src="https://cloud.githubusercontent.com/assets/25007700/26154717/85209836-3b08-11e7-8289-b51309887402.png">
| non_defect | success message styling success message should appear as below success message replaces appears in text box tick should be in top right corner img width alt src | 0 |
45,039 | 12,529,728,169 | IssuesEvent | 2020-06-04 11:52:54 | appknox/google-chartwrapper | https://api.github.com/repos/appknox/google-chartwrapper | closed | Line styles issue | Priority-Medium Type-Defect auto-migrated | ```
With G = GChart('lxy'),
G.line(1,1,0,2,1,0) produces:
G['chls'] : '1.0,1.0,0.0,2.0,1.0,0.0'
This should be:
G['chls'] : '1,1,0|2,1,0', in order to change the line styles correctly
```
Original issue reported on code.google.com by `hvanniekerk` on 21 Dec 2009 at 9:41
| 1.0 | Line styles issue - ```
With G = GChart('lxy'),
G.line(1,1,0,2,1,0) produces:
G['chls'] : '1.0,1.0,0.0,2.0,1.0,0.0'
This should be:
G['chls'] : '1,1,0|2,1,0', in order to change the line styles correctly
```
Original issue reported on code.google.com by `hvanniekerk` on 21 Dec 2009 at 9:41
| defect | line styles issue with g gchart lxy g line produces g this should be g in order to change the line styles correctly original issue reported on code google com by hvanniekerk on dec at | 1 |
28,204 | 5,221,370,996 | IssuesEvent | 2017-01-27 01:13:25 | elTiempoVuela/https-finder | https://api.github.com/repos/elTiempoVuela/https-finder | closed | add on removed from mozilla site | auto-migrated Priority-Medium Type-Defect | ```
fore some reason when i tried to find thw https fonder on mozilla i landed on a
message... add on removed by user . the add on is not listed anywhere
anymore.. please add it back.. its a very valuable add on that contributes
to free and donation based software philoaophy as well to internet security .
Thanks
```
Original issue reported on code.google.com by `dinorapt...@gmail.com` on 15 Feb 2015 at 1:10
| 1.0 | add on removed from mozilla site - ```
fore some reason when i tried to find thw https fonder on mozilla i landed on a
message... add on removed by user . the add on is not listed anywhere
anymore.. please add it back.. its a very valuable add on that contributes
to free and donation based software philoaophy as well to internet security .
Thanks
```
Original issue reported on code.google.com by `dinorapt...@gmail.com` on 15 Feb 2015 at 1:10
| defect | add on removed from mozilla site fore some reason when i tried to find thw https fonder on mozilla i landed on a message add on removed by user the add on is not listed anywhere anymore please add it back its a very valuable add on that contributes to free and donation based software philoaophy as well to internet security thanks original issue reported on code google com by dinorapt gmail com on feb at | 1 |
14,439 | 2,811,895,126 | IssuesEvent | 2015-05-18 02:52:17 | RenatoUtsch/nulldc | https://api.github.com/repos/RenatoUtsch/nulldc | closed | Game works in GDI format but not CDI | auto-migrated Priority-Medium Restrict-AddIssueComment-Commit Type-Defect | ```
Somes games like
chu rocket
ooga booga
works fine/perfectly in gdi format but in cdi format
chu rocket > see as audio CD
ooga booga > block at sega logo
I use NullDC 1.0.4 but sames problems on others emulators
GDRom Plugin: ImgReader_win31.dll (april2012)
I have tried differents cdi and convert GDI to CDI, always the sames bugs.
```
Original issue reported on code.google.com by `hollowma...@gmail.com` on 18 May 2013 at 2:15 | 1.0 | Game works in GDI format but not CDI - ```
Somes games like
chu rocket
ooga booga
works fine/perfectly in gdi format but in cdi format
chu rocket > see as audio CD
ooga booga > block at sega logo
I use NullDC 1.0.4 but sames problems on others emulators
GDRom Plugin: ImgReader_win31.dll (april2012)
I have tried differents cdi and convert GDI to CDI, always the sames bugs.
```
Original issue reported on code.google.com by `hollowma...@gmail.com` on 18 May 2013 at 2:15 | defect | game works in gdi format but not cdi somes games like chu rocket ooga booga works fine perfectly in gdi format but in cdi format chu rocket see as audio cd ooga booga block at sega logo i use nulldc but sames problems on others emulators gdrom plugin imgreader dll i have tried differents cdi and convert gdi to cdi always the sames bugs original issue reported on code google com by hollowma gmail com on may at | 1 |
50,487 | 13,537,041,798 | IssuesEvent | 2020-09-16 09:54:19 | haiwen/seafile | https://api.github.com/repos/haiwen/seafile | closed | Bad permissions for .seafile-data Directory | type-enhancement type-security | On Debian, the .seafile-data Directory has set "read" rights for "others".
I think this is not necessary and if this Directory is "stolen", an attacker can access all files which are not in an encrypted file container.
The Windows Client seems to have the same issue.
| True | Bad permissions for .seafile-data Directory - On Debian, the .seafile-data Directory has set "read" rights for "others".
I think this is not necessary and if this Directory is "stolen", an attacker can access all files which are not in an encrypted file container.
The Windows Client seems to have the same issue.
| non_defect | bad permissions for seafile data directory on debian the seafile data directory has set read rights for others i think this is not necessary and if this directory is stolen an attacker can access all files which are not in an encrypted file container the windows client seems to have the same issue | 0 |
526,929 | 15,305,030,187 | IssuesEvent | 2021-02-24 17:36:19 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | Overriding `DoCalcNextUpdateTime()` might have no effect. | component: simulator priority: high team: dynamics type: bug | From f2f with @sherm1, if we override `DoCalcNextUpdateTime()` to return a non-infinity value of `next_update_time`, our simulator should guarantee that time goes exactly through `next_update_time` whether new events were scheduled or not by `DoCalcNextUpdateTime()`. It seems that most of our overrides in Drake do schedule an event (see `LcmLogPlaybackSystem`,
`LcmSubscriberSystem`).
However `Simulator` uses this time value only if something was scheduled, see [here](https://github.com/RobotLocomotion/drake/blob/cdc37f451e2c97e32fb6670f96ebe0daae70a67a/systems/analysis/simulator.h#L1154). That value of time, in `next_timed_event_time_`, gets also passed to the integrators [here](https://github.com/RobotLocomotion/drake/blob/cdc37f451e2c97e32fb6670f96ebe0daae70a67a/systems/analysis/simulator.h#L1163), however [it gets never used](https://github.com/RobotLocomotion/drake/blob/cdc37f451e2c97e32fb6670f96ebe0daae70a67a/systems/analysis/simulator.h#L1452).
As a case where no new events are scheduled, we have `LcmInterfaceSystem::DoCalcNextUpdateTime()`, which might indicate a bug.
| 1.0 | Overriding `DoCalcNextUpdateTime()` might have no effect. - From f2f with @sherm1, if we override `DoCalcNextUpdateTime()` to return a non-infinity value of `next_update_time`, our simulator should guarantee that time goes exactly through `next_update_time` whether new events were scheduled or not by `DoCalcNextUpdateTime()`. It seems that most of our overrides in Drake do schedule an event (see `LcmLogPlaybackSystem`,
`LcmSubscriberSystem`).
However `Simulator` uses this time value only if something was scheduled, see [here](https://github.com/RobotLocomotion/drake/blob/cdc37f451e2c97e32fb6670f96ebe0daae70a67a/systems/analysis/simulator.h#L1154). That value of time, in `next_timed_event_time_`, gets also passed to the integrators [here](https://github.com/RobotLocomotion/drake/blob/cdc37f451e2c97e32fb6670f96ebe0daae70a67a/systems/analysis/simulator.h#L1163), however [it gets never used](https://github.com/RobotLocomotion/drake/blob/cdc37f451e2c97e32fb6670f96ebe0daae70a67a/systems/analysis/simulator.h#L1452).
As a case where no new events are scheduled, we have `LcmInterfaceSystem::DoCalcNextUpdateTime()`, which might indicate a bug.
| non_defect | overriding docalcnextupdatetime might have no effect from with if we override docalcnextupdatetime to return a non infinity value of next update time our simulator should guarantee that time goes exactly through next update time whether new events were scheduled or not by docalcnextupdatetime it seems that most of our overrides in drake do schedule an event see lcmlogplaybacksystem lcmsubscribersystem however simulator uses this time value only if something was scheduled see that value of time in next timed event time gets also passed to the integrators however as a case where no new events are scheduled we have lcminterfacesystem docalcnextupdatetime which might indicate a bug | 0 |
40,480 | 10,014,716,674 | IssuesEvent | 2019-07-15 18:14:09 | sm0svx/svxlink | https://api.github.com/repos/sm0svx/svxlink | closed | Version files does not always update without running cmake manually | T: defect | When updating a previously compiled source tree, the version files are not always updated which leads to the binaries ending up displaying an old software version number. Work around this by manually running `cmake` before `make`.
```
cd src/build
cmake ..
make
```
Note: You do not have to supply cmake with all arguments you did initially since all setup will be cached. | 1.0 | Version files does not always update without running cmake manually - When updating a previously compiled source tree, the version files are not always updated which leads to the binaries ending up displaying an old software version number. Work around this by manually running `cmake` before `make`.
```
cd src/build
cmake ..
make
```
Note: You do not have to supply cmake with all arguments you did initially since all setup will be cached. | defect | version files does not always update without running cmake manually when updating a previously compiled source tree the version files are not always updated which leads to the binaries ending up displaying an old software version number work around this by manually running cmake before make cd src build cmake make note you do not have to supply cmake with all arguments you did initially since all setup will be cached | 1 |
5,393 | 2,610,186,823 | IssuesEvent | 2015-02-26 18:59:16 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 纠结色斑怎么除掉 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
风也知盘旋,碟恋花相伴翩翩,星也懂缠绵,冰为水情愿己��
�。只有你,狠心走远。我纠结于,连这狂风暴雨般的疼爱都�
��归于平静,还有什么不能归于平静。繁华谢幕,生命崇尚,
盛夏之花。我承认,花很美。但那终究与我无关。流年低吟��
�叹,记忆久酿成伤。生命雨后屋檐下细嗅蔷薇,追忆旅途中�
��冬季畅饮雪扬,走到思绪的尽头,看这一世万丈尘寰。寂寞
在青春泛滥,烟花只得瞬间璀璨,流星唯有煞时迷眼。但是��
�到自己脸上陆续出现的小斑点,我又怎么能安静那!色斑怎�
��除掉,
《客户案例》
说说我的坎坷祛斑路吧,我从小就有黄褐斑,是遗传我��
�的,别人家的小孩的脸都是又白又嫩,红扑扑的,我倒好,�
��脸的小黄褐斑,害得我被喊了十多年的“芝麻小妹”,后来
上学,我们班的卫生委员小姑娘总说我不洗脸,到后来身边��
�姐妹都有男朋友,就只有我孤家寡人一个,这一筐筐的血泪�
��一天一夜都说不完,为了尽快解决我的终身大事,也为了我
那爱美的心,本姑娘决定把自己当小白鼠了,一定要把这该��
�的黄褐斑去掉。<br>
为了能知己知彼百战百胜,我特意上网查查怎么去黄褐��
�,可一查我那颗热血沸腾的心变的洼凉洼凉的,原来遗传性�
��黄褐斑只能被淡化不能完全消除,既然这样就退而求其次吧
,能淡化到看不见总可以吧。我又重振旗鼓,先到网上找个��
�多祛斑偏方,什么西红柿敷脸,牛奶蜂蜜一起上,折腾了半�
��,皮肤确实好点了,可斑还是那样,看来这个方法是不行的
,我又对准那些祛斑产品了,在网上一搜,密密麻麻的十多��
�,心想这么多总有一个管用的吧,什么七日祛斑,三天见效�
��广告做的挺不错,就是不敢相信,最后我觉得还是得听听网
友们的意见,就去论坛上发个帖子,过了几天,我就得到了��
�多回复,那么多回复中,我看很多人提到「黛芙薇尔精华液�
��,说用的效果不错,我当时感觉这个有那么好吗?就去他们��
�方网上详细的询问了客服,他们特别热心,在听我说了我的�
��况后,给了我很多祛斑的建议,我特别感动,当时就试着订
购了两个周期,过后感觉自己有点冲动,可既然订了就试试��
�,没想到用完一个周期的时候我的斑就开始淡化了,等我用�
��两个周期,黄褐斑已经淡化到接近肤色了,不自己看根本就
看不出来,而且让我更想不到的是,我的皮肤也变白了,这��
�我真的是太激动了,我终于不做“芝麻小妹”了。
阅读了色斑怎么除掉,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎么除掉,同时为您分享祛斑小方法
除面部色素斑:鲜西红柿汁\蜂蜜,按5:1混合,涂面部,过10分钟后�
��净,连用10-15日,能使黑色素分解,皮肤变白红润。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:01 | 1.0 | 纠结色斑怎么除掉 - ```
《摘要》
风也知盘旋,碟恋花相伴翩翩,星也懂缠绵,冰为水情愿己��
�。只有你,狠心走远。我纠结于,连这狂风暴雨般的疼爱都�
��归于平静,还有什么不能归于平静。繁华谢幕,生命崇尚,
盛夏之花。我承认,花很美。但那终究与我无关。流年低吟��
�叹,记忆久酿成伤。生命雨后屋檐下细嗅蔷薇,追忆旅途中�
��冬季畅饮雪扬,走到思绪的尽头,看这一世万丈尘寰。寂寞
在青春泛滥,烟花只得瞬间璀璨,流星唯有煞时迷眼。但是��
�到自己脸上陆续出现的小斑点,我又怎么能安静那!色斑怎�
��除掉,
《客户案例》
说说我的坎坷祛斑路吧,我从小就有黄褐斑,是遗传我��
�的,别人家的小孩的脸都是又白又嫩,红扑扑的,我倒好,�
��脸的小黄褐斑,害得我被喊了十多年的“芝麻小妹”,后来
上学,我们班的卫生委员小姑娘总说我不洗脸,到后来身边��
�姐妹都有男朋友,就只有我孤家寡人一个,这一筐筐的血泪�
��一天一夜都说不完,为了尽快解决我的终身大事,也为了我
那爱美的心,本姑娘决定把自己当小白鼠了,一定要把这该��
�的黄褐斑去掉。<br>
为了能知己知彼百战百胜,我特意上网查查怎么去黄褐��
�,可一查我那颗热血沸腾的心变的洼凉洼凉的,原来遗传性�
��黄褐斑只能被淡化不能完全消除,既然这样就退而求其次吧
,能淡化到看不见总可以吧。我又重振旗鼓,先到网上找个��
�多祛斑偏方,什么西红柿敷脸,牛奶蜂蜜一起上,折腾了半�
��,皮肤确实好点了,可斑还是那样,看来这个方法是不行的
,我又对准那些祛斑产品了,在网上一搜,密密麻麻的十多��
�,心想这么多总有一个管用的吧,什么七日祛斑,三天见效�
��广告做的挺不错,就是不敢相信,最后我觉得还是得听听网
友们的意见,就去论坛上发个帖子,过了几天,我就得到了��
�多回复,那么多回复中,我看很多人提到「黛芙薇尔精华液�
��,说用的效果不错,我当时感觉这个有那么好吗?就去他们��
�方网上详细的询问了客服,他们特别热心,在听我说了我的�
��况后,给了我很多祛斑的建议,我特别感动,当时就试着订
购了两个周期,过后感觉自己有点冲动,可既然订了就试试��
�,没想到用完一个周期的时候我的斑就开始淡化了,等我用�
��两个周期,黄褐斑已经淡化到接近肤色了,不自己看根本就
看不出来,而且让我更想不到的是,我的皮肤也变白了,这��
�我真的是太激动了,我终于不做“芝麻小妹”了。
阅读了色斑怎么除掉,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
色斑怎么除掉,同时为您分享祛斑小方法
除面部色素斑:鲜西红柿汁\蜂蜜,按5:1混合,涂面部,过10分钟后�
��净,连用10-15日,能使黑色素分解,皮肤变白红润。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 4:01 | defect | 纠结色斑怎么除掉 《摘要》 风也知盘旋,碟恋花相伴翩翩,星也懂缠绵,冰为水情愿己�� �。只有你,狠心走远。我纠结于,连这狂风暴雨般的疼爱都� ��归于平静,还有什么不能归于平静。繁华谢幕,生命崇尚, 盛夏之花。我承认,花很美。但那终究与我无关。流年低吟�� �叹,记忆久酿成伤。生命雨后屋檐下细嗅蔷薇,追忆旅途中� ��冬季畅饮雪扬,走到思绪的尽头,看这一世万丈尘寰。寂寞 在青春泛滥,烟花只得瞬间璀璨,流星唯有煞时迷眼。但是�� �到自己脸上陆续出现的小斑点,我又怎么能安静那!色斑怎� ��除掉, 《客户案例》 说说我的坎坷祛斑路吧,我从小就有黄褐斑,是遗传我�� �的,别人家的小孩的脸都是又白又嫩,红扑扑的,我倒好,� ��脸的小黄褐斑,害得我被喊了十多年的“芝麻小妹”,后来 上学,我们班的卫生委员小姑娘总说我不洗脸,到后来身边�� �姐妹都有男朋友,就只有我孤家寡人一个,这一筐筐的血泪� ��一天一夜都说不完,为了尽快解决我的终身大事,也为了我 那爱美的心,本姑娘决定把自己当小白鼠了,一定要把这该�� �的黄褐斑去掉。 为了能知己知彼百战百胜,我特意上网查查怎么去黄褐�� �,可一查我那颗热血沸腾的心变的洼凉洼凉的,原来遗传性� ��黄褐斑只能被淡化不能完全消除,既然这样就退而求其次吧 ,能淡化到看不见总可以吧。我又重振旗鼓,先到网上找个�� �多祛斑偏方,什么西红柿敷脸,牛奶蜂蜜一起上,折腾了半� ��,皮肤确实好点了,可斑还是那样,看来这个方法是不行的 ,我又对准那些祛斑产品了,在网上一搜,密密麻麻的十多�� �,心想这么多总有一个管用的吧,什么七日祛斑,三天见效� ��广告做的挺不错,就是不敢相信,最后我觉得还是得听听网 友们的意见,就去论坛上发个帖子,过了几天,我就得到了�� �多回复,那么多回复中,我看很多人提到「黛芙薇尔精华液� ��,说用的效果不错,我当时感觉这个有那么好吗 就去他们�� �方网上详细的询问了客服,他们特别热心,在听我说了我的� ��况后,给了我很多祛斑的建议,我特别感动,当时就试着订 购了两个周期,过后感觉自己有点冲动,可既然订了就试试�� �,没想到用完一个周期的时候我的斑就开始淡化了,等我用� ��两个周期,黄褐斑已经淡化到接近肤色了,不自己看根本就 看不出来,而且让我更想不到的是,我的皮肤也变白了,这�� �我真的是太激动了,我终于不做“芝麻小妹”了。 阅读了色斑怎么除掉,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 色斑怎么除掉,同时为您分享祛斑小方法 除面部色素斑 鲜西红柿汁 蜂蜜 涂面部 � ��净 能使黑色素分解 皮肤变白红润。 original issue reported on code google com by additive gmail com on jul at | 1 |
16,089 | 2,870,874,581 | IssuesEvent | 2015-06-07 16:18:00 | kubilus1/gendev | https://api.github.com/repos/kubilus1/gendev | closed | Build fails while building binutils | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
Just do a fresh svn checkout and then run make.
I'm running ArchLinux with gcc 4.9.1. I'm wondering if this builds with an
older version of GCC. The error:
gcc -DHAVE_CONFIG_H -I. -I../../binutils-2.24/gas -I.
-I../../binutils-2.24/gas -I../bfd -I../../binutils-2.24/gas/config
-I../../binutils-2.24/gas/../include -I../../binutils-2.24/gas/..
-I../../binutils-2.24/gas/../bfd
-DLOCALEDIR="\"/opt/toolchains/gen/sh-elf/share/locale\"" -W -Wall
-Wstrict-prototypes -Wmissing-prototypes -Wshadow -Werror -g -O2 -MT subsegs.o
-MD -MP -MF .deps/subsegs.Tpo -c -o subsegs.o ../../binutils-2.24/gas/subsegs.c
In file included from ../../binutils-2.24/gas/as.h:95:0,
from ../../binutils-2.24/gas/subsegs.c:25:
../../binutils-2.24/gas/subsegs.c: In function 'subseg_change':
../bfd/bfd.h:304:75: error: right-hand operand of comma expression has no
effect [-Werror=unused-value]
#define bfd_set_section_userdata(bfd, ptr, val) (((ptr)->userdata = (val)),TRUE)
^
../../binutils-2.24/gas/subsegs.c:70:7: note: in expansion of macro
'bfd_set_section_userdata'
bfd_set_section_userdata (stdoutput, seg, seginfo);
^
../../binutils-2.24/gas/subsegs.c: In function 'subseg_get':
../bfd/bfd.h:304:75: error: right-hand operand of comma expression has no
effect [-Werror=unused-value]
#define bfd_set_section_userdata(bfd, ptr, val) (((ptr)->userdata = (val)),TRUE)
^
../../binutils-2.24/gas/subsegs.c:172:7: note: in expansion of macro
'bfd_set_section_userdata'
bfd_set_section_userdata (stdoutput, secptr, seginfo);
^
cc1: all warnings being treated as errors
Makefile:891: recipe for target 'subsegs.o' failed
make[6]: *** [subsegs.o] Error 1
make[6]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24/gas'
Makefile:2165: recipe for target 'all-recursive' failed
make[5]: *** [all-recursive] Error 1
make[5]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24/gas'
Makefile:682: recipe for target 'all' failed
make[4]: *** [all] Error 2
make[4]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24/gas'
Makefile:4709: recipe for target 'all-gas' failed
make[3]: *** [all-gas] Error 2
make[3]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24'
Makefile:829: recipe for target 'all' failed
make[2]: *** [all] Error 2
make[2]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24'
makefile-gen:84: recipe for target 'build-sh2-binutils' failed
make[1]: *** [build-sh2-binutils] Error 2
make[1]: Leaving directory '/home/jalon/src/gendev/work'
Makefile:38: recipe for target 'build' failed
make: *** [build] Error 2
```
Original issue reported on code.google.com by `esakafor...@gmail.com` on 2 Nov 2014 at 7:25 | 1.0 | Build fails while building binutils - ```
What steps will reproduce the problem?
Just do a fresh svn checkout and then run make.
I'm running ArchLinux with gcc 4.9.1. I'm wondering if this builds with an
older version of GCC. The error:
gcc -DHAVE_CONFIG_H -I. -I../../binutils-2.24/gas -I.
-I../../binutils-2.24/gas -I../bfd -I../../binutils-2.24/gas/config
-I../../binutils-2.24/gas/../include -I../../binutils-2.24/gas/..
-I../../binutils-2.24/gas/../bfd
-DLOCALEDIR="\"/opt/toolchains/gen/sh-elf/share/locale\"" -W -Wall
-Wstrict-prototypes -Wmissing-prototypes -Wshadow -Werror -g -O2 -MT subsegs.o
-MD -MP -MF .deps/subsegs.Tpo -c -o subsegs.o ../../binutils-2.24/gas/subsegs.c
In file included from ../../binutils-2.24/gas/as.h:95:0,
from ../../binutils-2.24/gas/subsegs.c:25:
../../binutils-2.24/gas/subsegs.c: In function 'subseg_change':
../bfd/bfd.h:304:75: error: right-hand operand of comma expression has no
effect [-Werror=unused-value]
#define bfd_set_section_userdata(bfd, ptr, val) (((ptr)->userdata = (val)),TRUE)
^
../../binutils-2.24/gas/subsegs.c:70:7: note: in expansion of macro
'bfd_set_section_userdata'
bfd_set_section_userdata (stdoutput, seg, seginfo);
^
../../binutils-2.24/gas/subsegs.c: In function 'subseg_get':
../bfd/bfd.h:304:75: error: right-hand operand of comma expression has no
effect [-Werror=unused-value]
#define bfd_set_section_userdata(bfd, ptr, val) (((ptr)->userdata = (val)),TRUE)
^
../../binutils-2.24/gas/subsegs.c:172:7: note: in expansion of macro
'bfd_set_section_userdata'
bfd_set_section_userdata (stdoutput, secptr, seginfo);
^
cc1: all warnings being treated as errors
Makefile:891: recipe for target 'subsegs.o' failed
make[6]: *** [subsegs.o] Error 1
make[6]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24/gas'
Makefile:2165: recipe for target 'all-recursive' failed
make[5]: *** [all-recursive] Error 1
make[5]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24/gas'
Makefile:682: recipe for target 'all' failed
make[4]: *** [all] Error 2
make[4]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24/gas'
Makefile:4709: recipe for target 'all-gas' failed
make[3]: *** [all-gas] Error 2
make[3]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24'
Makefile:829: recipe for target 'all' failed
make[2]: *** [all] Error 2
make[2]: Leaving directory
'/home/jalon/src/gendev/work/build-binutils-sh-elf-2.24'
makefile-gen:84: recipe for target 'build-sh2-binutils' failed
make[1]: *** [build-sh2-binutils] Error 2
make[1]: Leaving directory '/home/jalon/src/gendev/work'
Makefile:38: recipe for target 'build' failed
make: *** [build] Error 2
```
Original issue reported on code.google.com by `esakafor...@gmail.com` on 2 Nov 2014 at 7:25 | defect | build fails while building binutils what steps will reproduce the problem just do a fresh svn checkout and then run make i m running archlinux with gcc i m wondering if this builds with an older version of gcc the error gcc dhave config h i i binutils gas i i binutils gas i bfd i binutils gas config i binutils gas include i binutils gas i binutils gas bfd dlocaledir opt toolchains gen sh elf share locale w wall wstrict prototypes wmissing prototypes wshadow werror g mt subsegs o md mp mf deps subsegs tpo c o subsegs o binutils gas subsegs c in file included from binutils gas as h from binutils gas subsegs c binutils gas subsegs c in function subseg change bfd bfd h error right hand operand of comma expression has no effect define bfd set section userdata bfd ptr val ptr userdata val true binutils gas subsegs c note in expansion of macro bfd set section userdata bfd set section userdata stdoutput seg seginfo binutils gas subsegs c in function subseg get bfd bfd h error right hand operand of comma expression has no effect define bfd set section userdata bfd ptr val ptr userdata val true binutils gas subsegs c note in expansion of macro bfd set section userdata bfd set section userdata stdoutput secptr seginfo all warnings being treated as errors makefile recipe for target subsegs o failed make error make leaving directory home jalon src gendev work build binutils sh elf gas makefile recipe for target all recursive failed make error make leaving directory home jalon src gendev work build binutils sh elf gas makefile recipe for target all failed make error make leaving directory home jalon src gendev work build binutils sh elf gas makefile recipe for target all gas failed make error make leaving directory home jalon src gendev work build binutils sh elf makefile recipe for target all failed make error make leaving directory home jalon src gendev work build binutils sh elf makefile gen recipe for target build binutils failed make error make leaving directory home jalon src gendev work makefile recipe for target build failed make error original issue reported on code google com by esakafor gmail com on nov at | 1 |
66,885 | 20,746,329,769 | IssuesEvent | 2022-03-14 23:39:10 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | Please update META again for kernel 5.16 | Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Fedora
Distribution Version | 35
Kernel Version | 5.16.11
Architecture | x86_64
OpenZFS Version | 2.1.99-837_g7901b6268
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
I've been using OpenZFS on Linux 5.16 with no trouble at all, compiling it from master, and upgrading Fedora normally. Yesterday dnf removed zfs, because it depended on kernel-devel-5.14, which was being removed from my setup.
I found the error to be in the META file, which limits kernel version to 5.15. I upgraded it to 5.16 and now I could install the latest git version, 2.1.99-913_gdad2b19ff without any compilation errors.
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
| 1.0 | Please update META again for kernel 5.16 - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Fedora
Distribution Version | 35
Kernel Version | 5.16.11
Architecture | x86_64
OpenZFS Version | 2.1.99-837_g7901b6268
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
I've been using OpenZFS on Linux 5.16 with no trouble at all, compiling it from master, and upgrading Fedora normally. Yesterday dnf removed zfs, because it depended on kernel-devel-5.14, which was being removed from my setup.
I found the error to be in the META file, which limits kernel version to 5.15. I upgraded it to 5.16 and now I could install the latest git version, 2.1.99-913_gdad2b19ff without any compilation errors.
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
| defect | please update meta again for kernel thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name fedora distribution version kernel version architecture openzfs version command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing i ve been using openzfs on linux with no trouble at all compiling it from master and upgrading fedora normally yesterday dnf removed zfs because it depended on kernel devel which was being removed from my setup i found the error to be in the meta file which limits kernel version to i upgraded it to and now i could install the latest git version without any compilation errors important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with | 1 |
74,492 | 3,440,526,210 | IssuesEvent | 2015-12-14 14:59:49 | ably/ably-js | https://api.github.com/repos/ably/ably-js | closed | High priority spec incompatibilities | bug high priority | Created from https://github.com/ably/ably-js/issues/43#issuecomment-159657024
##### Wrong names for things
- [ ] RSC12,TO3k2 - `host` should be `restHost`
- [ ] RTC1d,TO3k3 - `wsHost` should be `realtimeHost`
- [ ] RTL6c2, RTP16b, TO3g - `queueEvents` option should be `queueMessages`
- [ ] RTP6a/b - `presence#on` should be `presence#subscribe`
- [ ] RTP7a/b - `presence#off` should be `presence#unsubscribe`
##### Missing functionality (in roughly descending order of importance)
- [ ] RSN3a/b/c,RTS3a/c,TB1 - can't provide channelOptions when instantiating a channel, need to call channel.setOptions
- [ ] TG6/7 - no `PaginatedResult#hasNext` or `#isLast` functions
- [ ] RTN16b/c - no `Connection#recoveryKey` attribute
- [ ] TO3l3,TO3l6 - doesn't use `httpOpenTimeout` or `httpMaxRetryDuration` (rethink of timeouts & fallbacks in progress on https://github.com/ably/ably-js/issues/165)
##### Behaviour that doesn't match the spec (in roughly descending order of importance)
- [ ] RTP11b - `presence#get` doesn't raise for DETACHED or FAILED channels, it returns `[]`
- [ ] RSA7c - client doesn't check clientId in clientOptions can't be `*` itself, leaves it to realtime to give an error
##### 0.8.x final changes
See https://docs.google.com/spreadsheets/d/1bG_SEVlCi3blQBveGtsncxFGD4ArCtS-JCq6iLW3lGw/edit#gid=1243782805&vpid=B4
- [ ] Default token params
- [ ] clientId precedence when provided twice
- [ ] Support wildcard Auth#clientId
- [ ] Client library does not need to have REST limits
- [ ] API versioning to ensure future compatibility
- [ ] ProtocolMessage definition fix
- [ ] Implicit attach consistency
- [ ] Remove connect callback, not agreed
- [ ] Presence get spec defaults
- [ ] Clarification of meaning of null Auth#clientId | 1.0 | High priority spec incompatibilities - Created from https://github.com/ably/ably-js/issues/43#issuecomment-159657024
##### Wrong names for things
- [ ] RSC12,TO3k2 - `host` should be `restHost`
- [ ] RTC1d,TO3k3 - `wsHost` should be `realtimeHost`
- [ ] RTL6c2, RTP16b, TO3g - `queueEvents` option should be `queueMessages`
- [ ] RTP6a/b - `presence#on` should be `presence#subscribe`
- [ ] RTP7a/b - `presence#off` should be `presence#unsubscribe`
##### Missing functionality (in roughly descending order of importance)
- [ ] RSN3a/b/c,RTS3a/c,TB1 - can't provide channelOptions when instantiating a channel, need to call channel.setOptions
- [ ] TG6/7 - no `PaginatedResult#hasNext` or `#isLast` functions
- [ ] RTN16b/c - no `Connection#recoveryKey` attribute
- [ ] TO3l3,TO3l6 - doesn't use `httpOpenTimeout` or `httpMaxRetryDuration` (rethink of timeouts & fallbacks in progress on https://github.com/ably/ably-js/issues/165)
##### Behaviour that doesn't match the spec (in roughly descending order of importance)
- [ ] RTP11b - `presence#get` doesn't raise for DETACHED or FAILED channels, it returns `[]`
- [ ] RSA7c - client doesn't check clientId in clientOptions can't be `*` itself, leaves it to realtime to give an error
##### 0.8.x final changes
See https://docs.google.com/spreadsheets/d/1bG_SEVlCi3blQBveGtsncxFGD4ArCtS-JCq6iLW3lGw/edit#gid=1243782805&vpid=B4
- [ ] Default token params
- [ ] clientId precedence when provided twice
- [ ] Support wildcard Auth#clientId
- [ ] Client library does not need to have REST limits
- [ ] API versioning to ensure future compatibility
- [ ] ProtocolMessage definition fix
- [ ] Implicit attach consistency
- [ ] Remove connect callback, not agreed
- [ ] Presence get spec defaults
- [ ] Clarification of meaning of null Auth#clientId | non_defect | high priority spec incompatibilities created from wrong names for things host should be resthost wshost should be realtimehost queueevents option should be queuemessages b presence on should be presence subscribe b presence off should be presence unsubscribe missing functionality in roughly descending order of importance b c c can t provide channeloptions when instantiating a channel need to call channel setoptions no paginatedresult hasnext or islast functions c no connection recoverykey attribute doesn t use httpopentimeout or httpmaxretryduration rethink of timeouts fallbacks in progress on behaviour that doesn t match the spec in roughly descending order of importance presence get doesn t raise for detached or failed channels it returns client doesn t check clientid in clientoptions can t be itself leaves it to realtime to give an error x final changes see default token params clientid precedence when provided twice support wildcard auth clientid client library does not need to have rest limits api versioning to ensure future compatibility protocolmessage definition fix implicit attach consistency remove connect callback not agreed presence get spec defaults clarification of meaning of null auth clientid | 0 |
64,551 | 18,737,540,650 | IssuesEvent | 2021-11-04 09:38:27 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | no spinner in set email dialog after clicking continue | T-Defect X-Needs-Info P2 S-Major S-Tolerable | This is because the dialog gets recreated and so goes back to its original state | 1.0 | no spinner in set email dialog after clicking continue - This is because the dialog gets recreated and so goes back to its original state | defect | no spinner in set email dialog after clicking continue this is because the dialog gets recreated and so goes back to its original state | 1 |
324,832 | 27,824,482,688 | IssuesEvent | 2023-03-19 16:01:19 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix linalg.test_matmul | TensorFlow Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_linalg.py::test_matmul[cpu-ivy.functional.backends.torch-False-False]</summary>
2023-03-17T22:35:37.4727818Z E RuntimeError: Boolean value of Tensor with more than one value is ambiguous
2023-03-17T22:35:37.4731879Z E ivy.utils.exceptions.IvyBackendException: torch: matmul: Boolean value of Tensor with more than one value is ambiguous
2023-03-17T22:35:37.4732265Z E Falsifying example: test_matmul(
2023-03-17T22:35:37.4732584Z E x=(['int64'], array([[2, 2],
2023-03-17T22:35:37.4732822Z E [2, 2]]), False, False),
2023-03-17T22:35:37.4733098Z E y=(['int64'], array([[2, 2],
2023-03-17T22:35:37.4733329Z E [2, 2]]), False, False),
2023-03-17T22:35:37.4733588Z E fn_name='matmul',
2023-03-17T22:35:37.4733892Z E ground_truth_backend='tensorflow',
2023-03-17T22:35:37.4734169Z E test_flags=FunctionTestFlags(
2023-03-17T22:35:37.4734422Z E num_positional_args=2,
2023-03-17T22:35:37.4734650Z E with_out=True,
2023-03-17T22:35:37.4734874Z E instance_method=False,
2023-03-17T22:35:37.4735275Z E test_gradients=False,
2023-03-17T22:35:37.4735511Z E test_compile=False,
2023-03-17T22:35:37.4735739Z E as_variable=[False],
2023-03-17T22:35:37.4735983Z E native_arrays=[False],
2023-03-17T22:35:37.4736204Z E container=[False],
2023-03-17T22:35:37.4736413Z E ),
2023-03-17T22:35:37.4736861Z E backend_fw=<module 'ivy.functional.backends.torch' from '/ivy/ivy/functional/backends/torch/__init__.py'>,
2023-03-17T22:35:37.4737248Z E on_device='cpu',
2023-03-17T22:35:37.4737448Z E )
2023-03-17T22:35:37.4737618Z E
2023-03-17T22:35:37.4738123Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2BAA4wQCgAAHwAC') as a decorator on your test case
</details>
| 1.0 | Fix linalg.test_matmul - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4451752121/jobs/7818756564" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_functional/test_core/test_linalg.py::test_matmul[cpu-ivy.functional.backends.torch-False-False]</summary>
2023-03-17T22:35:37.4727818Z E RuntimeError: Boolean value of Tensor with more than one value is ambiguous
2023-03-17T22:35:37.4731879Z E ivy.utils.exceptions.IvyBackendException: torch: matmul: Boolean value of Tensor with more than one value is ambiguous
2023-03-17T22:35:37.4732265Z E Falsifying example: test_matmul(
2023-03-17T22:35:37.4732584Z E x=(['int64'], array([[2, 2],
2023-03-17T22:35:37.4732822Z E [2, 2]]), False, False),
2023-03-17T22:35:37.4733098Z E y=(['int64'], array([[2, 2],
2023-03-17T22:35:37.4733329Z E [2, 2]]), False, False),
2023-03-17T22:35:37.4733588Z E fn_name='matmul',
2023-03-17T22:35:37.4733892Z E ground_truth_backend='tensorflow',
2023-03-17T22:35:37.4734169Z E test_flags=FunctionTestFlags(
2023-03-17T22:35:37.4734422Z E num_positional_args=2,
2023-03-17T22:35:37.4734650Z E with_out=True,
2023-03-17T22:35:37.4734874Z E instance_method=False,
2023-03-17T22:35:37.4735275Z E test_gradients=False,
2023-03-17T22:35:37.4735511Z E test_compile=False,
2023-03-17T22:35:37.4735739Z E as_variable=[False],
2023-03-17T22:35:37.4735983Z E native_arrays=[False],
2023-03-17T22:35:37.4736204Z E container=[False],
2023-03-17T22:35:37.4736413Z E ),
2023-03-17T22:35:37.4736861Z E backend_fw=<module 'ivy.functional.backends.torch' from '/ivy/ivy/functional/backends/torch/__init__.py'>,
2023-03-17T22:35:37.4737248Z E on_device='cpu',
2023-03-17T22:35:37.4737448Z E )
2023-03-17T22:35:37.4737618Z E
2023-03-17T22:35:37.4738123Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.0', b'AXicY2BAA4wQCgAAHwAC') as a decorator on your test case
</details>
| non_defect | fix linalg test matmul tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test functional test core test linalg py test matmul e runtimeerror boolean value of tensor with more than one value is ambiguous e ivy utils exceptions ivybackendexception torch matmul boolean value of tensor with more than one value is ambiguous e falsifying example test matmul e x array e false false e y array e false false e fn name matmul e ground truth backend tensorflow e test flags functiontestflags e num positional args e with out true e instance method false e test gradients false e test compile false e as variable e native arrays e container e e backend fw e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case | 0 |
137,386 | 20,120,678,539 | IssuesEvent | 2022-02-08 01:44:17 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | Tech Debt: Remove and convert number-to-words in react-components | vsp-design-system-team | Remove and convert `number-to-words` dependency in `react-components`
File it is being used in `src/components/ProcessList/ProcessList.jsx`
The package it being used to convert integer to word ie: 1 to one
Strategy on how to do that with native JS: https://ourcodeworld.com/articles/read/353/how-to-convert-numbers-to-words-number-spelling-in-javascript
## Parent Issue
https://github.com/department-of-veterans-affairs/va.gov-team/issues/33188
## Acceptance Criteria
[] `number-to-words` dependency is removed
[] File where method is being used has been converted to native JS
[] Tested and confirmed working in all supported browsers: https://depo-platform-documentation.scrollhelp.site/developer-docs/Browser-Support-Policy.1847951366.html | 1.0 | Tech Debt: Remove and convert number-to-words in react-components - Remove and convert `number-to-words` dependency in `react-components`
File it is being used in `src/components/ProcessList/ProcessList.jsx`
The package it being used to convert integer to word ie: 1 to one
Strategy on how to do that with native JS: https://ourcodeworld.com/articles/read/353/how-to-convert-numbers-to-words-number-spelling-in-javascript
## Parent Issue
https://github.com/department-of-veterans-affairs/va.gov-team/issues/33188
## Acceptance Criteria
[] `number-to-words` dependency is removed
[] File where method is being used has been converted to native JS
[] Tested and confirmed working in all supported browsers: https://depo-platform-documentation.scrollhelp.site/developer-docs/Browser-Support-Policy.1847951366.html | non_defect | tech debt remove and convert number to words in react components remove and convert number to words dependency in react components file it is being used in src components processlist processlist jsx the package it being used to convert integer to word ie to one strategy on how to do that with native js parent issue acceptance criteria number to words dependency is removed file where method is being used has been converted to native js tested and confirmed working in all supported browsers | 0 |
23,415 | 3,813,887,057 | IssuesEvent | 2016-03-28 09:14:38 | night-ghost/minimosd-extra | https://api.github.com/repos/night-ghost/minimosd-extra | closed | DISARMED - ARMED warning in airplane firmware. | auto-migrated Priority-Medium Type-Defect | ```
Hi! I'd like to have the DISARMED/ARMED warning, for fixed wing models. I used
to have it on original MinimOSD firmware.
It is a nice feature when using apm:plane in an ardupilot board with no ground
station, because I have no audio or led feedback for the arming status of the
plane.
Thanks a lot for your great job!
```
Original issue reported on code.google.com by `fal...@gmail.com` on 12 Dec 2014 at 4:16 | 1.0 | DISARMED - ARMED warning in airplane firmware. - ```
Hi! I'd like to have the DISARMED/ARMED warning, for fixed wing models. I used
to have it on original MinimOSD firmware.
It is a nice feature when using apm:plane in an ardupilot board with no ground
station, because I have no audio or led feedback for the arming status of the
plane.
Thanks a lot for your great job!
```
Original issue reported on code.google.com by `fal...@gmail.com` on 12 Dec 2014 at 4:16 | defect | disarmed armed warning in airplane firmware hi i d like to have the disarmed armed warning for fixed wing models i used to have it on original minimosd firmware it is a nice feature when using apm plane in an ardupilot board with no ground station because i have no audio or led feedback for the arming status of the plane thanks a lot for your great job original issue reported on code google com by fal gmail com on dec at | 1 |
189,438 | 14,503,456,690 | IssuesEvent | 2020-12-11 22:45:00 | mapbox/rasterio | https://api.github.com/repos/mapbox/rasterio | closed | Test failures with gdal-3.2.0 + proj-7.2.0 | bug testing | While working on packaging gdal-3.2.0 + proj-7.2.0 for Fedora, rebuilding python-rasterio against these results in the following test failures, which don't occur with gdal-3.1.4 + proj-6.3.2:
=================================== FAILURES ===================================
_________________________________ test_shapes __________________________________
basic_image = array([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 1, 1, 1, 0, 0, 0, 0, 0],
...0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], dtype=uint8)
def test_shapes(basic_image):
"""Test creation of shapes from pixel values."""
results = list(shapes(basic_image))
assert len(results) == 2
shape, value = results[0]
assert shape == {
'coordinates': [
[(2, 2), (2, 5), (5, 5), (5, 2), (2, 2)]
],
'type': 'Polygon'
}
assert value == 1
shape, value = results[1]
> assert shape == {
'coordinates': [
[(0, 0), (0, 10), (10, 10), (10, 0), (0, 0)],
[(2, 2), (5, 2), (5, 5), (2, 5), (2, 2)]
],
'type': 'Polygon'
}
E AssertionError: assert {'coordinates...e': 'Polygon'} == {'coordinates...e': 'Polygon'}
E Omitting 1 identical items, use -vv to show
E Differing items:
E {'coordinates': [[(0.0, 0.0), (0.0, 10.0), (10.0, 10.0), (10.0, 0.0), (0.0, 0.0)], [(2.0, 2.0), (2.0, 5.0), (5.0, 5.0), (5.0, 2.0), (2.0, 2.0)]]} != {'coordinates': [[(0, 0), (0, 10), (10, 10), (10, 0), (0, 0)], [(2, 2), (5, 2), (5, 5), (2, 5), (2, 2)]]}
E Full diff:
E {
E - 'coordinates': [[(0,
E + 'coordinates': [[(0.0,...
E
E ...Full output truncated (61 lines hidden), use '-vv' to show
tests/test_features.py:824: AssertionError
------------------------------ Captured log call -------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59788bb0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59788bb0>.
______________________________ test_shapes_indent ______________________________
runner = <click.testing.CliRunner object at 0x7f7b59186f10>
pixelated_image_file = '/tmp/pytest-of-mockbuild/pytest-0/test_shapes_indent0/pixelated_image.tif'
def test_shapes_indent(runner, pixelated_image_file):
"""
--indent option should produce lots of newlines and contiguous spaces
"""
with pytest.warns(None):
result = runner.invoke(
main_group, ['shapes', '--collection', pixelated_image_file, '--indent', 2])
assert result.exit_code == 0
assert result.output.count('"FeatureCollection"') == 1
assert result.output.count('"Feature"') == 4
> assert result.output.count('\n') == 231
E assert 227 == 231
E +227
E -231
tests/test_rio_shapes.py:100: AssertionError
------------------------------ Captured log setup ------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302cd0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302cd0>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed49d0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed49d0>.
------------------------------ Captured log call -------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6460>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6460>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6310>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6310>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed43d0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed43d0>.
_______________________________ test_shapes_mask _______________________________
runner = <click.testing.CliRunner object at 0x7f7b58964130>
pixelated_image = array([[255, 255, 255, 255, 255, 255, 255, 255, 255, 255],
[255, 255, 255, 255, 255, 255, 255, 255, 255, 255],
... 255, 255, 0, 0, 0, 0, 0, 255, 255],
[255, 255, 255, 0, 0, 0, 0, 0, 255, 255]], dtype=uint8)
pixelated_image_file = '/tmp/pytest-of-mockbuild/pytest-0/test_shapes_mask0/pixelated_image.tif'
def test_shapes_mask(runner, pixelated_image, pixelated_image_file):
""" --mask should extract the nodata area of the image """
pixelated_image[0:5, 0:10] = 255
pixelated_image[0:10, 0:3] = 255
pixelated_image[8:10, 8:10] = 255
with rasterio.open(pixelated_image_file, 'r+') as out:
out.write(pixelated_image, indexes=1)
with pytest.warns(None):
result = runner.invoke(
main_group, ['shapes', '--collection', pixelated_image_file, '--mask'])
assert result.exit_code == 0
assert result.output.count('"FeatureCollection"') == 1
assert result.output.count('"Feature"') == 1
> assert np.allclose(
json.loads(result.output)['features'][0]['geometry']['coordinates'],
[[[3, 5], [3, 10], [8, 10], [8, 8], [9, 8], [10, 8], [10, 5], [3, 5]]])
tests/test_rio_shapes.py:155:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
<__array_function__ internals>:5: in allclose
???
/usr/lib64/python3.9/site-packages/numpy/core/numeric.py:2189: in allclose
res = all(isclose(a, b, rtol=rtol, atol=atol, equal_nan=equal_nan))
<__array_function__ internals>:5: in isclose
???
/usr/lib64/python3.9/site-packages/numpy/core/numeric.py:2290: in isclose
return within_tol(x, y, atol, rtol)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
x = array([[[ 3., 5.],
[ 3., 10.],
[ 8., 10.],
[ 8., 8.],
[10., 8.],
[10., 5.],
[ 3., 5.]]])
y = array([[[ 3., 5.],
[ 3., 10.],
[ 8., 10.],
[ 8., 8.],
[ 9., 8.],
[10., 8.],
[10., 5.],
[ 3., 5.]]])
atol = 1e-08, rtol = 1e-05
def within_tol(x, y, atol, rtol):
with errstate(invalid='ignore'):
> return less_equal(abs(x-y), atol + rtol * abs(y))
E ValueError: operands could not be broadcast together with shapes (1,7,2) (1,8,2)
/usr/lib64/python3.9/site-packages/numpy/core/numeric.py:2276: ValueError
------------------------------ Captured log setup ------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302850>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302850>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302d30>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302d30>.
------------------------------ Captured log call -------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302af0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302af0>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302940>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302940>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58e0b0d0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58e0b0d0>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6070>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6070>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58955460>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58955460>.
=============================== warnings summary ===============================
tests/test__crs.py: 3 warnings
tests/test_crs.py: 7 warnings
tests/test_rio_warp.py: 1 warning
tests/test_warp.py: 1 warning
/usr/lib/python3.9/site-packages/_pytest/skipping.py:129: DeprecationWarning: NotImplemented should not be used in a boolean context
result = bool(condition)
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
/builddir/build/BUILD/rasterio-1.1.8/tests/test_read.py:178: DeprecationWarning: tostring() is deprecated. Use tobytes() instead.
self.assertEqual([md5(x.tostring()).hexdigest() for x in a],
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
/builddir/build/BUILD/rasterio-1.1.8/tests/test_read.py:187: DeprecationWarning: tostring() is deprecated. Use tobytes() instead.
self.assertEqual([md5(x.tostring()).hexdigest() for x in a[:]],
tests/test_rio_blocks.py::test_windows
tests/test_rio_blocks.py::test_windows_sequence
tests/test_rio_blocks.py::test_windows_precision
tests/test_rio_blocks.py::test_windows_indent
tests/test_rio_blocks.py::test_windows_compact
tests/test_rio_blocks.py::test_windows_exception
tests/test_rio_blocks.py::test_windows_projected
/usr/lib/python3.9/site-packages/cligj/__init__.py:111: FutureWarning: Sequences of Features, not FeatureCollections, will be the default in version 1.0.0
callback=lambda ctx, param, value: warn(
-- Docs: https://docs.pytest.org/en/stable/warnings.html
=========================== short test summary info ============================
FAILED tests/test_features.py::test_shapes - AssertionError: assert {'coordin...
FAILED tests/test_rio_shapes.py::test_shapes_indent - assert 227 == 231
FAILED tests/test_rio_shapes.py::test_shapes_mask - ValueError: operands coul...
= 3 failed, 1764 passed, 26 skipped, 21 deselected, 11 xfailed, 3 xpassed, 25 warnings in 55.44s =
error: Bad exit status from /var/tmp/rpm-tmp.snIoQt (%check)
Bad exit status from /var/tmp/rpm-tmp.snIoQt (%check) | 1.0 | Test failures with gdal-3.2.0 + proj-7.2.0 - While working on packaging gdal-3.2.0 + proj-7.2.0 for Fedora, rebuilding python-rasterio against these results in the following test failures, which don't occur with gdal-3.1.4 + proj-6.3.2:
=================================== FAILURES ===================================
_________________________________ test_shapes __________________________________
basic_image = array([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 1, 1, 1, 0, 0, 0, 0, 0],
...0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], dtype=uint8)
def test_shapes(basic_image):
"""Test creation of shapes from pixel values."""
results = list(shapes(basic_image))
assert len(results) == 2
shape, value = results[0]
assert shape == {
'coordinates': [
[(2, 2), (2, 5), (5, 5), (5, 2), (2, 2)]
],
'type': 'Polygon'
}
assert value == 1
shape, value = results[1]
> assert shape == {
'coordinates': [
[(0, 0), (0, 10), (10, 10), (10, 0), (0, 0)],
[(2, 2), (5, 2), (5, 5), (2, 5), (2, 2)]
],
'type': 'Polygon'
}
E AssertionError: assert {'coordinates...e': 'Polygon'} == {'coordinates...e': 'Polygon'}
E Omitting 1 identical items, use -vv to show
E Differing items:
E {'coordinates': [[(0.0, 0.0), (0.0, 10.0), (10.0, 10.0), (10.0, 0.0), (0.0, 0.0)], [(2.0, 2.0), (2.0, 5.0), (5.0, 5.0), (5.0, 2.0), (2.0, 2.0)]]} != {'coordinates': [[(0, 0), (0, 10), (10, 10), (10, 0), (0, 0)], [(2, 2), (5, 2), (5, 5), (2, 5), (2, 2)]]}
E Full diff:
E {
E - 'coordinates': [[(0,
E + 'coordinates': [[(0.0,...
E
E ...Full output truncated (61 lines hidden), use '-vv' to show
tests/test_features.py:824: AssertionError
------------------------------ Captured log call -------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59788bb0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59788bb0>.
______________________________ test_shapes_indent ______________________________
runner = <click.testing.CliRunner object at 0x7f7b59186f10>
pixelated_image_file = '/tmp/pytest-of-mockbuild/pytest-0/test_shapes_indent0/pixelated_image.tif'
def test_shapes_indent(runner, pixelated_image_file):
"""
--indent option should produce lots of newlines and contiguous spaces
"""
with pytest.warns(None):
result = runner.invoke(
main_group, ['shapes', '--collection', pixelated_image_file, '--indent', 2])
assert result.exit_code == 0
assert result.output.count('"FeatureCollection"') == 1
assert result.output.count('"Feature"') == 4
> assert result.output.count('\n') == 231
E assert 227 == 231
E +227
E -231
tests/test_rio_shapes.py:100: AssertionError
------------------------------ Captured log setup ------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302cd0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302cd0>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed49d0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed49d0>.
------------------------------ Captured log call -------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6460>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6460>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6310>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6310>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed43d0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58ed43d0>.
_______________________________ test_shapes_mask _______________________________
runner = <click.testing.CliRunner object at 0x7f7b58964130>
pixelated_image = array([[255, 255, 255, 255, 255, 255, 255, 255, 255, 255],
[255, 255, 255, 255, 255, 255, 255, 255, 255, 255],
... 255, 255, 0, 0, 0, 0, 0, 255, 255],
[255, 255, 255, 0, 0, 0, 0, 0, 255, 255]], dtype=uint8)
pixelated_image_file = '/tmp/pytest-of-mockbuild/pytest-0/test_shapes_mask0/pixelated_image.tif'
def test_shapes_mask(runner, pixelated_image, pixelated_image_file):
""" --mask should extract the nodata area of the image """
pixelated_image[0:5, 0:10] = 255
pixelated_image[0:10, 0:3] = 255
pixelated_image[8:10, 8:10] = 255
with rasterio.open(pixelated_image_file, 'r+') as out:
out.write(pixelated_image, indexes=1)
with pytest.warns(None):
result = runner.invoke(
main_group, ['shapes', '--collection', pixelated_image_file, '--mask'])
assert result.exit_code == 0
assert result.output.count('"FeatureCollection"') == 1
assert result.output.count('"Feature"') == 1
> assert np.allclose(
json.loads(result.output)['features'][0]['geometry']['coordinates'],
[[[3, 5], [3, 10], [8, 10], [8, 8], [9, 8], [10, 8], [10, 5], [3, 5]]])
tests/test_rio_shapes.py:155:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
<__array_function__ internals>:5: in allclose
???
/usr/lib64/python3.9/site-packages/numpy/core/numeric.py:2189: in allclose
res = all(isclose(a, b, rtol=rtol, atol=atol, equal_nan=equal_nan))
<__array_function__ internals>:5: in isclose
???
/usr/lib64/python3.9/site-packages/numpy/core/numeric.py:2290: in isclose
return within_tol(x, y, atol, rtol)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
x = array([[[ 3., 5.],
[ 3., 10.],
[ 8., 10.],
[ 8., 8.],
[10., 8.],
[10., 5.],
[ 3., 5.]]])
y = array([[[ 3., 5.],
[ 3., 10.],
[ 8., 10.],
[ 8., 8.],
[ 9., 8.],
[10., 8.],
[10., 5.],
[ 3., 5.]]])
atol = 1e-08, rtol = 1e-05
def within_tol(x, y, atol, rtol):
with errstate(invalid='ignore'):
> return less_equal(abs(x-y), atol + rtol * abs(y))
E ValueError: operands could not be broadcast together with shapes (1,7,2) (1,8,2)
/usr/lib64/python3.9/site-packages/numpy/core/numeric.py:2276: ValueError
------------------------------ Captured log setup ------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302850>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302850>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302d30>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302d30>.
------------------------------ Captured log call -------------------------------
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302af0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302af0>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302940>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b59302940>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58e0b0d0>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58e0b0d0>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6070>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b590c6070>.
DEBUG rasterio._env:env.py:310 GDAL data files are available at built-in paths
DEBUG rasterio._env:env.py:310 PROJ data files are available at built-in paths
DEBUG rasterio._env:env.py:310 Started GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58955460>.
DEBUG rasterio._env:env.py:346 Stopped GDALEnv <rasterio._env.GDALEnv object at 0x7f7b58955460>.
=============================== warnings summary ===============================
tests/test__crs.py: 3 warnings
tests/test_crs.py: 7 warnings
tests/test_rio_warp.py: 1 warning
tests/test_warp.py: 1 warning
/usr/lib/python3.9/site-packages/_pytest/skipping.py:129: DeprecationWarning: NotImplemented should not be used in a boolean context
result = bool(condition)
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
/builddir/build/BUILD/rasterio-1.1.8/tests/test_read.py:178: DeprecationWarning: tostring() is deprecated. Use tobytes() instead.
self.assertEqual([md5(x.tostring()).hexdigest() for x in a],
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
tests/test_read.py::ReaderContextTest::test_read_window
/builddir/build/BUILD/rasterio-1.1.8/tests/test_read.py:187: DeprecationWarning: tostring() is deprecated. Use tobytes() instead.
self.assertEqual([md5(x.tostring()).hexdigest() for x in a[:]],
tests/test_rio_blocks.py::test_windows
tests/test_rio_blocks.py::test_windows_sequence
tests/test_rio_blocks.py::test_windows_precision
tests/test_rio_blocks.py::test_windows_indent
tests/test_rio_blocks.py::test_windows_compact
tests/test_rio_blocks.py::test_windows_exception
tests/test_rio_blocks.py::test_windows_projected
/usr/lib/python3.9/site-packages/cligj/__init__.py:111: FutureWarning: Sequences of Features, not FeatureCollections, will be the default in version 1.0.0
callback=lambda ctx, param, value: warn(
-- Docs: https://docs.pytest.org/en/stable/warnings.html
=========================== short test summary info ============================
FAILED tests/test_features.py::test_shapes - AssertionError: assert {'coordin...
FAILED tests/test_rio_shapes.py::test_shapes_indent - assert 227 == 231
FAILED tests/test_rio_shapes.py::test_shapes_mask - ValueError: operands coul...
= 3 failed, 1764 passed, 26 skipped, 21 deselected, 11 xfailed, 3 xpassed, 25 warnings in 55.44s =
error: Bad exit status from /var/tmp/rpm-tmp.snIoQt (%check)
Bad exit status from /var/tmp/rpm-tmp.snIoQt (%check) | non_defect | test failures with gdal proj while working on packaging gdal proj for fedora rebuilding python rasterio against these results in the following test failures which don t occur with gdal proj failures test shapes basic image array dtype def test shapes basic image test creation of shapes from pixel values results list shapes basic image assert len results shape value results assert shape coordinates type polygon assert value shape value results assert shape coordinates type polygon e assertionerror assert coordinates e polygon coordinates e polygon e omitting identical items use vv to show e differing items e coordinates coordinates e full diff e e coordinates e coordinates e e full output truncated lines hidden use vv to show tests test features py assertionerror captured log call debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv test shapes indent runner pixelated image file tmp pytest of mockbuild pytest test shapes pixelated image tif def test shapes indent runner pixelated image file indent option should produce lots of newlines and contiguous spaces with pytest warns none result runner invoke main group assert result exit code assert result output count featurecollection assert result output count feature assert result output count n e assert e e tests test rio shapes py assertionerror captured log setup debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv captured log call debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv test shapes mask runner pixelated image array dtype pixelated image file tmp pytest of mockbuild pytest test shapes pixelated image tif def test shapes mask runner pixelated image pixelated image file mask should extract the nodata area of the image pixelated image pixelated image pixelated image with rasterio open pixelated image file r as out out write pixelated image indexes with pytest warns none result runner invoke main group assert result exit code assert result output count featurecollection assert result output count feature assert np allclose json loads result output tests test rio shapes py in allclose usr site packages numpy core numeric py in allclose res all isclose a b rtol rtol atol atol equal nan equal nan in isclose usr site packages numpy core numeric py in isclose return within tol x y atol rtol x array y array atol rtol def within tol x y atol rtol with errstate invalid ignore return less equal abs x y atol rtol abs y e valueerror operands could not be broadcast together with shapes usr site packages numpy core numeric py valueerror captured log setup debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv captured log call debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv debug rasterio env env py gdal data files are available at built in paths debug rasterio env env py proj data files are available at built in paths debug rasterio env env py started gdalenv debug rasterio env env py stopped gdalenv warnings summary tests test crs py warnings tests test crs py warnings tests test rio warp py warning tests test warp py warning usr lib site packages pytest skipping py deprecationwarning notimplemented should not be used in a boolean context result bool condition tests test read py readercontexttest test read window tests test read py readercontexttest test read window tests test read py readercontexttest test read window builddir build build rasterio tests test read py deprecationwarning tostring is deprecated use tobytes instead self assertequal tests test read py readercontexttest test read window tests test read py readercontexttest test read window tests test read py readercontexttest test read window builddir build build rasterio tests test read py deprecationwarning tostring is deprecated use tobytes instead self assertequal tests test rio blocks py test windows tests test rio blocks py test windows sequence tests test rio blocks py test windows precision tests test rio blocks py test windows indent tests test rio blocks py test windows compact tests test rio blocks py test windows exception tests test rio blocks py test windows projected usr lib site packages cligj init py futurewarning sequences of features not featurecollections will be the default in version callback lambda ctx param value warn docs short test summary info failed tests test features py test shapes assertionerror assert coordin failed tests test rio shapes py test shapes indent assert failed tests test rio shapes py test shapes mask valueerror operands coul failed passed skipped deselected xfailed xpassed warnings in error bad exit status from var tmp rpm tmp snioqt check bad exit status from var tmp rpm tmp snioqt check | 0 |
67,113 | 20,908,508,947 | IssuesEvent | 2022-03-24 06:37:18 | martinrotter/rssguard | https://api.github.com/repos/martinrotter/rssguard | closed | [BUG]: Can't Import FEEDS | Type-Defect | ### Brief description of the issue
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### How to reproduce the bug?
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### What was the expected result?
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### What actually happened?
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### Other information
_No response_
### Operating system and version
* OS: Linux (Ubuntu/Pop_OS)
* RSS Guard version: 4.1.2
| 1.0 | [BUG]: Can't Import FEEDS - ### Brief description of the issue
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### How to reproduce the bug?
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### What was the expected result?
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### What actually happened?
Switching laptops. Exported FEEDS from RSSGuard into a file. Copied the OPML file to new laptop. Started RSS Guard and attempted to Import FEEDS. File is recognised, but no feeds are listed in the preview and therefore cannot import feeds.
### Other information
_No response_
### Operating system and version
* OS: Linux (Ubuntu/Pop_OS)
* RSS Guard version: 4.1.2
| defect | can t import feeds brief description of the issue switching laptops exported feeds from rssguard into a file copied the opml file to new laptop started rss guard and attempted to import feeds file is recognised but no feeds are listed in the preview and therefore cannot import feeds how to reproduce the bug switching laptops exported feeds from rssguard into a file copied the opml file to new laptop started rss guard and attempted to import feeds file is recognised but no feeds are listed in the preview and therefore cannot import feeds what was the expected result switching laptops exported feeds from rssguard into a file copied the opml file to new laptop started rss guard and attempted to import feeds file is recognised but no feeds are listed in the preview and therefore cannot import feeds what actually happened switching laptops exported feeds from rssguard into a file copied the opml file to new laptop started rss guard and attempted to import feeds file is recognised but no feeds are listed in the preview and therefore cannot import feeds other information no response operating system and version os linux ubuntu pop os rss guard version | 1 |
13,881 | 2,789,448,568 | IssuesEvent | 2015-05-08 19:27:55 | orwant/google-visualization-issues | https://api.github.com/repos/orwant/google-visualization-issues | opened | Motion Chart | Bug: The size of the bubbles is inconsistent when "Same Size" is selected | Priority-Medium Type-Defect | Original [issue 52](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=52) created by orwant on 2009-09-13T23:04:12.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Initially "Same Size" is default for the size of the bubbles. Note the size.
2. Switch to another column for sizing. Size changes.
3. Switch back to "Same Size"; the size of the bubbles are now smaller than
in step 1
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
Motion Chart
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
NO
<b>What operating system and browser are you using?</b>
OS X 10.5 (Leopard) and Firefox 3.5.2
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| 1.0 | Motion Chart | Bug: The size of the bubbles is inconsistent when "Same Size" is selected - Original [issue 52](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=52) created by orwant on 2009-09-13T23:04:12.000Z:
<b>What steps will reproduce the problem? Please provide a link to a</b>
<b>demonstration page if at all possible, or attach code.</b>
1. Initially "Same Size" is default for the size of the bubbles. Note the size.
2. Switch to another column for sizing. Size changes.
3. Switch back to "Same Size"; the size of the bubbles are now smaller than
in step 1
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
Motion Chart
<b>Are you using the test environment (version 1.1)?</b>
<b>(If you are not sure, answer NO)</b>
NO
<b>What operating system and browser are you using?</b>
OS X 10.5 (Leopard) and Firefox 3.5.2
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| defect | motion chart bug the size of the bubbles is inconsistent when same size is selected original created by orwant on what steps will reproduce the problem please provide a link to a demonstration page if at all possible or attach code initially quot same size quot is default for the size of the bubbles note the size switch to another column for sizing size changes switch back to quot same size quot the size of the bubbles are now smaller than in step what component is this issue related to piechart linechart datatable query etc motion chart are you using the test environment version if you are not sure answer no no what operating system and browser are you using os x leopard and firefox for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved | 1 |
78,901 | 27,811,804,183 | IssuesEvent | 2023-03-18 07:40:03 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | closed | NULL pointer dereference when rolling back a snapshot while active chmod/chown commands were running | Type: Defect Status: Stale | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Archlinux
Distribution Version | 2021-11-29
Kernel Version | 5.10.63-custom
Architecture | 64-bit
OpenZFS Version | zfs-2.1.99-448_gb9ec4a15e zfs-kmod-2.1.99-448_gb9ec4a15e
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
This crash spurred from zpl_setattr during a `chown` and `chmod -R` operation. I did a rollback while those operations were running in the background and it crashed as a result.
### Describe how to reproduce the problem
```root@server archive # find /tank/archive -type f -exec chmod 0664 {} \; &
root@server archive # find /tank/archive -type d -exec chmod 0775 {} \; &
root@server archive # chown root:administrators -R /tank/archive/
^Z
root@server archive # zfs rollback tank/encrypted/archive@backup-20211129-prechmod
find: 'chmod' terminated by signal 9
root@server ~ # jobs
[1] Running find /tank/archive -type f -exec chmod 0664 {} \; &
[2]- Running find /tank/archive -type d -exec chmod 0775 {} \; &
[3]+ Running chown root:administrators -R /tank/archive/ &
root@server archive # zpool iostat -v 1
^Z
<hung here>
```
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
```[6980561.605730] BUG: kernel NULL pointer dereference, address: 00000000000000a1
[6980561.664607] #PF: supervisor read access in kernel mode
[6980561.712320] #PF: error_code(0x0000) - not-present page
[6980561.759826] PGD 0 P4D 0
[6980561.791465] Oops: 0000 [#1] SMP NOPTI
[6980561.829638] CPU: 11 PID: 3226448 Comm: chmod Tainted: P W O 5.10.63-custom+ #263
[6980561.896445] Hardware name: Supermicro Super Server/H11SSL-i, BIOS 1.3 06/25/2019
[6980561.956982] RIP: 0010:dmu_tx_create_dd+0x3e/0xe0
[6980562.000508] Code: bf b0 00 00 00 48 83 ec 18 65 48 8b 04 25 28 00 00 00 48 89 44 24 10 31 c0 e8 be 75 f6 ff 48 89 58 28 49 89 c4 48 85 db 74 0c <48> 8b 83 a8 00 00 00 49 89 44 24 30 49 8d 44 24 10
48 89 e7 49 c7
[6980562.143762] RSP: 0018:ffffb8aea112fb98 EFLAGS: 00010286
[6980562.190868] RAX: ffff9552e734b000 RBX: fffffffffffffff9 RCX: 0000000000000001
[6980562.249198] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff954940041180
[6980562.307299] RBP: ffffb8aea112fdb0 R08: ffffb8aea112fb20 R09: ffff9552e734b000
[6980562.365205] R10: 0000000000000041 R11: 001e01bf10000000 R12: ffff9552e734b000
[6980562.422893] R13: ffff95559e44a040 R14: 0000000000000041 R15: ffffb8aea112fcc8
[6980562.480372] FS: 00007f1e1313a600(0000) GS:ffff95583ed80000(0000) knlGS:0000000000000000
[6980562.543462] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[6980562.592277] CR2: 00000000000000a1 CR3: 00000008fc238000 CR4: 00000000003506e0
[6980562.649398] Call Trace:
[6980562.678071] dmu_tx_create+0x13/0x20
[6980562.713254] zfs_setattr+0x6ff/0x21c0
[6980562.748734] ? fallback_alloc+0xe4/0x1e0
[6980562.785540] ? kmem_cache_alloc_node+0xf7/0x1f0
[6980562.825775] zpl_setattr+0xfa/0x170
[6980562.859440] notify_change+0x265/0x480
[6980562.894461] ? chmod_common+0x81/0x130
[6980562.929189] chmod_common+0x81/0x130
[6980562.962660] do_fchmodat+0x55/0xa0
[6980562.994818] __x64_sys_fchmodat+0x10/0x20
[6980563.030374] do_syscall_64+0x33/0x40
[6980563.063108] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[6980563.104615] RIP: 0033:0x7f1e1305c437
[6980563.136908] Code: 28 00 00 00 0f 85 b2 00 00 00 48 81 c4 c0 00 00 00 44 89 e0 5d 41 5c 41 5d c3 66 0f 1f 84 00 00 00 00 00 b8 0c 01 00 00 0f 05 <48> 3d 00 f0 ff ff 77 51 41 89 c4 eb bc 0f 1f 40 00
44 89 e7 41 bc
[6980563.270256] RSP: 002b:00007ffc079e67f0 EFLAGS: 00000246 ORIG_RAX: 000000000000010c
[6980563.326313] RAX: ffffffffffffffda RBX: 000055cf258c8320 RCX: 00007f1e1305c437
[6980563.379616] RDX: 00000000000001fd RSI: 000055cf258c83b0 RDI: 00000000ffffff9c
[6980563.432684] RBP: 000055cf258c95d0 R08: 0000000000000000 R09: 00000000000001fd
[6980563.485579] R10: 00000000000003ff R11: 0000000000000246 R12: 000055cf258c83b0
[6980563.538397] R13: 00000000000001fd R14: 0000000000000002 R15: 00000000000041fd
[6980563.590784] Modules linked in: efivarfs nvidia_drm(PO) nvidia_uvm(PO) nvidia_modeset(PO) nvidia(PO)
[6980563.654559] CR2: 00000000000000a1
[6980563.683805] ---[ end trace 29fb02afae20a3d2 ]---
[6980563.871527] RIP: 0010:dmu_tx_create_dd+0x3e/0xe0
[6980563.908180] Code: bf b0 00 00 00 48 83 ec 18 65 48 8b 04 25 28 00 00 00 48 89 44 24 10 31 c0 e8 be 75 f6 ff 48 89 58 28 49 89 c4 48 85 db 74 0c <48> 8b 83 a8 00 00 00 49 89 44 24 30 49 8d 44 24 10
48 89 e7 49 c7
[6980564.037914] RSP: 0018:ffffb8aea112fb98 EFLAGS: 00010286
[6980564.078069] RAX: ffff9552e734b000 RBX: fffffffffffffff9 RCX: 0000000000000001
[6980564.129856] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff954940041180
[6980564.181598] RBP: ffffb8aea112fdb0 R08: ffffb8aea112fb20 R09: ffff9552e734b000
[6980564.233306] R10: 0000000000000041 R11: 001e01bf10000000 R12: ffff9552e734b000
[6980564.285036] R13: ffff95559e44a040 R14: 0000000000000041 R15: ffffb8aea112fcc8
[6980564.336835] FS: 00007f1e1313a600(0000) GS:ffff95583ed80000(0000) knlGS:0000000000000000
[6980564.394245] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[6980564.437549] CR2: 00000000000000a1 CR3: 00000008fc238000 CR4: 00000000003506e0
``` | 1.0 | NULL pointer dereference when rolling back a snapshot while active chmod/chown commands were running - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Archlinux
Distribution Version | 2021-11-29
Kernel Version | 5.10.63-custom
Architecture | 64-bit
OpenZFS Version | zfs-2.1.99-448_gb9ec4a15e zfs-kmod-2.1.99-448_gb9ec4a15e
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
This crash spurred from zpl_setattr during a `chown` and `chmod -R` operation. I did a rollback while those operations were running in the background and it crashed as a result.
### Describe how to reproduce the problem
```root@server archive # find /tank/archive -type f -exec chmod 0664 {} \; &
root@server archive # find /tank/archive -type d -exec chmod 0775 {} \; &
root@server archive # chown root:administrators -R /tank/archive/
^Z
root@server archive # zfs rollback tank/encrypted/archive@backup-20211129-prechmod
find: 'chmod' terminated by signal 9
root@server ~ # jobs
[1] Running find /tank/archive -type f -exec chmod 0664 {} \; &
[2]- Running find /tank/archive -type d -exec chmod 0775 {} \; &
[3]+ Running chown root:administrators -R /tank/archive/ &
root@server archive # zpool iostat -v 1
^Z
<hung here>
```
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
```[6980561.605730] BUG: kernel NULL pointer dereference, address: 00000000000000a1
[6980561.664607] #PF: supervisor read access in kernel mode
[6980561.712320] #PF: error_code(0x0000) - not-present page
[6980561.759826] PGD 0 P4D 0
[6980561.791465] Oops: 0000 [#1] SMP NOPTI
[6980561.829638] CPU: 11 PID: 3226448 Comm: chmod Tainted: P W O 5.10.63-custom+ #263
[6980561.896445] Hardware name: Supermicro Super Server/H11SSL-i, BIOS 1.3 06/25/2019
[6980561.956982] RIP: 0010:dmu_tx_create_dd+0x3e/0xe0
[6980562.000508] Code: bf b0 00 00 00 48 83 ec 18 65 48 8b 04 25 28 00 00 00 48 89 44 24 10 31 c0 e8 be 75 f6 ff 48 89 58 28 49 89 c4 48 85 db 74 0c <48> 8b 83 a8 00 00 00 49 89 44 24 30 49 8d 44 24 10
48 89 e7 49 c7
[6980562.143762] RSP: 0018:ffffb8aea112fb98 EFLAGS: 00010286
[6980562.190868] RAX: ffff9552e734b000 RBX: fffffffffffffff9 RCX: 0000000000000001
[6980562.249198] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff954940041180
[6980562.307299] RBP: ffffb8aea112fdb0 R08: ffffb8aea112fb20 R09: ffff9552e734b000
[6980562.365205] R10: 0000000000000041 R11: 001e01bf10000000 R12: ffff9552e734b000
[6980562.422893] R13: ffff95559e44a040 R14: 0000000000000041 R15: ffffb8aea112fcc8
[6980562.480372] FS: 00007f1e1313a600(0000) GS:ffff95583ed80000(0000) knlGS:0000000000000000
[6980562.543462] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[6980562.592277] CR2: 00000000000000a1 CR3: 00000008fc238000 CR4: 00000000003506e0
[6980562.649398] Call Trace:
[6980562.678071] dmu_tx_create+0x13/0x20
[6980562.713254] zfs_setattr+0x6ff/0x21c0
[6980562.748734] ? fallback_alloc+0xe4/0x1e0
[6980562.785540] ? kmem_cache_alloc_node+0xf7/0x1f0
[6980562.825775] zpl_setattr+0xfa/0x170
[6980562.859440] notify_change+0x265/0x480
[6980562.894461] ? chmod_common+0x81/0x130
[6980562.929189] chmod_common+0x81/0x130
[6980562.962660] do_fchmodat+0x55/0xa0
[6980562.994818] __x64_sys_fchmodat+0x10/0x20
[6980563.030374] do_syscall_64+0x33/0x40
[6980563.063108] entry_SYSCALL_64_after_hwframe+0x44/0xa9
[6980563.104615] RIP: 0033:0x7f1e1305c437
[6980563.136908] Code: 28 00 00 00 0f 85 b2 00 00 00 48 81 c4 c0 00 00 00 44 89 e0 5d 41 5c 41 5d c3 66 0f 1f 84 00 00 00 00 00 b8 0c 01 00 00 0f 05 <48> 3d 00 f0 ff ff 77 51 41 89 c4 eb bc 0f 1f 40 00
44 89 e7 41 bc
[6980563.270256] RSP: 002b:00007ffc079e67f0 EFLAGS: 00000246 ORIG_RAX: 000000000000010c
[6980563.326313] RAX: ffffffffffffffda RBX: 000055cf258c8320 RCX: 00007f1e1305c437
[6980563.379616] RDX: 00000000000001fd RSI: 000055cf258c83b0 RDI: 00000000ffffff9c
[6980563.432684] RBP: 000055cf258c95d0 R08: 0000000000000000 R09: 00000000000001fd
[6980563.485579] R10: 00000000000003ff R11: 0000000000000246 R12: 000055cf258c83b0
[6980563.538397] R13: 00000000000001fd R14: 0000000000000002 R15: 00000000000041fd
[6980563.590784] Modules linked in: efivarfs nvidia_drm(PO) nvidia_uvm(PO) nvidia_modeset(PO) nvidia(PO)
[6980563.654559] CR2: 00000000000000a1
[6980563.683805] ---[ end trace 29fb02afae20a3d2 ]---
[6980563.871527] RIP: 0010:dmu_tx_create_dd+0x3e/0xe0
[6980563.908180] Code: bf b0 00 00 00 48 83 ec 18 65 48 8b 04 25 28 00 00 00 48 89 44 24 10 31 c0 e8 be 75 f6 ff 48 89 58 28 49 89 c4 48 85 db 74 0c <48> 8b 83 a8 00 00 00 49 89 44 24 30 49 8d 44 24 10
48 89 e7 49 c7
[6980564.037914] RSP: 0018:ffffb8aea112fb98 EFLAGS: 00010286
[6980564.078069] RAX: ffff9552e734b000 RBX: fffffffffffffff9 RCX: 0000000000000001
[6980564.129856] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff954940041180
[6980564.181598] RBP: ffffb8aea112fdb0 R08: ffffb8aea112fb20 R09: ffff9552e734b000
[6980564.233306] R10: 0000000000000041 R11: 001e01bf10000000 R12: ffff9552e734b000
[6980564.285036] R13: ffff95559e44a040 R14: 0000000000000041 R15: ffffb8aea112fcc8
[6980564.336835] FS: 00007f1e1313a600(0000) GS:ffff95583ed80000(0000) knlGS:0000000000000000
[6980564.394245] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[6980564.437549] CR2: 00000000000000a1 CR3: 00000008fc238000 CR4: 00000000003506e0
``` | defect | null pointer dereference when rolling back a snapshot while active chmod chown commands were running thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name archlinux distribution version kernel version custom architecture bit openzfs version zfs zfs kmod command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing this crash spurred from zpl setattr during a chown and chmod r operation i did a rollback while those operations were running in the background and it crashed as a result describe how to reproduce the problem root server archive find tank archive type f exec chmod root server archive find tank archive type d exec chmod root server archive chown root administrators r tank archive z root server archive zfs rollback tank encrypted archive backup prechmod find chmod terminated by signal root server jobs running find tank archive type f exec chmod running find tank archive type d exec chmod running chown root administrators r tank archive root server archive zpool iostat v z include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with bug kernel null pointer dereference address pf supervisor read access in kernel mode pf error code not present page pgd oops smp nopti cpu pid comm chmod tainted p w o custom hardware name supermicro super server i bios rip dmu tx create dd code bf ec be ff db rsp eflags rax rbx rcx rdx rsi rdi rbp fs gs knlgs cs ds es call trace dmu tx create zfs setattr fallback alloc kmem cache alloc node zpl setattr notify change chmod common chmod common do fchmodat sys fchmodat do syscall entry syscall after hwframe rip code ff ff eb bc bc rsp eflags orig rax rax ffffffffffffffda rbx rcx rdx rsi rdi rbp modules linked in efivarfs nvidia drm po nvidia uvm po nvidia modeset po nvidia po rip dmu tx create dd code bf ec be ff db rsp eflags rax rbx rcx rdx rsi rdi rbp fs gs knlgs cs ds es | 1 |
326,230 | 27,980,120,911 | IssuesEvent | 2023-03-26 03:03:19 | Fetyali7/MAK-soen341project2023 | https://api.github.com/repos/Fetyali7/MAK-soen341project2023 | closed | User Story #14 Task 2 | User acceptance test | Acceptance test for Login page! #56
- Check if inputs for email and password matche with the database
-
- Alert the user that login was sucessful
- Navigate to Home page
- Navigation bar should show that the user's name on the right
- If the user is an Applicant
- Find Job tab
- Submitted tab
- Inbox tab
- if the user is an Employer
- Job tab
- PostJob tab
- Inbox
| 1.0 | User Story #14 Task 2 - Acceptance test for Login page! #56
- Check if inputs for email and password matche with the database
-
- Alert the user that login was sucessful
- Navigate to Home page
- Navigation bar should show that the user's name on the right
- If the user is an Applicant
- Find Job tab
- Submitted tab
- Inbox tab
- if the user is an Employer
- Job tab
- PostJob tab
- Inbox
| non_defect | user story task acceptance test for login page check if inputs for email and password matche with the database alert the user that login was sucessful navigate to home page navigation bar should show that the user s name on the right if the user is an applicant find job tab submitted tab inbox tab if the user is an employer job tab postjob tab inbox | 0 |
75,273 | 9,834,060,762 | IssuesEvent | 2019-06-17 08:47:43 | PGM-Lab/InferPy | https://api.github.com/repos/PGM-Lab/InferPy | closed | some code fragments are not highlighted | bug documentation | Sphinx does not highlight some of the code fragments, such as
```
# definition of a generic model
@inf.probmodel
def pca(k,d):
w = inf.Normal(loc=np.zeros([k,d]), scale=1, name="w") # shape = [k,d]
with inf.datamodel():
z = inf.Normal(np.ones(k),1, name="z") # shape = [N,k]
x = inf.Normal(z @ w , 1, name="x") # shape = [N,d]
# create an instance of the model
m = pca(k=1,d=2)
```
at https://inferpy.readthedocs.io/en/latest/notes/guidemodels.html | 1.0 | some code fragments are not highlighted - Sphinx does not highlight some of the code fragments, such as
```
# definition of a generic model
@inf.probmodel
def pca(k,d):
w = inf.Normal(loc=np.zeros([k,d]), scale=1, name="w") # shape = [k,d]
with inf.datamodel():
z = inf.Normal(np.ones(k),1, name="z") # shape = [N,k]
x = inf.Normal(z @ w , 1, name="x") # shape = [N,d]
# create an instance of the model
m = pca(k=1,d=2)
```
at https://inferpy.readthedocs.io/en/latest/notes/guidemodels.html | non_defect | some code fragments are not highlighted sphinx does not highlight some of the code fragments such as definition of a generic model inf probmodel def pca k d w inf normal loc np zeros scale name w shape with inf datamodel z inf normal np ones k name z shape x inf normal z w name x shape create an instance of the model m pca k d at | 0 |
78,034 | 22,093,209,074 | IssuesEvent | 2022-06-01 07:55:22 | PaddlePaddle/Paddle | https://api.github.com/repos/PaddlePaddle/Paddle | opened | 3090运行nn.Conv2D(3, 16, 3)报错 | status/new-issue type/build | ### 问题描述 Issue Description
代码:
import paddle.nn as nn
nn.Conv2D(3, 16, 3)
log:
W0601 15:46:55.629635 21509 device_context.cc:447] Please NOTE: device: 0, GPU Compute Capability: 8.6, Driver API Version: 11.2, Runtime API Version: 11.2
W0601 15:46:55.633584 21509 device_context.cc:465] device: 0, cuDNN Version: 7.6.
Traceback (most recent call last):
File "111.py", line 3, in <module>
nn.Conv2D(3, 16, 3)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/nn/layer/conv.py", line 656, in __init__
data_format=data_format)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/nn/layer/conv.py", line 135, in __init__
default_initializer=_get_default_param_initializer())
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py", line 422, in create_parameter
default_initializer)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/layer_helper_base.py", line 378, in create_parameter
**attr._to_kwargs(with_initializer=True))
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/framework.py", line 3137, in create_parameter
initializer(param, self)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/initializer.py", line 362, in __call__
stop_gradient=True)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/framework.py", line 3167, in append_op
kwargs.get("stop_gradient", False))
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/dygraph/tracer.py", line 45, in trace_op
not stop_gradient)
OSError: (External) CUBLAS error(7).
[Hint: 'CUBLAS_STATUS_INVALID_VALUE'. An unsupported value or parameter was passed to the function (a negative vector size, for example). To correct: ensure that all the parameters being passed have valid values. ] (at /paddle/paddle/fluid/platform/cuda_helper.h:107)
### 版本&环境信息 Version & Environment Information
Paddle version: 2.2.2
Paddle With CUDA: True
OS: Ubuntu 16.04
Python version: 3.7.1
CUDA version: 11.2.67
Build cuda_11.2.r11.2/compiler.29373293_0
cuDNN version: 7.6.5
Nvidia driver version: 460.84
显卡使用的3090,CUDA也曾实验过CUDA11.1,python也用过python3.8,都会报错 | 1.0 | 3090运行nn.Conv2D(3, 16, 3)报错 - ### 问题描述 Issue Description
代码:
import paddle.nn as nn
nn.Conv2D(3, 16, 3)
log:
W0601 15:46:55.629635 21509 device_context.cc:447] Please NOTE: device: 0, GPU Compute Capability: 8.6, Driver API Version: 11.2, Runtime API Version: 11.2
W0601 15:46:55.633584 21509 device_context.cc:465] device: 0, cuDNN Version: 7.6.
Traceback (most recent call last):
File "111.py", line 3, in <module>
nn.Conv2D(3, 16, 3)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/nn/layer/conv.py", line 656, in __init__
data_format=data_format)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/nn/layer/conv.py", line 135, in __init__
default_initializer=_get_default_param_initializer())
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/dygraph/layers.py", line 422, in create_parameter
default_initializer)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/layer_helper_base.py", line 378, in create_parameter
**attr._to_kwargs(with_initializer=True))
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/framework.py", line 3137, in create_parameter
initializer(param, self)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/initializer.py", line 362, in __call__
stop_gradient=True)
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/framework.py", line 3167, in append_op
kwargs.get("stop_gradient", False))
File "/opt/python3.7/lib/python3.7/site-packages/paddle/fluid/dygraph/tracer.py", line 45, in trace_op
not stop_gradient)
OSError: (External) CUBLAS error(7).
[Hint: 'CUBLAS_STATUS_INVALID_VALUE'. An unsupported value or parameter was passed to the function (a negative vector size, for example). To correct: ensure that all the parameters being passed have valid values. ] (at /paddle/paddle/fluid/platform/cuda_helper.h:107)
### 版本&环境信息 Version & Environment Information
Paddle version: 2.2.2
Paddle With CUDA: True
OS: Ubuntu 16.04
Python version: 3.7.1
CUDA version: 11.2.67
Build cuda_11.2.r11.2/compiler.29373293_0
cuDNN version: 7.6.5
Nvidia driver version: 460.84
显卡使用的3090,CUDA也曾实验过CUDA11.1,python也用过python3.8,都会报错 | non_defect | 报错 问题描述 issue description 代码 import paddle nn as nn nn log device context cc please note device gpu compute capability driver api version runtime api version device context cc device cudnn version traceback most recent call last file py line in nn file opt lib site packages paddle nn layer conv py line in init data format data format file opt lib site packages paddle nn layer conv py line in init default initializer get default param initializer file opt lib site packages paddle fluid dygraph layers py line in create parameter default initializer file opt lib site packages paddle fluid layer helper base py line in create parameter attr to kwargs with initializer true file opt lib site packages paddle fluid framework py line in create parameter initializer param self file opt lib site packages paddle fluid initializer py line in call stop gradient true file opt lib site packages paddle fluid framework py line in append op kwargs get stop gradient false file opt lib site packages paddle fluid dygraph tracer py line in trace op not stop gradient oserror external cublas error at paddle paddle fluid platform cuda helper h 版本 环境信息 version environment information paddle version paddle with cuda true os ubuntu python version cuda version build cuda compiler cudnn version nvidia driver version , , ,都会报错 | 0 |
59,470 | 14,593,555,725 | IssuesEvent | 2020-12-19 23:33:08 | sandboxie-plus/Sandboxie | https://api.github.com/repos/sandboxie-plus/Sandboxie | closed | File dialog window for "Add File/Folder" in "Resource Access" | enhancement fixed in next build | It would be nice to have the option instead of typing the strings in the "Path" field. Thanks | 1.0 | File dialog window for "Add File/Folder" in "Resource Access" - It would be nice to have the option instead of typing the strings in the "Path" field. Thanks | non_defect | file dialog window for add file folder in resource access it would be nice to have the option instead of typing the strings in the path field thanks | 0 |
119,700 | 25,562,155,269 | IssuesEvent | 2022-11-30 11:37:20 | vegaprotocol/specs | https://api.github.com/repos/vegaprotocol/specs | closed | Add ACs for `reward.staking.delegation` network behaviour after parameter changes | ac-code-remediation | In order to write QA tests to prove that the network behaviour is as expected after a network parameter change we need to add ACs to the relevant specifications for the following network parameters:
- [ ] reward.staking.delegation.delegatorShare
- [ ] reward.staking.delegation.optimalStakeMultiplier
NOTE this should be done **AFTER**:
- https://github.com/vegaprotocol/specs/issues/1405
- https://github.com/vegaprotocol/specs/issues/1406
- https://github.com/vegaprotocol/specs/issues/1407
- https://github.com/vegaprotocol/specs/issues/1408
- https://github.com/vegaprotocol/specs/issues/1409 | 1.0 | Add ACs for `reward.staking.delegation` network behaviour after parameter changes - In order to write QA tests to prove that the network behaviour is as expected after a network parameter change we need to add ACs to the relevant specifications for the following network parameters:
- [ ] reward.staking.delegation.delegatorShare
- [ ] reward.staking.delegation.optimalStakeMultiplier
NOTE this should be done **AFTER**:
- https://github.com/vegaprotocol/specs/issues/1405
- https://github.com/vegaprotocol/specs/issues/1406
- https://github.com/vegaprotocol/specs/issues/1407
- https://github.com/vegaprotocol/specs/issues/1408
- https://github.com/vegaprotocol/specs/issues/1409 | non_defect | add acs for reward staking delegation network behaviour after parameter changes in order to write qa tests to prove that the network behaviour is as expected after a network parameter change we need to add acs to the relevant specifications for the following network parameters reward staking delegation delegatorshare reward staking delegation optimalstakemultiplier note this should be done after | 0 |
232,915 | 25,707,010,129 | IssuesEvent | 2022-12-07 01:53:02 | NirShaharabani/python-mini-projects | https://api.github.com/repos/NirShaharabani/python-mini-projects | opened | requests-2.24.0-py2.py3-none-any.whl: 1 vulnerabilities (highest severity is: 7.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>requests-2.24.0-py2.py3-none-any.whl</b></p></summary>
<p></p>
<p>Path to dependency file: /projects/download GeeksForGeeks articles/requirements.txt</p>
<p>Path to vulnerable library: /projects/Scraping Medium Articles/requirements.txt,/projects/Scrape_quotes/requirements.txt,/projects/All_links_from_given_webpage/requirements.txt,/projects/racing_barchart_animation/requirements.txt,/projects/Movie Information Scraper/requirements.txt,/projects/download GeeksForGeeks articles/requirements.txt</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (requests version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-33503](https://www.mend.io/vulnerability-database/CVE-2021-33503) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | urllib3-1.25.11-py2.py3-none-any.whl | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33503</summary>
### Vulnerable Library - <b>urllib3-1.25.11-py2.py3-none-any.whl</b></p>
<p>HTTP library with thread-safe connection pooling, file post, and more.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /projects/Scraping Medium Articles/requirements.txt</p>
<p>Path to vulnerable library: /projects/Scraping Medium Articles/requirements.txt,/projects/Scrape_quotes/requirements.txt,/projects/All_links_from_given_webpage/requirements.txt,/projects/racing_barchart_animation/requirements.txt,/projects/Movie Information Scraper/requirements.txt,/projects/download GeeksForGeeks articles/requirements.txt</p>
<p>
Dependency Hierarchy:
- requests-2.24.0-py2.py3-none-any.whl (Root Library)
- :x: **urllib3-1.25.11-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue was discovered in urllib3 before 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL were passed as a parameter or redirected to via an HTTP redirect.
<p>Publish Date: 2021-06-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-33503>CVE-2021-33503</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p>
<p>Release Date: 2021-06-29</p>
<p>Fix Resolution: urllib3 - 1.26.5</p>
</p>
<p></p>
</details> | True | requests-2.24.0-py2.py3-none-any.whl: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>requests-2.24.0-py2.py3-none-any.whl</b></p></summary>
<p></p>
<p>Path to dependency file: /projects/download GeeksForGeeks articles/requirements.txt</p>
<p>Path to vulnerable library: /projects/Scraping Medium Articles/requirements.txt,/projects/Scrape_quotes/requirements.txt,/projects/All_links_from_given_webpage/requirements.txt,/projects/racing_barchart_animation/requirements.txt,/projects/Movie Information Scraper/requirements.txt,/projects/download GeeksForGeeks articles/requirements.txt</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (requests version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-33503](https://www.mend.io/vulnerability-database/CVE-2021-33503) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | urllib3-1.25.11-py2.py3-none-any.whl | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-33503</summary>
### Vulnerable Library - <b>urllib3-1.25.11-py2.py3-none-any.whl</b></p>
<p>HTTP library with thread-safe connection pooling, file post, and more.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/56/aa/4ef5aa67a9a62505db124a5cb5262332d1d4153462eb8fd89c9fa41e5d92/urllib3-1.25.11-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /projects/Scraping Medium Articles/requirements.txt</p>
<p>Path to vulnerable library: /projects/Scraping Medium Articles/requirements.txt,/projects/Scrape_quotes/requirements.txt,/projects/All_links_from_given_webpage/requirements.txt,/projects/racing_barchart_animation/requirements.txt,/projects/Movie Information Scraper/requirements.txt,/projects/download GeeksForGeeks articles/requirements.txt</p>
<p>
Dependency Hierarchy:
- requests-2.24.0-py2.py3-none-any.whl (Root Library)
- :x: **urllib3-1.25.11-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue was discovered in urllib3 before 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL were passed as a parameter or redirected to via an HTTP redirect.
<p>Publish Date: 2021-06-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-33503>CVE-2021-33503</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p>
<p>Release Date: 2021-06-29</p>
<p>Fix Resolution: urllib3 - 1.26.5</p>
</p>
<p></p>
</details> | non_defect | requests none any whl vulnerabilities highest severity is vulnerable library requests none any whl path to dependency file projects download geeksforgeeks articles requirements txt path to vulnerable library projects scraping medium articles requirements txt projects scrape quotes requirements txt projects all links from given webpage requirements txt projects racing barchart animation requirements txt projects movie information scraper requirements txt projects download geeksforgeeks articles requirements txt vulnerabilities cve severity cvss dependency type fixed in requests version remediation available high none any whl transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library none any whl http library with thread safe connection pooling file post and more library home page a href path to dependency file projects scraping medium articles requirements txt path to vulnerable library projects scraping medium articles requirements txt projects scrape quotes requirements txt projects all links from given webpage requirements txt projects racing barchart animation requirements txt projects movie information scraper requirements txt projects download geeksforgeeks articles requirements txt dependency hierarchy requests none any whl root library x none any whl vulnerable library found in base branch master vulnerability details an issue was discovered in before when provided with a url containing many characters in the authority component the authority regular expression exhibits catastrophic backtracking causing a denial of service if a url were passed as a parameter or redirected to via an http redirect publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
285,285 | 31,150,348,577 | IssuesEvent | 2023-08-16 09:31:08 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [Security Solution] Alerts table does not show threat enrichments values | bug triage_needed Team: SecuritySolution | **Describe the bug:**
This issue was discovered while fixing skipped cypress test https://github.com/elastic/kibana/issues/162818
Here is the link to the slack discussion which describes all the details and root causes https://elastic.slack.com/archives/C056TQ5J81Y/p1692115257121509
**Steps to reproduce:**
1. Create threat indicator rule which generates alerts with threat enrichments
2. Open Rule details page
3. Scroll down to Alerts Table
4. Add `threat.enrichments.*` fields
**Current behavior:**
Alerts table does not show `threat.enrichments.*` fields values and instead displays `-`
**Expected behavior:**
Alerts table should show `threat.enrichments.*` fields values if those exist
**Screenshots (if relevant):**
https://github.com/elastic/kibana/assets/2700761/ea5518a7-d0cf-442e-b253-dc805d4bf8dc
In this recording you can see that alerts table does not show any of `threat.enrichments.*` fields values. Second part of the recording shows that values exist by adding the same fields in timeline.
cc @michaelolo24 | True | [Security Solution] Alerts table does not show threat enrichments values - **Describe the bug:**
This issue was discovered while fixing skipped cypress test https://github.com/elastic/kibana/issues/162818
Here is the link to the slack discussion which describes all the details and root causes https://elastic.slack.com/archives/C056TQ5J81Y/p1692115257121509
**Steps to reproduce:**
1. Create threat indicator rule which generates alerts with threat enrichments
2. Open Rule details page
3. Scroll down to Alerts Table
4. Add `threat.enrichments.*` fields
**Current behavior:**
Alerts table does not show `threat.enrichments.*` fields values and instead displays `-`
**Expected behavior:**
Alerts table should show `threat.enrichments.*` fields values if those exist
**Screenshots (if relevant):**
https://github.com/elastic/kibana/assets/2700761/ea5518a7-d0cf-442e-b253-dc805d4bf8dc
In this recording you can see that alerts table does not show any of `threat.enrichments.*` fields values. Second part of the recording shows that values exist by adding the same fields in timeline.
cc @michaelolo24 | non_defect | alerts table does not show threat enrichments values describe the bug this issue was discovered while fixing skipped cypress test here is the link to the slack discussion which describes all the details and root causes steps to reproduce create threat indicator rule which generates alerts with threat enrichments open rule details page scroll down to alerts table add threat enrichments fields current behavior alerts table does not show threat enrichments fields values and instead displays expected behavior alerts table should show threat enrichments fields values if those exist screenshots if relevant in this recording you can see that alerts table does not show any of threat enrichments fields values second part of the recording shows that values exist by adding the same fields in timeline cc | 0 |
439,549 | 30,703,020,728 | IssuesEvent | 2023-07-27 02:17:10 | hyc-sg/Helios-public | https://api.github.com/repos/hyc-sg/Helios-public | reopened | Conveyor Belt Speed Change | documentation CIP request (functional) request (data) | **Machine ID**
SA031/alpha system
**Describe the bug**
Footprint Mike Miles said, for the Conveyor Belt Speed change, they would like a software revision that permits the user to set the belt speed anywhere between 30-105.
Currently we have 3 speeds, 30, 60, 105. They want to be able to adjust the conveyor belt to run a bit faster than their infeed product transfer belt.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Run at ...
4. Bug present
**Expected behavior**
A clear and concise description of what you expected to happen.
**Images or Screenshots**
Please try to capture images or screenshots of the bug for better understanding and resolution.
**Additional context**
Add any other context about the problem here.
| 1.0 | Conveyor Belt Speed Change - **Machine ID**
SA031/alpha system
**Describe the bug**
Footprint Mike Miles said, for the Conveyor Belt Speed change, they would like a software revision that permits the user to set the belt speed anywhere between 30-105.
Currently we have 3 speeds, 30, 60, 105. They want to be able to adjust the conveyor belt to run a bit faster than their infeed product transfer belt.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Run at ...
4. Bug present
**Expected behavior**
A clear and concise description of what you expected to happen.
**Images or Screenshots**
Please try to capture images or screenshots of the bug for better understanding and resolution.
**Additional context**
Add any other context about the problem here.
| non_defect | conveyor belt speed change machine id alpha system describe the bug footprint mike miles said for the conveyor belt speed change they would like a software revision that permits the user to set the belt speed anywhere between currently we have speeds they want to be able to adjust the conveyor belt to run a bit faster than their infeed product transfer belt to reproduce steps to reproduce the behavior go to click on run at bug present expected behavior a clear and concise description of what you expected to happen images or screenshots please try to capture images or screenshots of the bug for better understanding and resolution additional context add any other context about the problem here | 0 |
69,556 | 22,494,374,309 | IssuesEvent | 2022-06-23 05:59:19 | hazelcast/hazelcast-go-client | https://api.github.com/repos/hazelcast/hazelcast-go-client | opened | Failover does not respect individual failover cluster SSL settings | Type: Defect Source: Internal to-jira | #### Expected behaviour
Each SSL setting in the individual failover cluster items should be respected when trying to connect to the cluster:
```
config := hazelcast.Config{}
c1 := cluster.Config{}
c1.Network.SSL = sslConfig1
c2 := cluster.Config{}
c1.Network.SSL = sslConfig2
config.Failover.Enabled = true
config.Failover.SetConfigs(c1, c2)
client, err := hazelcast.StartNewClientWithConfig(context.Background(), config)
```
#### Actual behaviour
Only the SSL setting in the root `Cluster` configuration is used.
| 1.0 | Failover does not respect individual failover cluster SSL settings - #### Expected behaviour
Each SSL setting in the individual failover cluster items should be respected when trying to connect to the cluster:
```
config := hazelcast.Config{}
c1 := cluster.Config{}
c1.Network.SSL = sslConfig1
c2 := cluster.Config{}
c1.Network.SSL = sslConfig2
config.Failover.Enabled = true
config.Failover.SetConfigs(c1, c2)
client, err := hazelcast.StartNewClientWithConfig(context.Background(), config)
```
#### Actual behaviour
Only the SSL setting in the root `Cluster` configuration is used.
| defect | failover does not respect individual failover cluster ssl settings expected behaviour each ssl setting in the individual failover cluster items should be respected when trying to connect to the cluster config hazelcast config cluster config network ssl cluster config network ssl config failover enabled true config failover setconfigs client err hazelcast startnewclientwithconfig context background config actual behaviour only the ssl setting in the root cluster configuration is used | 1 |
150,923 | 13,385,560,468 | IssuesEvent | 2020-09-02 13:37:48 | dojot/dojot | https://api.github.com/repos/dojot/dojot | opened | [Auth] Improve documentation | Priority:Low Type:Documentation Type:Enhancement | INDIVIDUAL USER SETTINGS
Add the user identification (user_id or username) in `Get user info` and `Remove user`


| 1.0 | [Auth] Improve documentation - INDIVIDUAL USER SETTINGS
Add the user identification (user_id or username) in `Get user info` and `Remove user`


| non_defect | improve documentation individual user settings add the user identification user id or username in get user info and remove user | 0 |
124,167 | 12,226,383,123 | IssuesEvent | 2020-05-03 10:39:20 | PhuongPhg/SAVABLE | https://api.github.com/repos/PhuongPhg/SAVABLE | reopened | Test PHP Crud | documentation | Use cypress to test the function add, delete, update items. Also give more test cases to the login and signup | 1.0 | Test PHP Crud - Use cypress to test the function add, delete, update items. Also give more test cases to the login and signup | non_defect | test php crud use cypress to test the function add delete update items also give more test cases to the login and signup | 0 |
80,312 | 30,220,318,972 | IssuesEvent | 2023-07-05 18:49:35 | dotCMS/core | https://api.github.com/repos/dotCMS/core | closed | Fix: Bayesian calculation in Goal to minimize a Metric | Type : Defect Merged QA : Passed Internal QA : Needs Internal Team : Falcon dotCMS : Experiments Next Release | ### Parent Issue
https://github.com/dotCMS/core/issues/22222
### Problem Statement
We have to type of Goal:
- Goal to minimize for Example Bounce Rate, when you have a Bounce Rate goal you want to minimize the amount of Bounce Rate for a specific page, so in this case each Bounce Rate is a fail.
- Goal to maximize for example Reach Page, when you have a Reach Page Goal you want to Maximixe the amount of times the a specific page is reach, so in this case is Reach Page is a Success.
Right now we are taking the amount from the Goal as a Success no matter the kind of Goal the it is, so in the case of we are taking the fail as success.
### Steps to Reproduce
- Create a Experiment, Add a Bounce Rate, a Variant, and 100% of traffic location.
- Start it.
- Go to the Front end to be assign inside the Experiment.
- Navigate into the Experiment's Page, and Close the Tab.
- Check the Experiment Result, You must have a Bounce Rate inside the Variant that you was assing, and this Variant is going to be the winner.
### Acceptance Criteria
The Winner must be the Variant with less Bounce Rate.
### dotCMS Version
latest in master
### Proposed Objective
Please Select
### Proposed Priority
Please Select
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | Fix: Bayesian calculation in Goal to minimize a Metric - ### Parent Issue
https://github.com/dotCMS/core/issues/22222
### Problem Statement
We have to type of Goal:
- Goal to minimize for Example Bounce Rate, when you have a Bounce Rate goal you want to minimize the amount of Bounce Rate for a specific page, so in this case each Bounce Rate is a fail.
- Goal to maximize for example Reach Page, when you have a Reach Page Goal you want to Maximixe the amount of times the a specific page is reach, so in this case is Reach Page is a Success.
Right now we are taking the amount from the Goal as a Success no matter the kind of Goal the it is, so in the case of we are taking the fail as success.
### Steps to Reproduce
- Create a Experiment, Add a Bounce Rate, a Variant, and 100% of traffic location.
- Start it.
- Go to the Front end to be assign inside the Experiment.
- Navigate into the Experiment's Page, and Close the Tab.
- Check the Experiment Result, You must have a Bounce Rate inside the Variant that you was assing, and this Variant is going to be the winner.
### Acceptance Criteria
The Winner must be the Variant with less Bounce Rate.
### dotCMS Version
latest in master
### Proposed Objective
Please Select
### Proposed Priority
Please Select
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
_No response_
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | defect | fix bayesian calculation in goal to minimize a metric parent issue problem statement we have to type of goal goal to minimize for example bounce rate when you have a bounce rate goal you want to minimize the amount of bounce rate for a specific page so in this case each bounce rate is a fail goal to maximize for example reach page when you have a reach page goal you want to maximixe the amount of times the a specific page is reach so in this case is reach page is a success right now we are taking the amount from the goal as a success no matter the kind of goal the it is so in the case of we are taking the fail as success steps to reproduce create a experiment add a bounce rate a variant and of traffic location start it go to the front end to be assign inside the experiment navigate into the experiment s page and close the tab check the experiment result you must have a bounce rate inside the variant that you was assing and this variant is going to be the winner acceptance criteria the winner must be the variant with less bounce rate dotcms version latest in master proposed objective please select proposed priority please select external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response | 1 |
13,522 | 2,763,720,885 | IssuesEvent | 2015-04-29 11:27:02 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | [LINQ] OfType<double>() gets different results compared to .Net | defect | ```C#
using Bridge;
using Bridge.Html5;
using System;
using System.Collections.Generic;
using System.Linq;
namespace Demo
{
public class App
{
[Ready]
public static void Main()
{
object[] numbers = { null, 1.0, "two", 3, "four", 5, "six", 7.0 };
var doubleNumbers = numbers.OfType<double>().ToArray();
Console.Log(doubleNumbers);
}
}
}
```
Actual result: [1, 3, 5, 7]
.Net result: [1.0, 7.0]
| 1.0 | [LINQ] OfType<double>() gets different results compared to .Net - ```C#
using Bridge;
using Bridge.Html5;
using System;
using System.Collections.Generic;
using System.Linq;
namespace Demo
{
public class App
{
[Ready]
public static void Main()
{
object[] numbers = { null, 1.0, "two", 3, "four", 5, "six", 7.0 };
var doubleNumbers = numbers.OfType<double>().ToArray();
Console.Log(doubleNumbers);
}
}
}
```
Actual result: [1, 3, 5, 7]
.Net result: [1.0, 7.0]
| defect | oftype gets different results compared to net c using bridge using bridge using system using system collections generic using system linq namespace demo public class app public static void main object numbers null two four six var doublenumbers numbers oftype toarray console log doublenumbers actual result net result | 1 |
536,406 | 15,708,389,218 | IssuesEvent | 2021-03-26 20:27:13 | silknow/converter | https://api.github.com/repos/silknow/converter | closed | Address Knowledge Graph issues reported in D7.1 | enhancement priority | The various issues reported in the[ Deliverable D7.1, section 5](https://drive.google.com/file/d/1pKLZyd86fbpDnBTDqc5hj3azdkCymQeM/view) should be addressed:
- check how to represent object dimensions when the record does not describe a SINGLE object but set of textiles
- check how construction and techniques are represented in the KG
- represent in a structured way warp and weft information which are available from the `P3_has_note` textual property
- check that we appropriately distinguish the production location from the storage location
- check the production location for the objects coming from Monreale (UNIPA) | 1.0 | Address Knowledge Graph issues reported in D7.1 - The various issues reported in the[ Deliverable D7.1, section 5](https://drive.google.com/file/d/1pKLZyd86fbpDnBTDqc5hj3azdkCymQeM/view) should be addressed:
- check how to represent object dimensions when the record does not describe a SINGLE object but set of textiles
- check how construction and techniques are represented in the KG
- represent in a structured way warp and weft information which are available from the `P3_has_note` textual property
- check that we appropriately distinguish the production location from the storage location
- check the production location for the objects coming from Monreale (UNIPA) | non_defect | address knowledge graph issues reported in the various issues reported in the should be addressed check how to represent object dimensions when the record does not describe a single object but set of textiles check how construction and techniques are represented in the kg represent in a structured way warp and weft information which are available from the has note textual property check that we appropriately distinguish the production location from the storage location check the production location for the objects coming from monreale unipa | 0 |
45,354 | 12,741,923,956 | IssuesEvent | 2020-06-26 07:21:41 | line/armeria | https://api.github.com/repos/line/armeria | closed | Some default gRPC request examples do not have default fields set. | defect | For example, `CreateGreetingRequest` is serialized into an empty JSON object when `GrpcDocServicePlugin` builds its default example request:
```protobuf
message CreateGreetingRequest {
FullName full_name = 1;
}
message FullName {
string first_name = 1;
string last_name = 2;
}
```
A workaround is to specify an example object explicitly when building a `DocService`, as documented in https://line.github.io/armeria/server-docservice.html#example-requests-and-headers
/cc @gary-lo | 1.0 | Some default gRPC request examples do not have default fields set. - For example, `CreateGreetingRequest` is serialized into an empty JSON object when `GrpcDocServicePlugin` builds its default example request:
```protobuf
message CreateGreetingRequest {
FullName full_name = 1;
}
message FullName {
string first_name = 1;
string last_name = 2;
}
```
A workaround is to specify an example object explicitly when building a `DocService`, as documented in https://line.github.io/armeria/server-docservice.html#example-requests-and-headers
/cc @gary-lo | defect | some default grpc request examples do not have default fields set for example creategreetingrequest is serialized into an empty json object when grpcdocserviceplugin builds its default example request protobuf message creategreetingrequest fullname full name message fullname string first name string last name a workaround is to specify an example object explicitly when building a docservice as documented in cc gary lo | 1 |
427,840 | 29,865,908,324 | IssuesEvent | 2023-06-20 03:50:52 | morazanm/fsm | https://api.github.com/repos/morazanm/fsm | closed | chore: Make CI tests Blocking on Pull Requests | enhancement Clarity/Documentation DevOps | ## Describe the bug
We should make the CI tests blocking on Pull Request. The repo owner @morazanm has to enable this
| 1.0 | chore: Make CI tests Blocking on Pull Requests - ## Describe the bug
We should make the CI tests blocking on Pull Request. The repo owner @morazanm has to enable this
| non_defect | chore make ci tests blocking on pull requests describe the bug we should make the ci tests blocking on pull request the repo owner morazanm has to enable this | 0 |
38,216 | 8,701,506,734 | IssuesEvent | 2018-12-05 11:46:39 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | HazelcastClient JavaDoc should use "unicast client" instead of "dumb client" | Priority: Medium Team: Client Team: Documentation Type: Defect | See: https://github.com/jerrinot/hazelcast/blob/3438e05a2e8e8a1ac21b9f92798d55ae3b9a9f78/hazelcast-client/src/main/java/com/hazelcast/client/HazelcastClient.java#L33
```
* <h1>Smart vs dumb clients</h1>
* Hazelcast Client enables you to do all Hazelcast operations without being a member of the cluster. Clients can be:
* <ol>
* <li>smart: this means that they immediately can send an operation like map.get(key) to the member that owns that
* specific key.
* </li>
* <li>
* dumb: it will connect to a random member in the cluster and send requests to this member. This member then needs
* to send the request to the correct member.
``` | 1.0 | HazelcastClient JavaDoc should use "unicast client" instead of "dumb client" - See: https://github.com/jerrinot/hazelcast/blob/3438e05a2e8e8a1ac21b9f92798d55ae3b9a9f78/hazelcast-client/src/main/java/com/hazelcast/client/HazelcastClient.java#L33
```
* <h1>Smart vs dumb clients</h1>
* Hazelcast Client enables you to do all Hazelcast operations without being a member of the cluster. Clients can be:
* <ol>
* <li>smart: this means that they immediately can send an operation like map.get(key) to the member that owns that
* specific key.
* </li>
* <li>
* dumb: it will connect to a random member in the cluster and send requests to this member. This member then needs
* to send the request to the correct member.
``` | defect | hazelcastclient javadoc should use unicast client instead of dumb client see smart vs dumb clients hazelcast client enables you to do all hazelcast operations without being a member of the cluster clients can be smart this means that they immediately can send an operation like map get key to the member that owns that specific key dumb it will connect to a random member in the cluster and send requests to this member this member then needs to send the request to the correct member | 1 |
9,103 | 2,615,129,816 | IssuesEvent | 2015-03-01 05:59:34 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | closed | the classpath entry of the google-api-services-calendar's manifest is wrong | auto-migrated Component-Google-APIs Milestone-Version1.12.0 Priority-High Type-Defect | ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
google-api-client-1.11.0-beta.jar
google-api-services-calendar-v3-rev16-1.8.0-beta.jar
Java environment (e.g. Java 6, Android 2.3, App Engine)?
JAVA 6
I'm trying to import the google api calendar by maven into my project. However,
when I'm trying to deploy in a JAVAEE container like jboss as7, I have a
classloading error caused by the content of the manifest file of
google-api-services-calendar-v3-rev16-1.8.0-beta.jar
An example :
Caused by: java.io.FileNotFoundException:
\home\codegen\dependencies\java\stable\dependencies\libs\google-http-client-1.11
.0-beta.jar (Le chemin d'accès spécifié est introuvable)
at java.util.zip.ZipFile.open(Native Method) [rt.jar:1.6.0_23]
at java.util.zip.ZipFile.<init>(ZipFile.java:127) [rt.jar:1.6.0_23]
at java.util.jar.JarFile.<init>(JarFile.java:135) [rt.jar:1.6.0_23]
at java.util.jar.JarFile.<init>(JarFile.java:99) [rt.jar:1.6.0_23]
at org.jboss.as.server.moduleservice.ExternalModuleSpecService.start(ExternalModuleSpecService.java:63) [jboss-as-server-7.1.1.Final.jar:7.1.1.Final]
... 5 more
As you can see below, the classpath contain absolute paths and not relative
paths.
Class-Path:
/home/codegen/dependencies/java/stable/dependencies/libs/google-api-client-1.11.
0-beta.jar
/home/codegen/dependenciesjava/stable/dependencies/libs/google-oauth-client-1.11
.0-beta.jar
/home/codegen/dependencies/java/stable/dependencieslibs/google-http-client-1.11.
0-beta.jar
/home/codegen/dependencies/java/stable/dependencies/libs/guava-11.0.1.jar
Such as it is, this artifact is not deployable by the maven way.
It would be nice to deploy an another release of this artifact with a correct
manifest content.
```
Original issue reported on code.google.com by `alexan...@touret.info` on 24 Oct 2012 at 7:35 | 1.0 | the classpath entry of the google-api-services-calendar's manifest is wrong - ```
Version of google-api-java-client (e.g. 1.5.0-beta)?
google-api-client-1.11.0-beta.jar
google-api-services-calendar-v3-rev16-1.8.0-beta.jar
Java environment (e.g. Java 6, Android 2.3, App Engine)?
JAVA 6
I'm trying to import the google api calendar by maven into my project. However,
when I'm trying to deploy in a JAVAEE container like jboss as7, I have a
classloading error caused by the content of the manifest file of
google-api-services-calendar-v3-rev16-1.8.0-beta.jar
An example :
Caused by: java.io.FileNotFoundException:
\home\codegen\dependencies\java\stable\dependencies\libs\google-http-client-1.11
.0-beta.jar (Le chemin d'accès spécifié est introuvable)
at java.util.zip.ZipFile.open(Native Method) [rt.jar:1.6.0_23]
at java.util.zip.ZipFile.<init>(ZipFile.java:127) [rt.jar:1.6.0_23]
at java.util.jar.JarFile.<init>(JarFile.java:135) [rt.jar:1.6.0_23]
at java.util.jar.JarFile.<init>(JarFile.java:99) [rt.jar:1.6.0_23]
at org.jboss.as.server.moduleservice.ExternalModuleSpecService.start(ExternalModuleSpecService.java:63) [jboss-as-server-7.1.1.Final.jar:7.1.1.Final]
... 5 more
As you can see below, the classpath contain absolute paths and not relative
paths.
Class-Path:
/home/codegen/dependencies/java/stable/dependencies/libs/google-api-client-1.11.
0-beta.jar
/home/codegen/dependenciesjava/stable/dependencies/libs/google-oauth-client-1.11
.0-beta.jar
/home/codegen/dependencies/java/stable/dependencieslibs/google-http-client-1.11.
0-beta.jar
/home/codegen/dependencies/java/stable/dependencies/libs/guava-11.0.1.jar
Such as it is, this artifact is not deployable by the maven way.
It would be nice to deploy an another release of this artifact with a correct
manifest content.
```
Original issue reported on code.google.com by `alexan...@touret.info` on 24 Oct 2012 at 7:35 | defect | the classpath entry of the google api services calendar s manifest is wrong version of google api java client e g beta google api client beta jar google api services calendar beta jar java environment e g java android app engine java i m trying to import the google api calendar by maven into my project however when i m trying to deploy in a javaee container like jboss i have a classloading error caused by the content of the manifest file of google api services calendar beta jar an example caused by java io filenotfoundexception home codegen dependencies java stable dependencies libs google http client beta jar le chemin d accès spécifié est introuvable at java util zip zipfile open native method at java util zip zipfile zipfile java at java util jar jarfile jarfile java at java util jar jarfile jarfile java at org jboss as server moduleservice externalmodulespecservice start externalmodulespecservice java more as you can see below the classpath contain absolute paths and not relative paths class path home codegen dependencies java stable dependencies libs google api client beta jar home codegen dependenciesjava stable dependencies libs google oauth client beta jar home codegen dependencies java stable dependencieslibs google http client beta jar home codegen dependencies java stable dependencies libs guava jar such as it is this artifact is not deployable by the maven way it would be nice to deploy an another release of this artifact with a correct manifest content original issue reported on code google com by alexan touret info on oct at | 1 |
22,406 | 3,643,564,640 | IssuesEvent | 2016-02-15 02:50:25 | prettydiff/prettydiff | https://api.github.com/repos/prettydiff/prettydiff | closed | Improperly inserted space in markup | Defect Parsing QA | From: https://github.com/Glavin001/atom-beautify/issues/778
<p>
Thanks for joining ProductName<sup>®</sup> 2016.
</p>
Becomes
<p>
Thanks for joining ProductName
<sup>®</sup>
2016.
</p> | 1.0 | Improperly inserted space in markup - From: https://github.com/Glavin001/atom-beautify/issues/778
<p>
Thanks for joining ProductName<sup>®</sup> 2016.
</p>
Becomes
<p>
Thanks for joining ProductName
<sup>®</sup>
2016.
</p> | defect | improperly inserted space in markup from thanks for joining productname reg becomes thanks for joining productname reg | 1 |
21,068 | 3,455,776,753 | IssuesEvent | 2015-12-17 21:40:53 | netty/netty | https://api.github.com/repos/netty/netty | opened | HTTP/2 DefaultHttp2RemoteFlowController Stream writability notification broken | defect | `DefaultHttp2RemoteFlowController.ListenerWritabilityMonitor` no longer reliably detects when a stream's writability change occurs. `ListenerWritabilityMonitor` was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the `PriorityStreamByteDistributor` to call `write` for each stream during its iteration process. However the new `StreamByteDistributor` classes do not do an iteration over all active streams and so this assumption is now invalid.
The impact is isolated to stream writability change notifications and no impact unless you explicitly add a listener. | 1.0 | HTTP/2 DefaultHttp2RemoteFlowController Stream writability notification broken - `DefaultHttp2RemoteFlowController.ListenerWritabilityMonitor` no longer reliably detects when a stream's writability change occurs. `ListenerWritabilityMonitor` was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the `PriorityStreamByteDistributor` to call `write` for each stream during its iteration process. However the new `StreamByteDistributor` classes do not do an iteration over all active streams and so this assumption is now invalid.
The impact is isolated to stream writability change notifications and no impact unless you explicitly add a listener. | defect | http stream writability notification broken listenerwritabilitymonitor no longer reliably detects when a stream s writability change occurs listenerwritabilitymonitor was implemented to avoid duplicating iteration over all streams when possible and instead was relying on the prioritystreambytedistributor to call write for each stream during its iteration process however the new streambytedistributor classes do not do an iteration over all active streams and so this assumption is now invalid the impact is isolated to stream writability change notifications and no impact unless you explicitly add a listener | 1 |
452,100 | 13,046,085,685 | IssuesEvent | 2020-07-29 08:24:59 | onaio/reveal-frontend | https://api.github.com/repos/onaio/reveal-frontend | closed | IRS Map sidebar improvements | Priority: High blocked | - [x] For Namibia, Spray Coverage should be listed above Found Coverage
- [x] For Zambia, Spray Coverage should be listed above Found Coverage
- [ ] For Zambia, the not sprayed reasons and duplicate reasons should be listed similar to how it was done in mSpray e.g. https://macepa.mspray.onalabs.org/reactive/irs/chw/3106/3122 (click on target area stats top right) | 1.0 | IRS Map sidebar improvements - - [x] For Namibia, Spray Coverage should be listed above Found Coverage
- [x] For Zambia, Spray Coverage should be listed above Found Coverage
- [ ] For Zambia, the not sprayed reasons and duplicate reasons should be listed similar to how it was done in mSpray e.g. https://macepa.mspray.onalabs.org/reactive/irs/chw/3106/3122 (click on target area stats top right) | non_defect | irs map sidebar improvements for namibia spray coverage should be listed above found coverage for zambia spray coverage should be listed above found coverage for zambia the not sprayed reasons and duplicate reasons should be listed similar to how it was done in mspray e g click on target area stats top right | 0 |
42,639 | 11,194,684,353 | IssuesEvent | 2020-01-03 02:22:21 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Oracle Sequences can't get after upgrading JOOQ version from 3.10.4 to 3.12.1 | T: Defect | ### Expected behavior and actual behavior:
Before, I used the version of 3.10.4 in my project. We created a util for checking the SQL field is exist or not, and the code as below:
_Record.field("SEQ_USERID"."nextval")_
In the version of 3.10.4, when checking sequence by Record.field("SEQ_USERID"."nextval"), it returns nextval. And then we think this statement can be executed.
But now, we upgraded the version to 3.12.1, and found this code can't work well.
Record.field("SEQ_USERID"."nextval") returns null, and we defaulted it as an exception.
Could you please share the changes from 3.10.* to 3.12.* for this path? And could you please provide us some suggestions for this function?
### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve):
### Versions:
- jOOQ: 3.12.1
- Java:
- Database (include vendor): Oracle 11
- OS:
- JDBC Driver (include name if inofficial driver):
| 1.0 | Oracle Sequences can't get after upgrading JOOQ version from 3.10.4 to 3.12.1 - ### Expected behavior and actual behavior:
Before, I used the version of 3.10.4 in my project. We created a util for checking the SQL field is exist or not, and the code as below:
_Record.field("SEQ_USERID"."nextval")_
In the version of 3.10.4, when checking sequence by Record.field("SEQ_USERID"."nextval"), it returns nextval. And then we think this statement can be executed.
But now, we upgraded the version to 3.12.1, and found this code can't work well.
Record.field("SEQ_USERID"."nextval") returns null, and we defaulted it as an exception.
Could you please share the changes from 3.10.* to 3.12.* for this path? And could you please provide us some suggestions for this function?
### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve):
### Versions:
- jOOQ: 3.12.1
- Java:
- Database (include vendor): Oracle 11
- OS:
- JDBC Driver (include name if inofficial driver):
| defect | oracle sequences can t get after upgrading jooq version from to expected behavior and actual behavior before i used the version of in my project we created a util for checking the sql field is exist or not and the code as below record field seq userid nextval in the version of when checking sequence by record field seq userid nextval it returns nextval and then we think this statement can be executed but now we upgraded the version to and found this code can t work well record field seq userid nextval returns null and we defaulted it as an exception could you please share the changes from to for this path and could you please provide us some suggestions for this function steps to reproduce the problem if possible create an mcve versions jooq java database include vendor oracle os jdbc driver include name if inofficial driver | 1 |
236,022 | 18,068,278,856 | IssuesEvent | 2021-09-20 21:58:09 | girlscript/winter-of-contributing | https://api.github.com/repos/girlscript/winter-of-contributing | opened | GoLang: 8.6 DOC Concurrency - Select | documentation GWOC21 | <hr>
## Description 📜
**Buffered Channels**
Basic-Intermediate-Advanced information about `Select` in Golang, also needs to create some examples.
This must be inside `Concurrency` folder.
<!-- Please describe the issue in brief. -->
**References:**
- [Geeks for Geeks](https://www.geeksforgeeks.org/select-statement-in-go-language/)
- [A Tour of Go](https://tour.golang.org/concurrency/5)
- [gobot](https://golangbot.com/select)
<hr>
## Domain of Contribution 📊
<!----Please delete options that are not relevant.And in order to tick the check box just but x inside them for example [x] like this----->
- [x] Golang
<hr>
## Instructions ✔
* Kindly make a README going over the subtopic.
* Changes should be made under the **Golang** branch only and push your PR to the same.
* This issue is for **GWOC'21** contributors only.
* You need to push in a single commit. If there exists more than one commit then you need to **squash** your commits into one commit. | 1.0 | GoLang: 8.6 DOC Concurrency - Select - <hr>
## Description 📜
**Buffered Channels**
Basic-Intermediate-Advanced information about `Select` in Golang, also needs to create some examples.
This must be inside `Concurrency` folder.
<!-- Please describe the issue in brief. -->
**References:**
- [Geeks for Geeks](https://www.geeksforgeeks.org/select-statement-in-go-language/)
- [A Tour of Go](https://tour.golang.org/concurrency/5)
- [gobot](https://golangbot.com/select)
<hr>
## Domain of Contribution 📊
<!----Please delete options that are not relevant.And in order to tick the check box just but x inside them for example [x] like this----->
- [x] Golang
<hr>
## Instructions ✔
* Kindly make a README going over the subtopic.
* Changes should be made under the **Golang** branch only and push your PR to the same.
* This issue is for **GWOC'21** contributors only.
* You need to push in a single commit. If there exists more than one commit then you need to **squash** your commits into one commit. | non_defect | golang doc concurrency select description 📜 buffered channels basic intermediate advanced information about select in golang also needs to create some examples this must be inside concurrency folder references domain of contribution 📊 golang instructions ✔ kindly make a readme going over the subtopic changes should be made under the golang branch only and push your pr to the same this issue is for gwoc contributors only you need to push in a single commit if there exists more than one commit then you need to squash your commits into one commit | 0 |
22,303 | 3,626,601,585 | IssuesEvent | 2016-02-10 01:54:12 | prettydiff/prettydiff | https://api.github.com/repos/prettydiff/prettydiff | closed | JSX As Argument | Defect Parsing QA | First up, thanks for the awesome work.
The code in question it the following:
```
renderer.render(<MyComponent items={[1, 2, 3]}/>);
```
When I run it through prettydiff I get the following output which breaks the code.
```
renderer.render(< MyComponent items = {
[1, 2, 3]
} />);
```
It seems to work find if I first declare the variable and pass that in. For some reason it's not enjoying the JSX being passed in as an argument. The code below gets formatted just fine in the end:
```
const componentToRender = <MyComponent items={[1, 2, 3]}/>
renderer.render(componentToRender);
```
Any ideas on what I can do to fix this? I tried a few different components but nothing seems to work.
Thanks! | 1.0 | JSX As Argument - First up, thanks for the awesome work.
The code in question it the following:
```
renderer.render(<MyComponent items={[1, 2, 3]}/>);
```
When I run it through prettydiff I get the following output which breaks the code.
```
renderer.render(< MyComponent items = {
[1, 2, 3]
} />);
```
It seems to work find if I first declare the variable and pass that in. For some reason it's not enjoying the JSX being passed in as an argument. The code below gets formatted just fine in the end:
```
const componentToRender = <MyComponent items={[1, 2, 3]}/>
renderer.render(componentToRender);
```
Any ideas on what I can do to fix this? I tried a few different components but nothing seems to work.
Thanks! | defect | jsx as argument first up thanks for the awesome work the code in question it the following renderer render when i run it through prettydiff i get the following output which breaks the code renderer render mycomponent items it seems to work find if i first declare the variable and pass that in for some reason it s not enjoying the jsx being passed in as an argument the code below gets formatted just fine in the end const componenttorender renderer render componenttorender any ideas on what i can do to fix this i tried a few different components but nothing seems to work thanks | 1 |
256,851 | 27,561,731,123 | IssuesEvent | 2023-03-07 22:42:47 | samqws-marketing/box_box-ui-elements | https://api.github.com/repos/samqws-marketing/box_box-ui-elements | closed | CVE-2021-23368 (Medium) detected in multiple libraries - autoclosed | security vulnerability | ## CVE-2021-23368 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.14.tgz</b>, <b>postcss-7.0.23.tgz</b>, <b>postcss-7.0.26.tgz</b>, <b>postcss-7.0.17.tgz</b>, <b>postcss-7.0.2.tgz</b></p></summary>
<p>
<details><summary><b>postcss-7.0.14.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.14.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-12.0.0.tgz (Root Library)
- postcss-less-3.1.4.tgz
- :x: **postcss-7.0.14.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.23.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.23.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-5.3.9.tgz (Root Library)
- core-5.3.9.tgz
- css-loader-3.2.0.tgz
- postcss-modules-scope-2.1.0.tgz
- :x: **postcss-7.0.23.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.26.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.26.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.26.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- autoprefixer-9.7.4.tgz (Root Library)
- :x: **postcss-7.0.26.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.17.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.17.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.17.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-5.3.9.tgz (Root Library)
- core-5.3.9.tgz
- css-loader-3.2.0.tgz
- :x: **postcss-7.0.17.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.2.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.2.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-5.3.9.tgz (Root Library)
- core-5.3.9.tgz
- postcss-flexbugs-fixes-4.1.0.tgz
- :x: **postcss-7.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23368>CVE-2021-23368</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (stylelint): 12.0.1</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (@storybook/react): 5.3.10</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (autoprefixer): 9.7.5</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (@storybook/react): 5.3.10</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (@storybook/react): 5.3.10</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2021-23368 (Medium) detected in multiple libraries - autoclosed - ## CVE-2021-23368 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.14.tgz</b>, <b>postcss-7.0.23.tgz</b>, <b>postcss-7.0.26.tgz</b>, <b>postcss-7.0.17.tgz</b>, <b>postcss-7.0.2.tgz</b></p></summary>
<p>
<details><summary><b>postcss-7.0.14.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.14.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.14.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-12.0.0.tgz (Root Library)
- postcss-less-3.1.4.tgz
- :x: **postcss-7.0.14.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.23.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.23.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-5.3.9.tgz (Root Library)
- core-5.3.9.tgz
- css-loader-3.2.0.tgz
- postcss-modules-scope-2.1.0.tgz
- :x: **postcss-7.0.23.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.26.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.26.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.26.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- autoprefixer-9.7.4.tgz (Root Library)
- :x: **postcss-7.0.26.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.17.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.17.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.17.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-5.3.9.tgz (Root Library)
- core-5.3.9.tgz
- css-loader-3.2.0.tgz
- :x: **postcss-7.0.17.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.2.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.2.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-5.3.9.tgz (Root Library)
- core-5.3.9.tgz
- postcss-flexbugs-fixes-4.1.0.tgz
- :x: **postcss-7.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23368>CVE-2021-23368</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (stylelint): 12.0.1</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (@storybook/react): 5.3.10</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (autoprefixer): 9.7.5</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (@storybook/react): 5.3.10</p><p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (@storybook/react): 5.3.10</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_defect | cve medium detected in multiple libraries autoclosed cve medium severity vulnerability vulnerable libraries postcss tgz postcss tgz postcss tgz postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy stylelint tgz root library postcss less tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy react tgz root library core tgz css loader tgz postcss modules scope tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy autoprefixer tgz root library x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy react tgz root library core tgz css loader tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy react tgz root library core tgz postcss flexbugs fixes tgz x postcss tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package postcss from and before are vulnerable to regular expression denial of service redos during source map parsing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss direct dependency fix resolution stylelint fix resolution postcss direct dependency fix resolution storybook react fix resolution postcss direct dependency fix resolution autoprefixer fix resolution postcss direct dependency fix resolution storybook react fix resolution postcss direct dependency fix resolution storybook react check this box to open an automated fix pr | 0 |
9,921 | 2,616,010,491 | IssuesEvent | 2015-03-02 00:53:58 | jasonhall/bwapi | https://api.github.com/repos/jasonhall/bwapi | closed | UnitType.maxGroundHits() return 0 for Tank (sieged and unsieged) | auto-migrated Component-Logic Priority-Low Type-Defect Usability | ```
What steps will reproduce the problem?
Broodwar->printf("Tank SiegeMode Max hits:
%d",UnitTypes::Terran_Siege_Tank_Tank_Mode.maxGroundHits());
will output: Tank SiegeMode Max hits: 0.
For other units, that I tested it works (marine, firebat, ghost, zealot, goon).
What is the expected output? What do you see instead?
If I understand correctly, this is the number of hits per attack (a zealot
has 16 damage, but per each cooldown (every attack) it makes 2 hits, 8
damage each). For Siege Tank maybe it must be 1 (30 or 70 damage in 1 hit).
Useful, when calculating approximate hp/damage/range/cooldown-based unit
strenght.
What version of the product are you using? On what operating system?
BWAPI r2157 on Windows XP
```
Original issue reported on code.google.com by `Nuclear...@gmail.com` on 24 Mar 2010 at 10:28 | 1.0 | UnitType.maxGroundHits() return 0 for Tank (sieged and unsieged) - ```
What steps will reproduce the problem?
Broodwar->printf("Tank SiegeMode Max hits:
%d",UnitTypes::Terran_Siege_Tank_Tank_Mode.maxGroundHits());
will output: Tank SiegeMode Max hits: 0.
For other units, that I tested it works (marine, firebat, ghost, zealot, goon).
What is the expected output? What do you see instead?
If I understand correctly, this is the number of hits per attack (a zealot
has 16 damage, but per each cooldown (every attack) it makes 2 hits, 8
damage each). For Siege Tank maybe it must be 1 (30 or 70 damage in 1 hit).
Useful, when calculating approximate hp/damage/range/cooldown-based unit
strenght.
What version of the product are you using? On what operating system?
BWAPI r2157 on Windows XP
```
Original issue reported on code.google.com by `Nuclear...@gmail.com` on 24 Mar 2010 at 10:28 | defect | unittype maxgroundhits return for tank sieged and unsieged what steps will reproduce the problem broodwar printf tank siegemode max hits d unittypes terran siege tank tank mode maxgroundhits will output tank siegemode max hits for other units that i tested it works marine firebat ghost zealot goon what is the expected output what do you see instead if i understand correctly this is the number of hits per attack a zealot has damage but per each cooldown every attack it makes hits damage each for siege tank maybe it must be or damage in hit useful when calculating approximate hp damage range cooldown based unit strenght what version of the product are you using on what operating system bwapi on windows xp original issue reported on code google com by nuclear gmail com on mar at | 1 |
637,026 | 20,618,078,385 | IssuesEvent | 2022-03-07 15:00:28 | Bone008/orbiteye | https://api.github.com/repos/Bone008/orbiteye | closed | Visualize by-country heatmap with normal countries | Visualization Logic priority: high | Create by-country heatmap (e.g. number of satellites matching the global filter with this country as an owner) | 1.0 | Visualize by-country heatmap with normal countries - Create by-country heatmap (e.g. number of satellites matching the global filter with this country as an owner) | non_defect | visualize by country heatmap with normal countries create by country heatmap e g number of satellites matching the global filter with this country as an owner | 0 |
108,217 | 16,762,817,852 | IssuesEvent | 2021-06-14 03:16:40 | gms-ws-sandbox/nibrs-pr-test | https://api.github.com/repos/gms-ws-sandbox/nibrs-pr-test | opened | CVE-2018-8037 (Medium) detected in tomcat-embed-core-8.5.20.jar | security vulnerability | ## CVE-2018-8037 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.20.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.20/tomcat-embed-core-8.5.20.jar,nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/tomcat-embed-core-8.5.20.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-core-8.5.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
If an async request was completed by the application at the same time as the container triggered the async timeout, a race condition existed that could result in a user seeing a response intended for a different user. An additional issue was present in the NIO and NIO2 connectors that did not correctly track the closure of the connection when an async request was completed by the application and timed out by the container at the same time. This could also result in a user seeing a response intended for another user. Versions Affected: Apache Tomcat 9.0.0.M9 to 9.0.9 and 8.5.5 to 8.5.31.
<p>Publish Date: 2018-08-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8037>CVE-2018-8037</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-8037">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-8037</a></p>
<p>Release Date: 2018-08-02</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.10,8.5.32,org.apache.tomcat:tomcat-coyote:9.0.10,8.5.32</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.20","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat.embed:tomcat-embed-core:8.5.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.10,8.5.32,org.apache.tomcat:tomcat-coyote:9.0.10,8.5.32"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-8037","vulnerabilityDetails":"If an async request was completed by the application at the same time as the container triggered the async timeout, a race condition existed that could result in a user seeing a response intended for a different user. An additional issue was present in the NIO and NIO2 connectors that did not correctly track the closure of the connection when an async request was completed by the application and timed out by the container at the same time. This could also result in a user seeing a response intended for another user. Versions Affected: Apache Tomcat 9.0.0.M9 to 9.0.9 and 8.5.5 to 8.5.31.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8037","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-8037 (Medium) detected in tomcat-embed-core-8.5.20.jar - ## CVE-2018-8037 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-8.5.20.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Library home page: <a href="http://tomcat.apache.org/">http://tomcat.apache.org/</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/8.5.20/tomcat-embed-core-8.5.20.jar,nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/tomcat-embed-core-8.5.20.jar</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-core-8.5.20.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
If an async request was completed by the application at the same time as the container triggered the async timeout, a race condition existed that could result in a user seeing a response intended for a different user. An additional issue was present in the NIO and NIO2 connectors that did not correctly track the closure of the connection when an async request was completed by the application and timed out by the container at the same time. This could also result in a user seeing a response intended for another user. Versions Affected: Apache Tomcat 9.0.0.M9 to 9.0.9 and 8.5.5 to 8.5.31.
<p>Publish Date: 2018-08-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8037>CVE-2018-8037</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-8037">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-8037</a></p>
<p>Release Date: 2018-08-02</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.10,8.5.32,org.apache.tomcat:tomcat-coyote:9.0.10,8.5.32</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.tomcat.embed","packageName":"tomcat-embed-core","packageVersion":"8.5.20","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.tomcat.embed:tomcat-embed-core:8.5.20","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.tomcat.embed:tomcat-embed-core:9.0.10,8.5.32,org.apache.tomcat:tomcat-coyote:9.0.10,8.5.32"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-8037","vulnerabilityDetails":"If an async request was completed by the application at the same time as the container triggered the async timeout, a race condition existed that could result in a user seeing a response intended for a different user. An additional issue was present in the NIO and NIO2 connectors that did not correctly track the closure of the connection when an async request was completed by the application and timed out by the container at the same time. This could also result in a user seeing a response intended for another user. Versions Affected: Apache Tomcat 9.0.0.M9 to 9.0.9 and 8.5.5 to 8.5.31.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8037","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_defect | cve medium detected in tomcat embed core jar cve medium severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file nibrs pr test tools nibrs fbi service pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar nibrs pr test tools nibrs fbi service target nibrs fbi service web inf lib tomcat embed core jar dependency hierarchy x tomcat embed core jar vulnerable library found in head commit a href found in base branch master vulnerability details if an async request was completed by the application at the same time as the container triggered the async timeout a race condition existed that could result in a user seeing a response intended for a different user an additional issue was present in the nio and connectors that did not correctly track the closure of the connection when an async request was completed by the application and timed out by the container at the same time this could also result in a user seeing a response intended for another user versions affected apache tomcat to and to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache tomcat embed tomcat embed core isminimumfixversionavailable true minimumfixversion org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote basebranches vulnerabilityidentifier cve vulnerabilitydetails if an async request was completed by the application at the same time as the container triggered the async timeout a race condition existed that could result in a user seeing a response intended for a different user an additional issue was present in the nio and connectors that did not correctly track the closure of the connection when an async request was completed by the application and timed out by the container at the same time this could also result in a user seeing a response intended for another user versions affected apache tomcat to and to vulnerabilityurl | 0 |
51,910 | 13,211,336,369 | IssuesEvent | 2020-08-15 22:24:06 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | [millipede] fix the FIXME in python/artists.py (Trac #1249) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1249">https://code.icecube.wisc.edu/projects/icecube/ticket/1249</a>, reported by hdembinskiand owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-02-19T19:00:53",
"_ts": "1455908453690620",
"description": "the FIXME should be fixed or converted into a TODO.\n\nhere is my suggestion on how to make the tables configurable by the user. i would add a string-setting, like so\n\n(in __init__)\nself.addSetting(\"Photon table\", \"$I3_DATA/photon-tables/splines/ems_mie_z20_a10\")\n\nand then parse and complete the string in create(...)",
"reporter": "hdembinski",
"cc": "",
"resolution": "fixed",
"time": "2015-08-20T17:07:42",
"component": "combo reconstruction",
"summary": "[millipede] fix the FIXME in python/artists.py",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [millipede] fix the FIXME in python/artists.py (Trac #1249) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1249">https://code.icecube.wisc.edu/projects/icecube/ticket/1249</a>, reported by hdembinskiand owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-02-19T19:00:53",
"_ts": "1455908453690620",
"description": "the FIXME should be fixed or converted into a TODO.\n\nhere is my suggestion on how to make the tables configurable by the user. i would add a string-setting, like so\n\n(in __init__)\nself.addSetting(\"Photon table\", \"$I3_DATA/photon-tables/splines/ems_mie_z20_a10\")\n\nand then parse and complete the string in create(...)",
"reporter": "hdembinski",
"cc": "",
"resolution": "fixed",
"time": "2015-08-20T17:07:42",
"component": "combo reconstruction",
"summary": "[millipede] fix the FIXME in python/artists.py",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| defect | fix the fixme in python artists py trac migrated from json status closed changetime ts description the fixme should be fixed or converted into a todo n nhere is my suggestion on how to make the tables configurable by the user i would add a string setting like so n n in init nself addsetting photon table data photon tables splines ems mie n nand then parse and complete the string in create reporter hdembinski cc resolution fixed time component combo reconstruction summary fix the fixme in python artists py priority normal keywords milestone owner jbraun type defect | 1 |
743,802 | 25,915,620,484 | IssuesEvent | 2022-12-15 17:06:54 | bcgov/entity | https://api.github.com/repos/bcgov/entity | closed | Filings UI: Investigate business not in good standing but cannot file Annual Report | bug Priority1 ENTITY UX Assurance | ### Description
There are a few things going on here that need to be verified/fixed/implemented:
1. does the API provide the "not in good standing" flag? **yes, see #14612**
2. does the UI display the NIGS alert? **no, that will be fixed in THIS ticket; see also #13881**
3. does the UI handle entity-type-alterations and BEN/BC/CCC/ULC alterations correctly (with goodStanding=true or false)? **no, that will be fixed in THIS ticket**
### Details
Ref: https://dev.bcregistry.ca/business/BC1230074

This business has a draft alteration (Change of Company Information) filing with a "This business is not in good standing" warning indicating that it is probably due to an overdue annual report. This is not wrong, however the Resume button is grayed out so this filing cannot be actioned.
The Annual Report todo task cannot be filed because its action button is disabled (because it is not the first task in the list).
Therefore, a user (and staff) cannot do anything with this business atm.
This situation can happen if a business has an unfinished (draft) alteration when its Annual Report becomes due, and then the AR becomes overdue.
Per discussions below (in the comments);
- [x] NIGS alert should be displayed when flag is set (according to BE business logic).
- [x] NIGS state should not stop a business from filing an alteration (except entity type change alterations).
- Entity type change alterations should display special-case todo items (as per screenshots above and in comments below).
- Other alterations should display normal draft todo item.
- [x] Compliance warnings should not display NIGS alert -- they should display a different alert.
- BE does not provide compliance warnings yet, so new alert is for future use.
| 1.0 | Filings UI: Investigate business not in good standing but cannot file Annual Report - ### Description
There are a few things going on here that need to be verified/fixed/implemented:
1. does the API provide the "not in good standing" flag? **yes, see #14612**
2. does the UI display the NIGS alert? **no, that will be fixed in THIS ticket; see also #13881**
3. does the UI handle entity-type-alterations and BEN/BC/CCC/ULC alterations correctly (with goodStanding=true or false)? **no, that will be fixed in THIS ticket**
### Details
Ref: https://dev.bcregistry.ca/business/BC1230074

This business has a draft alteration (Change of Company Information) filing with a "This business is not in good standing" warning indicating that it is probably due to an overdue annual report. This is not wrong, however the Resume button is grayed out so this filing cannot be actioned.
The Annual Report todo task cannot be filed because its action button is disabled (because it is not the first task in the list).
Therefore, a user (and staff) cannot do anything with this business atm.
This situation can happen if a business has an unfinished (draft) alteration when its Annual Report becomes due, and then the AR becomes overdue.
Per discussions below (in the comments);
- [x] NIGS alert should be displayed when flag is set (according to BE business logic).
- [x] NIGS state should not stop a business from filing an alteration (except entity type change alterations).
- Entity type change alterations should display special-case todo items (as per screenshots above and in comments below).
- Other alterations should display normal draft todo item.
- [x] Compliance warnings should not display NIGS alert -- they should display a different alert.
- BE does not provide compliance warnings yet, so new alert is for future use.
| non_defect | filings ui investigate business not in good standing but cannot file annual report description there are a few things going on here that need to be verified fixed implemented does the api provide the not in good standing flag yes see does the ui display the nigs alert no that will be fixed in this ticket see also does the ui handle entity type alterations and ben bc ccc ulc alterations correctly with goodstanding true or false no that will be fixed in this ticket details ref this business has a draft alteration change of company information filing with a this business is not in good standing warning indicating that it is probably due to an overdue annual report this is not wrong however the resume button is grayed out so this filing cannot be actioned the annual report todo task cannot be filed because its action button is disabled because it is not the first task in the list therefore a user and staff cannot do anything with this business atm this situation can happen if a business has an unfinished draft alteration when its annual report becomes due and then the ar becomes overdue per discussions below in the comments nigs alert should be displayed when flag is set according to be business logic nigs state should not stop a business from filing an alteration except entity type change alterations entity type change alterations should display special case todo items as per screenshots above and in comments below other alterations should display normal draft todo item compliance warnings should not display nigs alert they should display a different alert be does not provide compliance warnings yet so new alert is for future use | 0 |
1,214 | 2,601,759,944 | IssuesEvent | 2015-02-24 00:34:38 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | SC crash on BWAPI 3.4 | auto-migrated Component-Logic Priority-Critical Type-Defect Usability | ```
My bot which works under BWAPI 3.0.3 and 3.2 crashes immediately under BWAPI
3.4.
The map starts, but before any output is shown in the window SC crashes. The
global log shows the following:
10:09:55 BWAPI initialisation started
10:09:58 calling onGameStart
10:09:58 Loading AI DLL from: bwapi-data\AI\BTHAIModule.dll
10:09:58 Loaded AI Module
10:09:58 Importing by Virtual Function Table from AI DLL
10:09:58 Creating an Object of AIModule
10:09:58 Created an Object of AIModule
```
-----
Original issue reported on code.google.com by `johan.hagelback@gmail.com` on 5 Jan 2011 at 9:25 | 1.0 | SC crash on BWAPI 3.4 - ```
My bot which works under BWAPI 3.0.3 and 3.2 crashes immediately under BWAPI
3.4.
The map starts, but before any output is shown in the window SC crashes. The
global log shows the following:
10:09:55 BWAPI initialisation started
10:09:58 calling onGameStart
10:09:58 Loading AI DLL from: bwapi-data\AI\BTHAIModule.dll
10:09:58 Loaded AI Module
10:09:58 Importing by Virtual Function Table from AI DLL
10:09:58 Creating an Object of AIModule
10:09:58 Created an Object of AIModule
```
-----
Original issue reported on code.google.com by `johan.hagelback@gmail.com` on 5 Jan 2011 at 9:25 | defect | sc crash on bwapi my bot which works under bwapi and crashes immediately under bwapi the map starts but before any output is shown in the window sc crashes the global log shows the following bwapi initialisation started calling ongamestart loading ai dll from bwapi data ai bthaimodule dll loaded ai module importing by virtual function table from ai dll creating an object of aimodule created an object of aimodule original issue reported on code google com by johan hagelback gmail com on jan at | 1 |
196,298 | 22,441,282,010 | IssuesEvent | 2022-06-21 01:36:25 | rvvergara/bookstore-cms-next-app | https://api.github.com/repos/rvvergara/bookstore-cms-next-app | closed | CVE-2021-29060 (Medium) detected in color-string-1.5.3.tgz - autoclosed | security vulnerability | ## CVE-2021-29060 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>color-string-1.5.3.tgz</b></p></summary>
<p>Parser and generator for CSS color strings</p>
<p>Library home page: <a href="https://registry.npmjs.org/color-string/-/color-string-1.5.3.tgz">https://registry.npmjs.org/color-string/-/color-string-1.5.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/color-string/package.json</p>
<p>
Dependency Hierarchy:
- next-9.4.4.tgz (Root Library)
- toolbox-optimizer-2.4.0.tgz
- cssnano-4.1.10.tgz
- cssnano-preset-default-4.0.7.tgz
- postcss-colormin-4.0.3.tgz
- color-3.1.2.tgz
- :x: **color-string-1.5.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore-cms-next-app/commit/5614dac4aa122d53b811e7642095e49cf274e6ef">5614dac4aa122d53b811e7642095e49cf274e6ef</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular Expression Denial of Service (ReDOS) vulnerability was discovered in Color-String version 1.5.5 and below which occurs when the application is provided and checks a crafted invalid HWB string.
<p>Publish Date: 2021-06-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29060>CVE-2021-29060</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-257v-vj4p-3w2h">https://github.com/advisories/GHSA-257v-vj4p-3w2h</a></p>
<p>Release Date: 2021-06-21</p>
<p>Fix Resolution (color-string): 1.5.5</p>
<p>Direct dependency fix Resolution (next): 9.4.5-canary.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-29060 (Medium) detected in color-string-1.5.3.tgz - autoclosed - ## CVE-2021-29060 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>color-string-1.5.3.tgz</b></p></summary>
<p>Parser and generator for CSS color strings</p>
<p>Library home page: <a href="https://registry.npmjs.org/color-string/-/color-string-1.5.3.tgz">https://registry.npmjs.org/color-string/-/color-string-1.5.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/color-string/package.json</p>
<p>
Dependency Hierarchy:
- next-9.4.4.tgz (Root Library)
- toolbox-optimizer-2.4.0.tgz
- cssnano-4.1.10.tgz
- cssnano-preset-default-4.0.7.tgz
- postcss-colormin-4.0.3.tgz
- color-3.1.2.tgz
- :x: **color-string-1.5.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/bookstore-cms-next-app/commit/5614dac4aa122d53b811e7642095e49cf274e6ef">5614dac4aa122d53b811e7642095e49cf274e6ef</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular Expression Denial of Service (ReDOS) vulnerability was discovered in Color-String version 1.5.5 and below which occurs when the application is provided and checks a crafted invalid HWB string.
<p>Publish Date: 2021-06-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29060>CVE-2021-29060</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-257v-vj4p-3w2h">https://github.com/advisories/GHSA-257v-vj4p-3w2h</a></p>
<p>Release Date: 2021-06-21</p>
<p>Fix Resolution (color-string): 1.5.5</p>
<p>Direct dependency fix Resolution (next): 9.4.5-canary.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in color string tgz autoclosed cve medium severity vulnerability vulnerable library color string tgz parser and generator for css color strings library home page a href path to dependency file package json path to vulnerable library node modules color string package json dependency hierarchy next tgz root library toolbox optimizer tgz cssnano tgz cssnano preset default tgz postcss colormin tgz color tgz x color string tgz vulnerable library found in head commit a href vulnerability details a regular expression denial of service redos vulnerability was discovered in color string version and below which occurs when the application is provided and checks a crafted invalid hwb string publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution color string direct dependency fix resolution next canary step up your open source security game with whitesource | 0 |
66,374 | 27,431,369,790 | IssuesEvent | 2023-03-02 01:47:45 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | Managed HSM is returning 400 BadParameter for release policy version | KeyVault Service Attention Client | For several tests where 7.4-preview.1 support was added, anything with a release policy that does define a version (as they did before) is returning:
```
Status: 400 (Bad Request)
ErrorCode: BadParameter
Content:
{"error":{"code":"BadParameter","message":"Key Release Policy: missing 'version' property (Activity ID: 31b04ff2-5b0a-11ed-94f7-0022488df0c4)"}}
```
Re-running these tests on 7.3 works. For now, I'll add a `Max` version for live tests and have a thread going with the service team. | 1.0 | Managed HSM is returning 400 BadParameter for release policy version - For several tests where 7.4-preview.1 support was added, anything with a release policy that does define a version (as they did before) is returning:
```
Status: 400 (Bad Request)
ErrorCode: BadParameter
Content:
{"error":{"code":"BadParameter","message":"Key Release Policy: missing 'version' property (Activity ID: 31b04ff2-5b0a-11ed-94f7-0022488df0c4)"}}
```
Re-running these tests on 7.3 works. For now, I'll add a `Max` version for live tests and have a thread going with the service team. | non_defect | managed hsm is returning badparameter for release policy version for several tests where preview support was added anything with a release policy that does define a version as they did before is returning status bad request errorcode badparameter content error code badparameter message key release policy missing version property activity id re running these tests on works for now i ll add a max version for live tests and have a thread going with the service team | 0 |
38,327 | 8,759,428,412 | IssuesEvent | 2018-12-15 16:06:44 | jsslutils/jsslutils | https://api.github.com/repos/jsslutils/jsslutils | closed | The reload of CRL happens only one time instead repeat at a fixed interval when an reload interval is specified | Priority-Medium Type-Defect auto-migrated | ```
What steps will reproduce the problem?
1. Specify system property org.jsslutils.prop.crlReloadInterval to n value.
This is the number of seconds CRLs are reloaded
2. CRLs are reloaded after n seconds
3. There are no CRLs reloaded after that.
What is the expected output? What do you see instead?
Expected CRLs to reloaded every n seconds but CRLs reloaded only once
What version of the product are you using? On what operating system?
happens in JsslUtils 1.0.5 on both windows and linux.
Please provide any additional information below.
In PKIXSSLContextFactory class the call made in line 361
crlReloaderScheduledThreadPoolExecutor.schedule(reloader,
reloadInterval, TimeUnit.SECONDS);
should be
crlReloaderScheduledThreadPoolExecutor.scheduleAtFixedRate(...)
```
Original issue reported on code.google.com by `hnguyen6...@gmail.com` on 3 May 2011 at 3:03
| 1.0 | The reload of CRL happens only one time instead repeat at a fixed interval when an reload interval is specified - ```
What steps will reproduce the problem?
1. Specify system property org.jsslutils.prop.crlReloadInterval to n value.
This is the number of seconds CRLs are reloaded
2. CRLs are reloaded after n seconds
3. There are no CRLs reloaded after that.
What is the expected output? What do you see instead?
Expected CRLs to reloaded every n seconds but CRLs reloaded only once
What version of the product are you using? On what operating system?
happens in JsslUtils 1.0.5 on both windows and linux.
Please provide any additional information below.
In PKIXSSLContextFactory class the call made in line 361
crlReloaderScheduledThreadPoolExecutor.schedule(reloader,
reloadInterval, TimeUnit.SECONDS);
should be
crlReloaderScheduledThreadPoolExecutor.scheduleAtFixedRate(...)
```
Original issue reported on code.google.com by `hnguyen6...@gmail.com` on 3 May 2011 at 3:03
| defect | the reload of crl happens only one time instead repeat at a fixed interval when an reload interval is specified what steps will reproduce the problem specify system property org jsslutils prop crlreloadinterval to n value this is the number of seconds crls are reloaded crls are reloaded after n seconds there are no crls reloaded after that what is the expected output what do you see instead expected crls to reloaded every n seconds but crls reloaded only once what version of the product are you using on what operating system happens in jsslutils on both windows and linux please provide any additional information below in pkixsslcontextfactory class the call made in line crlreloaderscheduledthreadpoolexecutor schedule reloader reloadinterval timeunit seconds should be crlreloaderscheduledthreadpoolexecutor scheduleatfixedrate original issue reported on code google com by gmail com on may at | 1 |
144,937 | 19,318,934,958 | IssuesEvent | 2021-12-14 01:41:10 | txh51591/tm-repo | https://api.github.com/repos/txh51591/tm-repo | opened | CVE-2019-12814 (Medium) detected in jackson-databind-2.9.9.jar | security vulnerability | ## CVE-2019-12814 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: tm-repo/pom.xml</p>
<p>Path to vulnerable library: m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server.
<p>Publish Date: 2019-06-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12814>CVE-2019-12814</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2341">https://github.com/FasterXML/jackson-databind/issues/2341</a></p>
<p>Release Date: 2020-10-20</p>
<p>Fix Resolution: 2.7.9.6, 2.8.11.4, 2.9.9.1, 2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-12814 (Medium) detected in jackson-databind-2.9.9.jar - ## CVE-2019-12814 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: tm-repo/pom.xml</p>
<p>Path to vulnerable library: m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server.
<p>Publish Date: 2019-06-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12814>CVE-2019-12814</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2341">https://github.com/FasterXML/jackson-databind/issues/2341</a></p>
<p>Release Date: 2020-10-20</p>
<p>Fix Resolution: 2.7.9.6, 2.8.11.4, 2.9.9.1, 2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tm repo pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind x through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has jdom x or x jar in the classpath an attacker can send a specifically crafted json message that allows them to read arbitrary local files on the server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
814,440 | 30,507,862,131 | IssuesEvent | 2023-07-18 18:19:05 | GoogleCloudPlatform/professional-services-data-validator | https://api.github.com/repos/GoogleCloudPlatform/professional-services-data-validator | opened | Upgrade to pyproject.toml to configure setuptools | type: cleanup priority: p1 | The setup.py based install is being deprecated, and now pyproject.toml is the standard way of specifiying project metadata.
This issue will involve migrating to pyproject.toml file. | 1.0 | Upgrade to pyproject.toml to configure setuptools - The setup.py based install is being deprecated, and now pyproject.toml is the standard way of specifiying project metadata.
This issue will involve migrating to pyproject.toml file. | non_defect | upgrade to pyproject toml to configure setuptools the setup py based install is being deprecated and now pyproject toml is the standard way of specifiying project metadata this issue will involve migrating to pyproject toml file | 0 |
65,696 | 6,972,810,537 | IssuesEvent | 2017-12-11 18:16:44 | zcash/zcash | https://api.github.com/repos/zcash/zcash | closed | Test suite fail with two errors | portability testing | ### Describe the issue
Test suite fails.
### Can you reliably reproduce the issue?
#### If so, please list the steps to reproduce below:
1. Download ubuntu-16.04.2-desktop-amd64.iso
2. Run steps from https://github.com/zcash/zcash/wiki/1.0-User-Guide#configuration
3. Run Tests -> FAIL.
### Expected behaviour
Tests should pass.
### Actual behaviour + errors
Tests Fail.
### The version of Zcash you were using:
Zcash Daemon version v1.0.10-
### Machine specs:
- OS ubuntu-16.04.2-desktop-amd64
- CPU: amd64
- RAM: 2GiB
- Disk size: 30 GiB
- Disk Type (HD/SDD): HD
- Linux kernel version (uname -a): Linux good-VirtualBox 4.8.0-36-generic #36~16.04.1-Ubuntu SMP Sun Feb 5 09:39:57 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
- Compiler version (gcc -version):
gcc (Ubuntu 5.4.0-6ubuntu1~16.04.4) 5.4.0 20160609
### Any extra information that might be useful in the debugging process.
---
### Do you have a back up of `~/.zcash` directory and/or take a VM snapshot?
I have both, just ping me for a copy.
```
========================================
Zcash 1.0.10-1: src/test-suite.log
========================================
# TOTAL: 2
# PASS: 0
# SKIP: 0
# XFAIL: 0
# FAIL: 2
# XPASS: 0
# ERROR: 0
.. contents:: :depth: 2
FAIL: test/test_bitcoin
=======================
FAIL test/test_bitcoin (exit status: 137)
FAIL: zcash-gtest
=================
[==========] Running 128 tests from 25 test cases.
[----------] Global test environment set-up.
[----------] 24 tests from checktransaction_tests
[ RUN ] checktransaction_tests.check_vpub_not_both_nonzero
[ OK ] checktransaction_tests.check_vpub_not_both_nonzero (0 ms)
[ RUN ] checktransaction_tests.valid_transaction
[ OK ] checktransaction_tests.valid_transaction (1 ms)
[ RUN ] checktransaction_tests.BadVersionTooLow
[ OK ] checktransaction_tests.BadVersionTooLow (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vin_empty
[ OK ] checktransaction_tests.bad_txns_vin_empty (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vout_empty
[ OK ] checktransaction_tests.bad_txns_vout_empty (1 ms)
[ RUN ] checktransaction_tests.bad_txns_oversize
[ OK ] checktransaction_tests.bad_txns_oversize (3 ms)
[ RUN ] checktransaction_tests.bad_txns_vout_negative
[ OK ] checktransaction_tests.bad_txns_vout_negative (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vout_toolarge
[ OK ] checktransaction_tests.bad_txns_vout_toolarge (0 ms)
[ RUN ] checktransaction_tests.bad_txns_txouttotal_toolarge_outputs
[ OK ] checktransaction_tests.bad_txns_txouttotal_toolarge_outputs (1 ms)
[ RUN ] checktransaction_tests.bad_txns_txouttotal_toolarge_joinsplit
[ OK ] checktransaction_tests.bad_txns_txouttotal_toolarge_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_txns_txintotal_toolarge_joinsplit
[ OK ] checktransaction_tests.bad_txns_txintotal_toolarge_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_old_negative
[ OK ] checktransaction_tests.bad_txns_vpub_old_negative (1 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_new_negative
[ OK ] checktransaction_tests.bad_txns_vpub_new_negative (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_old_toolarge
[ OK ] checktransaction_tests.bad_txns_vpub_old_toolarge (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_new_toolarge
[ OK ] checktransaction_tests.bad_txns_vpub_new_toolarge (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpubs_both_nonzero
[ OK ] checktransaction_tests.bad_txns_vpubs_both_nonzero (1 ms)
[ RUN ] checktransaction_tests.bad_txns_inputs_duplicate
[ OK ] checktransaction_tests.bad_txns_inputs_duplicate (2 ms)
[ RUN ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_same_joinsplit
[ OK ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_same_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_different_joinsplit
[ OK ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_different_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_cb_has_joinsplits
[ OK ] checktransaction_tests.bad_cb_has_joinsplits (1 ms)
[ RUN ] checktransaction_tests.bad_cb_empty_scriptsig
[ OK ] checktransaction_tests.bad_cb_empty_scriptsig (0 ms)
[ RUN ] checktransaction_tests.bad_txns_prevout_null
[ OK ] checktransaction_tests.bad_txns_prevout_null (1 ms)
[ RUN ] checktransaction_tests.bad_txns_invalid_joinsplit_signature
[ OK ] checktransaction_tests.bad_txns_invalid_joinsplit_signature (0 ms)
[ RUN ] checktransaction_tests.non_canonical_ed25519_signature
[ OK ] checktransaction_tests.non_canonical_ed25519_signature (1 ms)
[----------] 24 tests from checktransaction_tests (13 ms total)
[----------] 7 tests from founders_reward_test
[ RUN ] founders_reward_test.general
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[ OK ] founders_reward_test.general (1026 ms)
[ RUN ] founders_reward_test.mainnet
[ OK ] founders_reward_test.mainnet (90 ms)
[ RUN ] founders_reward_test.testnet
[ OK ] founders_reward_test.testnet (86 ms)
[ RUN ] founders_reward_test.regtest
[ OK ] founders_reward_test.regtest (0 ms)
[ RUN ] founders_reward_test.slow_start_subsidy
[ OK ] founders_reward_test.slow_start_subsidy (6 ms)
[ RUN ] founders_reward_test.per_address_reward_mainnet
[ OK ] founders_reward_test.per_address_reward_mainnet (474 ms)
[ RUN ] founders_reward_test.per_address_reward_testnet
[ OK ] founders_reward_test.per_address_reward_testnet (513 ms)
[----------] 7 tests from founders_reward_test (2195 ms total)
[----------] 3 tests from wallet_zkeys_tests
[ RUN ] wallet_zkeys_tests.store_and_load_zkeys
[ OK ] wallet_zkeys_tests.store_and_load_zkeys (1 ms)
[ RUN ] wallet_zkeys_tests.write_zkey_direct_to_db
[ OK ] wallet_zkeys_tests.write_zkey_direct_to_db (59 ms)
[ RUN ] wallet_zkeys_tests.write_cryptedzkey_direct_to_db
[ OK ] wallet_zkeys_tests.write_cryptedzkey_direct_to_db (818 ms)
[----------] 3 tests from wallet_zkeys_tests (878 ms total)
[----------] 1 test from tautologies
[ RUN ] tautologies.seven_eq_seven
[ OK ] tautologies.seven_eq_seven (0 ms)
[----------] 1 test from tautologies (0 ms total)
[----------] 9 tests from DeprecationTest
[ RUN ] DeprecationTest.NonDeprecatedNodeKeepsRunning
[ OK ] DeprecationTest.NonDeprecatedNodeKeepsRunning (0 ms)
[ RUN ] DeprecationTest.NodeNearDeprecationIsWarned
[ OK ] DeprecationTest.NodeNearDeprecationIsWarned (1 ms)
[ RUN ] DeprecationTest.NodeNearDeprecationWarningIsNotDuplicated
[ OK ] DeprecationTest.NodeNearDeprecationWarningIsNotDuplicated (0 ms)
[ RUN ] DeprecationTest.NodeNearDeprecationWarningIsRepeatedOnStartup
[ OK ] DeprecationTest.NodeNearDeprecationWarningIsRepeatedOnStartup (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeShutsDown
[ OK ] DeprecationTest.DeprecatedNodeShutsDown (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeErrorIsNotDuplicated
[ OK ] DeprecationTest.DeprecatedNodeErrorIsNotDuplicated (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeErrorIsRepeatedOnStartup
[ OK ] DeprecationTest.DeprecatedNodeErrorIsRepeatedOnStartup (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeShutsDownIfOldVersionDisabled
[ OK ] DeprecationTest.DeprecatedNodeShutsDownIfOldVersionDisabled (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeKeepsRunningIfCurrentVersionDisabled
[ OK ] DeprecationTest.DeprecatedNodeKeepsRunningIfCurrentVersionDisabled (0 ms)
[----------] 9 tests from DeprecationTest (1 ms total)
[----------] 5 tests from equihash_tests
[ RUN ] equihash_tests.expand_and_contract_arrays
[ OK ] equihash_tests.expand_and_contract_arrays (0 ms)
[ RUN ] equihash_tests.minimal_solution_representation
[ OK ] equihash_tests.minimal_solution_representation (0 ms)
[ RUN ] equihash_tests.is_probably_duplicate
[ OK ] equihash_tests.is_probably_duplicate (0 ms)
[ RUN ] equihash_tests.check_basic_solver_cancelled
[ OK ] equihash_tests.check_basic_solver_cancelled (9 ms)
[ RUN ] equihash_tests.check_optimised_solver_cancelled
[ OK ] equihash_tests.check_optimised_solver_cancelled (4 ms)
[----------] 5 tests from equihash_tests (13 ms total)
[----------] 3 tests from joinsplit
[ RUN ] joinsplit.h_sig
[ OK ] joinsplit.h_sig (0 ms)
[ RUN ] joinsplit.full_api_test
FAIL zcash-gtest (exit status: 137)
```
| 1.0 | Test suite fail with two errors - ### Describe the issue
Test suite fails.
### Can you reliably reproduce the issue?
#### If so, please list the steps to reproduce below:
1. Download ubuntu-16.04.2-desktop-amd64.iso
2. Run steps from https://github.com/zcash/zcash/wiki/1.0-User-Guide#configuration
3. Run Tests -> FAIL.
### Expected behaviour
Tests should pass.
### Actual behaviour + errors
Tests Fail.
### The version of Zcash you were using:
Zcash Daemon version v1.0.10-
### Machine specs:
- OS ubuntu-16.04.2-desktop-amd64
- CPU: amd64
- RAM: 2GiB
- Disk size: 30 GiB
- Disk Type (HD/SDD): HD
- Linux kernel version (uname -a): Linux good-VirtualBox 4.8.0-36-generic #36~16.04.1-Ubuntu SMP Sun Feb 5 09:39:57 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
- Compiler version (gcc -version):
gcc (Ubuntu 5.4.0-6ubuntu1~16.04.4) 5.4.0 20160609
### Any extra information that might be useful in the debugging process.
---
### Do you have a back up of `~/.zcash` directory and/or take a VM snapshot?
I have both, just ping me for a copy.
```
========================================
Zcash 1.0.10-1: src/test-suite.log
========================================
# TOTAL: 2
# PASS: 0
# SKIP: 0
# XFAIL: 0
# FAIL: 2
# XPASS: 0
# ERROR: 0
.. contents:: :depth: 2
FAIL: test/test_bitcoin
=======================
FAIL test/test_bitcoin (exit status: 137)
FAIL: zcash-gtest
=================
[==========] Running 128 tests from 25 test cases.
[----------] Global test environment set-up.
[----------] 24 tests from checktransaction_tests
[ RUN ] checktransaction_tests.check_vpub_not_both_nonzero
[ OK ] checktransaction_tests.check_vpub_not_both_nonzero (0 ms)
[ RUN ] checktransaction_tests.valid_transaction
[ OK ] checktransaction_tests.valid_transaction (1 ms)
[ RUN ] checktransaction_tests.BadVersionTooLow
[ OK ] checktransaction_tests.BadVersionTooLow (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vin_empty
[ OK ] checktransaction_tests.bad_txns_vin_empty (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vout_empty
[ OK ] checktransaction_tests.bad_txns_vout_empty (1 ms)
[ RUN ] checktransaction_tests.bad_txns_oversize
[ OK ] checktransaction_tests.bad_txns_oversize (3 ms)
[ RUN ] checktransaction_tests.bad_txns_vout_negative
[ OK ] checktransaction_tests.bad_txns_vout_negative (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vout_toolarge
[ OK ] checktransaction_tests.bad_txns_vout_toolarge (0 ms)
[ RUN ] checktransaction_tests.bad_txns_txouttotal_toolarge_outputs
[ OK ] checktransaction_tests.bad_txns_txouttotal_toolarge_outputs (1 ms)
[ RUN ] checktransaction_tests.bad_txns_txouttotal_toolarge_joinsplit
[ OK ] checktransaction_tests.bad_txns_txouttotal_toolarge_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_txns_txintotal_toolarge_joinsplit
[ OK ] checktransaction_tests.bad_txns_txintotal_toolarge_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_old_negative
[ OK ] checktransaction_tests.bad_txns_vpub_old_negative (1 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_new_negative
[ OK ] checktransaction_tests.bad_txns_vpub_new_negative (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_old_toolarge
[ OK ] checktransaction_tests.bad_txns_vpub_old_toolarge (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpub_new_toolarge
[ OK ] checktransaction_tests.bad_txns_vpub_new_toolarge (0 ms)
[ RUN ] checktransaction_tests.bad_txns_vpubs_both_nonzero
[ OK ] checktransaction_tests.bad_txns_vpubs_both_nonzero (1 ms)
[ RUN ] checktransaction_tests.bad_txns_inputs_duplicate
[ OK ] checktransaction_tests.bad_txns_inputs_duplicate (2 ms)
[ RUN ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_same_joinsplit
[ OK ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_same_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_different_joinsplit
[ OK ] checktransaction_tests.bad_joinsplits_nullifiers_duplicate_different_joinsplit (0 ms)
[ RUN ] checktransaction_tests.bad_cb_has_joinsplits
[ OK ] checktransaction_tests.bad_cb_has_joinsplits (1 ms)
[ RUN ] checktransaction_tests.bad_cb_empty_scriptsig
[ OK ] checktransaction_tests.bad_cb_empty_scriptsig (0 ms)
[ RUN ] checktransaction_tests.bad_txns_prevout_null
[ OK ] checktransaction_tests.bad_txns_prevout_null (1 ms)
[ RUN ] checktransaction_tests.bad_txns_invalid_joinsplit_signature
[ OK ] checktransaction_tests.bad_txns_invalid_joinsplit_signature (0 ms)
[ RUN ] checktransaction_tests.non_canonical_ed25519_signature
[ OK ] checktransaction_tests.non_canonical_ed25519_signature (1 ms)
[----------] 24 tests from checktransaction_tests (13 ms total)
[----------] 7 tests from founders_reward_test
[ RUN ] founders_reward_test.general
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[WARNING] ../src/gtest-death-test.cc:825:: Death tests use fork(), which is unsafe particularly in a threaded context. For this test, Google Test couldn't detect the number of threads.
[ OK ] founders_reward_test.general (1026 ms)
[ RUN ] founders_reward_test.mainnet
[ OK ] founders_reward_test.mainnet (90 ms)
[ RUN ] founders_reward_test.testnet
[ OK ] founders_reward_test.testnet (86 ms)
[ RUN ] founders_reward_test.regtest
[ OK ] founders_reward_test.regtest (0 ms)
[ RUN ] founders_reward_test.slow_start_subsidy
[ OK ] founders_reward_test.slow_start_subsidy (6 ms)
[ RUN ] founders_reward_test.per_address_reward_mainnet
[ OK ] founders_reward_test.per_address_reward_mainnet (474 ms)
[ RUN ] founders_reward_test.per_address_reward_testnet
[ OK ] founders_reward_test.per_address_reward_testnet (513 ms)
[----------] 7 tests from founders_reward_test (2195 ms total)
[----------] 3 tests from wallet_zkeys_tests
[ RUN ] wallet_zkeys_tests.store_and_load_zkeys
[ OK ] wallet_zkeys_tests.store_and_load_zkeys (1 ms)
[ RUN ] wallet_zkeys_tests.write_zkey_direct_to_db
[ OK ] wallet_zkeys_tests.write_zkey_direct_to_db (59 ms)
[ RUN ] wallet_zkeys_tests.write_cryptedzkey_direct_to_db
[ OK ] wallet_zkeys_tests.write_cryptedzkey_direct_to_db (818 ms)
[----------] 3 tests from wallet_zkeys_tests (878 ms total)
[----------] 1 test from tautologies
[ RUN ] tautologies.seven_eq_seven
[ OK ] tautologies.seven_eq_seven (0 ms)
[----------] 1 test from tautologies (0 ms total)
[----------] 9 tests from DeprecationTest
[ RUN ] DeprecationTest.NonDeprecatedNodeKeepsRunning
[ OK ] DeprecationTest.NonDeprecatedNodeKeepsRunning (0 ms)
[ RUN ] DeprecationTest.NodeNearDeprecationIsWarned
[ OK ] DeprecationTest.NodeNearDeprecationIsWarned (1 ms)
[ RUN ] DeprecationTest.NodeNearDeprecationWarningIsNotDuplicated
[ OK ] DeprecationTest.NodeNearDeprecationWarningIsNotDuplicated (0 ms)
[ RUN ] DeprecationTest.NodeNearDeprecationWarningIsRepeatedOnStartup
[ OK ] DeprecationTest.NodeNearDeprecationWarningIsRepeatedOnStartup (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeShutsDown
[ OK ] DeprecationTest.DeprecatedNodeShutsDown (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeErrorIsNotDuplicated
[ OK ] DeprecationTest.DeprecatedNodeErrorIsNotDuplicated (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeErrorIsRepeatedOnStartup
[ OK ] DeprecationTest.DeprecatedNodeErrorIsRepeatedOnStartup (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeShutsDownIfOldVersionDisabled
[ OK ] DeprecationTest.DeprecatedNodeShutsDownIfOldVersionDisabled (0 ms)
[ RUN ] DeprecationTest.DeprecatedNodeKeepsRunningIfCurrentVersionDisabled
[ OK ] DeprecationTest.DeprecatedNodeKeepsRunningIfCurrentVersionDisabled (0 ms)
[----------] 9 tests from DeprecationTest (1 ms total)
[----------] 5 tests from equihash_tests
[ RUN ] equihash_tests.expand_and_contract_arrays
[ OK ] equihash_tests.expand_and_contract_arrays (0 ms)
[ RUN ] equihash_tests.minimal_solution_representation
[ OK ] equihash_tests.minimal_solution_representation (0 ms)
[ RUN ] equihash_tests.is_probably_duplicate
[ OK ] equihash_tests.is_probably_duplicate (0 ms)
[ RUN ] equihash_tests.check_basic_solver_cancelled
[ OK ] equihash_tests.check_basic_solver_cancelled (9 ms)
[ RUN ] equihash_tests.check_optimised_solver_cancelled
[ OK ] equihash_tests.check_optimised_solver_cancelled (4 ms)
[----------] 5 tests from equihash_tests (13 ms total)
[----------] 3 tests from joinsplit
[ RUN ] joinsplit.h_sig
[ OK ] joinsplit.h_sig (0 ms)
[ RUN ] joinsplit.full_api_test
FAIL zcash-gtest (exit status: 137)
```
| non_defect | test suite fail with two errors describe the issue test suite fails can you reliably reproduce the issue if so please list the steps to reproduce below download ubuntu desktop iso run steps from run tests fail expected behaviour tests should pass actual behaviour errors tests fail the version of zcash you were using zcash daemon version machine specs os ubuntu desktop cpu ram disk size gib disk type hd sdd hd linux kernel version uname a linux good virtualbox generic ubuntu smp sun feb utc gnu linux compiler version gcc version gcc ubuntu any extra information that might be useful in the debugging process do you have a back up of zcash directory and or take a vm snapshot i have both just ping me for a copy zcash src test suite log total pass skip xfail fail xpass error contents depth fail test test bitcoin fail test test bitcoin exit status fail zcash gtest running tests from test cases global test environment set up tests from checktransaction tests checktransaction tests check vpub not both nonzero checktransaction tests check vpub not both nonzero ms checktransaction tests valid transaction checktransaction tests valid transaction ms checktransaction tests badversiontoolow checktransaction tests badversiontoolow ms checktransaction tests bad txns vin empty checktransaction tests bad txns vin empty ms checktransaction tests bad txns vout empty checktransaction tests bad txns vout empty ms checktransaction tests bad txns oversize checktransaction tests bad txns oversize ms checktransaction tests bad txns vout negative checktransaction tests bad txns vout negative ms checktransaction tests bad txns vout toolarge checktransaction tests bad txns vout toolarge ms checktransaction tests bad txns txouttotal toolarge outputs checktransaction tests bad txns txouttotal toolarge outputs ms checktransaction tests bad txns txouttotal toolarge joinsplit checktransaction tests bad txns txouttotal toolarge joinsplit ms checktransaction tests bad txns txintotal toolarge joinsplit checktransaction tests bad txns txintotal toolarge joinsplit ms checktransaction tests bad txns vpub old negative checktransaction tests bad txns vpub old negative ms checktransaction tests bad txns vpub new negative checktransaction tests bad txns vpub new negative ms checktransaction tests bad txns vpub old toolarge checktransaction tests bad txns vpub old toolarge ms checktransaction tests bad txns vpub new toolarge checktransaction tests bad txns vpub new toolarge ms checktransaction tests bad txns vpubs both nonzero checktransaction tests bad txns vpubs both nonzero ms checktransaction tests bad txns inputs duplicate checktransaction tests bad txns inputs duplicate ms checktransaction tests bad joinsplits nullifiers duplicate same joinsplit checktransaction tests bad joinsplits nullifiers duplicate same joinsplit ms checktransaction tests bad joinsplits nullifiers duplicate different joinsplit checktransaction tests bad joinsplits nullifiers duplicate different joinsplit ms checktransaction tests bad cb has joinsplits checktransaction tests bad cb has joinsplits ms checktransaction tests bad cb empty scriptsig checktransaction tests bad cb empty scriptsig ms checktransaction tests bad txns prevout null checktransaction tests bad txns prevout null ms checktransaction tests bad txns invalid joinsplit signature checktransaction tests bad txns invalid joinsplit signature ms checktransaction tests non canonical signature checktransaction tests non canonical signature ms tests from checktransaction tests ms total tests from founders reward test founders reward test general src gtest death test cc death tests use fork which is unsafe particularly in a threaded context for this test google test couldn t detect the number of threads src gtest death test cc death tests use fork which is unsafe particularly in a threaded context for this test google test couldn t detect the number of threads src gtest death test cc death tests use fork which is unsafe particularly in a threaded context for this test google test couldn t detect the number of threads src gtest death test cc death tests use fork which is unsafe particularly in a threaded context for this test google test couldn t detect the number of threads founders reward test general ms founders reward test mainnet founders reward test mainnet ms founders reward test testnet founders reward test testnet ms founders reward test regtest founders reward test regtest ms founders reward test slow start subsidy founders reward test slow start subsidy ms founders reward test per address reward mainnet founders reward test per address reward mainnet ms founders reward test per address reward testnet founders reward test per address reward testnet ms tests from founders reward test ms total tests from wallet zkeys tests wallet zkeys tests store and load zkeys wallet zkeys tests store and load zkeys ms wallet zkeys tests write zkey direct to db wallet zkeys tests write zkey direct to db ms wallet zkeys tests write cryptedzkey direct to db wallet zkeys tests write cryptedzkey direct to db ms tests from wallet zkeys tests ms total test from tautologies tautologies seven eq seven tautologies seven eq seven ms test from tautologies ms total tests from deprecationtest deprecationtest nondeprecatednodekeepsrunning deprecationtest nondeprecatednodekeepsrunning ms deprecationtest nodeneardeprecationiswarned deprecationtest nodeneardeprecationiswarned ms deprecationtest nodeneardeprecationwarningisnotduplicated deprecationtest nodeneardeprecationwarningisnotduplicated ms deprecationtest nodeneardeprecationwarningisrepeatedonstartup deprecationtest nodeneardeprecationwarningisrepeatedonstartup ms deprecationtest deprecatednodeshutsdown deprecationtest deprecatednodeshutsdown ms deprecationtest deprecatednodeerrorisnotduplicated deprecationtest deprecatednodeerrorisnotduplicated ms deprecationtest deprecatednodeerrorisrepeatedonstartup deprecationtest deprecatednodeerrorisrepeatedonstartup ms deprecationtest deprecatednodeshutsdownifoldversiondisabled deprecationtest deprecatednodeshutsdownifoldversiondisabled ms deprecationtest deprecatednodekeepsrunningifcurrentversiondisabled deprecationtest deprecatednodekeepsrunningifcurrentversiondisabled ms tests from deprecationtest ms total tests from equihash tests equihash tests expand and contract arrays equihash tests expand and contract arrays ms equihash tests minimal solution representation equihash tests minimal solution representation ms equihash tests is probably duplicate equihash tests is probably duplicate ms equihash tests check basic solver cancelled equihash tests check basic solver cancelled ms equihash tests check optimised solver cancelled equihash tests check optimised solver cancelled ms tests from equihash tests ms total tests from joinsplit joinsplit h sig joinsplit h sig ms joinsplit full api test fail zcash gtest exit status | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.