Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
55,881 | 3,075,100,780 | IssuesEvent | 2015-08-20 11:39:54 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Opera не перехватывает ссылки | bug imported Priority-Medium Usability | _From [stupidcu...@gmail.com](https://code.google.com/u/109602399200609281346/) on February 16, 2011 13:31:56_
What steps will reproduce the problem? В браузере Opera при нажатии ссылки пишет отсутствует связанное приложение для данного расширение, хотя в флайлинке стоит галочка перехватывать ссылки What version of the product are you using? On what operating system? FlylinkDC-x64- r500 -beta74-build-6047 Windows 7 x64
**Attachment:** [bug.jpg](http://code.google.com/p/flylinkdc/issues/detail?id=362)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=362_ | 1.0 | Opera не перехватывает ссылки - _From [stupidcu...@gmail.com](https://code.google.com/u/109602399200609281346/) on February 16, 2011 13:31:56_
What steps will reproduce the problem? В браузере Opera при нажатии ссылки пишет отсутствует связанное приложение для данного расширение, хотя в флайлинке стоит галочка перехватывать ссылки What version of the product are you using? On what operating system? FlylinkDC-x64- r500 -beta74-build-6047 Windows 7 x64
**Attachment:** [bug.jpg](http://code.google.com/p/flylinkdc/issues/detail?id=362)
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=362_ | non_defect | opera не перехватывает ссылки from on february what steps will reproduce the problem в браузере opera при нажатии ссылки пишет отсутствует связанное приложение для данного расширение хотя в флайлинке стоит галочка перехватывать ссылки what version of the product are you using on what operating system flylinkdc build windows attachment original issue | 0 |
22,483 | 4,807,153,139 | IssuesEvent | 2016-11-02 20:36:53 | ipfs/go-ipfs | https://api.github.com/repos/ipfs/go-ipfs | closed | Inconsistent usage of IPFS and ipfs | documentation ready | We generally agreed on fact that capitalised IPFS should be used in reference to project and ipfs in reference to CLI or implementation.
Unfortunately now we are inconsistent across codebase. One of great examples is help text of `ipfs tar`:
```
ipfs tar -h
USAGE
ipfs tar - Utility functions for tar files in IPFS.
SUBCOMMANDS
ipfs tar add <file> - Import a tar file into ipfs.
ipfs tar cat <path> - Export a tar file from IPFS.
Use 'ipfs tar --help' for more information about this command.
```
| 1.0 | Inconsistent usage of IPFS and ipfs - We generally agreed on fact that capitalised IPFS should be used in reference to project and ipfs in reference to CLI or implementation.
Unfortunately now we are inconsistent across codebase. One of great examples is help text of `ipfs tar`:
```
ipfs tar -h
USAGE
ipfs tar - Utility functions for tar files in IPFS.
SUBCOMMANDS
ipfs tar add <file> - Import a tar file into ipfs.
ipfs tar cat <path> - Export a tar file from IPFS.
Use 'ipfs tar --help' for more information about this command.
```
| non_defect | inconsistent usage of ipfs and ipfs we generally agreed on fact that capitalised ipfs should be used in reference to project and ipfs in reference to cli or implementation unfortunately now we are inconsistent across codebase one of great examples is help text of ipfs tar ipfs tar h usage ipfs tar utility functions for tar files in ipfs subcommands ipfs tar add import a tar file into ipfs ipfs tar cat export a tar file from ipfs use ipfs tar help for more information about this command | 0 |
29,389 | 5,664,009,102 | IssuesEvent | 2017-04-11 00:26:12 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | MDAnalysisTests raises exception when imported | defect testing | ### Expected behaviour
```python
import MDAnalysisTests
```
and
```python
import MDAnalysis.tests
```
should import (and make eg files available in `datafiles`.
### Actual behaviour
Both imports fail with
```
In [3]: import MDAnalysisTests
---------------------------------------------------------------------------
Exception Traceback (most recent call last)
<ipython-input-3-52912a4f0416> in <module>()
----> 1 import MDAnalysisTests
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/__init__.py in <module>()
141 pass
142
--> 143 from MDAnalysisTests.util import (
144 block_import,
145 executable_not_found,
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/util.py in <module>()
37 from functools import wraps
38 import importlib
---> 39 import mock
40 import os
41
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/__init__.py in <module>()
1 from __future__ import absolute_import
----> 2 import mock.mock as _mock
3 from mock.mock import *
4 __all__ = _mock.__all__
5 #import mock.mock as _mock
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/mock.py in <module>()
69 from pbr.version import VersionInfo
70
---> 71 _v = VersionInfo('mock').semantic_version()
72 __version__ = _v.release_string()
73 version_info = _v.version_tuple()
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in semantic_version(self)
458 """Return the SemanticVersion object for this version."""
459 if self._semantic is None:
--> 460 self._semantic = self._get_version_from_pkg_resources()
461 return self._semantic
462
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in _get_version_from_pkg_resources(self)
445 # installed into anything. Revert to setup-time logic.
446 from pbr import packaging
--> 447 result_string = packaging.get_version(self.package)
448 return SemanticVersion.from_pip_string(result_string)
449
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/packaging.pyc in get_version(package_name, pre_version)
748 " to pbr.version.VersionInfo. Project name {name} was"
749 " given, but was not able to be found.".format(
--> 750 name=package_name))
751
752
Exception: Versioning for this project requires either an sdist tarball, or access to an upstream git repository. It's also possible that there is a mismatch between the package name in setup.cfg and the argument given to pbr.version.VersionInfo. Project name mock was given, but was not able to be found.
```
### Code to reproduce the behaviour
```python
import MDAnalysisTests
```
### Currently version of MDAnalysis:
(run `python -c "import MDAnalysis as mda; print(mda.__version__)"`)
0.16.0 (pip upgraded in a virtualenv) | 1.0 | MDAnalysisTests raises exception when imported - ### Expected behaviour
```python
import MDAnalysisTests
```
and
```python
import MDAnalysis.tests
```
should import (and make eg files available in `datafiles`.
### Actual behaviour
Both imports fail with
```
In [3]: import MDAnalysisTests
---------------------------------------------------------------------------
Exception Traceback (most recent call last)
<ipython-input-3-52912a4f0416> in <module>()
----> 1 import MDAnalysisTests
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/__init__.py in <module>()
141 pass
142
--> 143 from MDAnalysisTests.util import (
144 block_import,
145 executable_not_found,
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/util.py in <module>()
37 from functools import wraps
38 import importlib
---> 39 import mock
40 import os
41
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/__init__.py in <module>()
1 from __future__ import absolute_import
----> 2 import mock.mock as _mock
3 from mock.mock import *
4 __all__ = _mock.__all__
5 #import mock.mock as _mock
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/mock.py in <module>()
69 from pbr.version import VersionInfo
70
---> 71 _v = VersionInfo('mock').semantic_version()
72 __version__ = _v.release_string()
73 version_info = _v.version_tuple()
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in semantic_version(self)
458 """Return the SemanticVersion object for this version."""
459 if self._semantic is None:
--> 460 self._semantic = self._get_version_from_pkg_resources()
461 return self._semantic
462
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in _get_version_from_pkg_resources(self)
445 # installed into anything. Revert to setup-time logic.
446 from pbr import packaging
--> 447 result_string = packaging.get_version(self.package)
448 return SemanticVersion.from_pip_string(result_string)
449
/Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/packaging.pyc in get_version(package_name, pre_version)
748 " to pbr.version.VersionInfo. Project name {name} was"
749 " given, but was not able to be found.".format(
--> 750 name=package_name))
751
752
Exception: Versioning for this project requires either an sdist tarball, or access to an upstream git repository. It's also possible that there is a mismatch between the package name in setup.cfg and the argument given to pbr.version.VersionInfo. Project name mock was given, but was not able to be found.
```
### Code to reproduce the behaviour
```python
import MDAnalysisTests
```
### Currently version of MDAnalysis:
(run `python -c "import MDAnalysis as mda; print(mda.__version__)"`)
0.16.0 (pip upgraded in a virtualenv) | defect | mdanalysistests raises exception when imported expected behaviour python import mdanalysistests and python import mdanalysis tests should import and make eg files available in datafiles actual behaviour both imports fail with in import mdanalysistests exception traceback most recent call last in import mdanalysistests users oliver virtualenvs mda clean lib site packages mdanalysistests init py in pass from mdanalysistests util import block import executable not found users oliver virtualenvs mda clean lib site packages mdanalysistests util py in from functools import wraps import importlib import mock import os users oliver virtualenvs mda clean lib site packages mock init py in from future import absolute import import mock mock as mock from mock mock import all mock all import mock mock as mock users oliver virtualenvs mda clean lib site packages mock mock py in from pbr version import versioninfo v versioninfo mock semantic version version v release string version info v version tuple users oliver virtualenvs mda clean lib site packages pbr version pyc in semantic version self return the semanticversion object for this version if self semantic is none self semantic self get version from pkg resources return self semantic users oliver virtualenvs mda clean lib site packages pbr version pyc in get version from pkg resources self installed into anything revert to setup time logic from pbr import packaging result string packaging get version self package return semanticversion from pip string result string users oliver virtualenvs mda clean lib site packages pbr packaging pyc in get version package name pre version to pbr version versioninfo project name name was given but was not able to be found format name package name exception versioning for this project requires either an sdist tarball or access to an upstream git repository it s also possible that there is a mismatch between the package name in setup cfg and the argument given to pbr version versioninfo project name mock was given but was not able to be found code to reproduce the behaviour python import mdanalysistests currently version of mdanalysis run python c import mdanalysis as mda print mda version pip upgraded in a virtualenv | 1 |
59,999 | 17,023,307,846 | IssuesEvent | 2021-07-03 01:20:56 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | highway=motorway_junction nodes render with an incorrect font - t@h | Component: osmarender Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 8.35am, Thursday, 9th October 2008]**
Some stray t@h clients are rendering these nodes with an incorrect font. Some examples of this can be found at http://wiki.openstreetmap.org/index.php?title=User:Chriscf/bad_renders
In most of the cases, it can be seen that only highway=motorway_junction nodes are affected. It would seem that this one style, and no others, specifies "sans-serif" as a fallback font. It is unclear *why* some clients are falling back in this case, but it is causing undesirable results, especially on those captions that cross tile boundaries.
In the meantime, proposing that sans-serif be removed from this one feature, since there doesn't appear to be a reason why this feature and this one alone is marked out in this way. | 1.0 | highway=motorway_junction nodes render with an incorrect font - t@h - **[Submitted to the original trac issue database at 8.35am, Thursday, 9th October 2008]**
Some stray t@h clients are rendering these nodes with an incorrect font. Some examples of this can be found at http://wiki.openstreetmap.org/index.php?title=User:Chriscf/bad_renders
In most of the cases, it can be seen that only highway=motorway_junction nodes are affected. It would seem that this one style, and no others, specifies "sans-serif" as a fallback font. It is unclear *why* some clients are falling back in this case, but it is causing undesirable results, especially on those captions that cross tile boundaries.
In the meantime, proposing that sans-serif be removed from this one feature, since there doesn't appear to be a reason why this feature and this one alone is marked out in this way. | defect | highway motorway junction nodes render with an incorrect font t h some stray t h clients are rendering these nodes with an incorrect font some examples of this can be found at in most of the cases it can be seen that only highway motorway junction nodes are affected it would seem that this one style and no others specifies sans serif as a fallback font it is unclear why some clients are falling back in this case but it is causing undesirable results especially on those captions that cross tile boundaries in the meantime proposing that sans serif be removed from this one feature since there doesn t appear to be a reason why this feature and this one alone is marked out in this way | 1 |
594,594 | 18,049,158,666 | IssuesEvent | 2021-09-19 12:40:45 | codeforcroatia/imamopravoznati-theme | https://api.github.com/repos/codeforcroatia/imamopravoznati-theme | closed | Update donate URL | tasks ⚡ priority | **Is your feature request related to a problem? Please describe.**
<!--
A clear and concise description of what the problem is; e.g. I'm always frustrated when [...]
-->
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen.
-->
New URL (new WP web): https://gong.hr/tko-smo/podrzi-gong/
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
| 1.0 | Update donate URL - **Is your feature request related to a problem? Please describe.**
<!--
A clear and concise description of what the problem is; e.g. I'm always frustrated when [...]
-->
**Describe the solution you'd like**
<!--
A clear and concise description of what you want to happen.
-->
New URL (new WP web): https://gong.hr/tko-smo/podrzi-gong/
**Describe alternatives you've considered**
<!--
A clear and concise description of any alternative solutions or features you've considered.
-->
**Additional context**
<!--
Add any other context or screenshots about the feature request here.
-->
| non_defect | update donate url is your feature request related to a problem please describe a clear and concise description of what the problem is e g i m always frustrated when describe the solution you d like a clear and concise description of what you want to happen new url new wp web describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 0 |
203,022 | 15,340,155,633 | IssuesEvent | 2021-02-27 05:32:59 | momentum-mod/game | https://api.github.com/repos/momentum-mod/game | closed | Rendering issues when alt-tabbing in fullscreen | Blocked: Needs testing & verification Type: Bug | **Describe the bug**
ALt-tabbing in momentum on chaos causes the game to flicker and/or freeze for roughly half a second, Happens anywhere in the game. Under some circumstances it can show the setting panel for half a second even if it was not open prior to alt-tabbing.
**To Reproduce**
1. set game to fullscreen
2. alt-tab out and in, notice flickering/freezes (freezes seem to happen more often when in map compared to main menu)
**Expected behavior**
No visual corruption.
**Desktop/Branch (please complete the following information):**
- OS: Windows
- Branch: 0.8.8-rc
**Additional context**
Does not happened in windowed mode.
See if `-d3d9ex` changes anything since I know that affects alt-tabbing in other source games to some degree.
| 1.0 | Rendering issues when alt-tabbing in fullscreen - **Describe the bug**
ALt-tabbing in momentum on chaos causes the game to flicker and/or freeze for roughly half a second, Happens anywhere in the game. Under some circumstances it can show the setting panel for half a second even if it was not open prior to alt-tabbing.
**To Reproduce**
1. set game to fullscreen
2. alt-tab out and in, notice flickering/freezes (freezes seem to happen more often when in map compared to main menu)
**Expected behavior**
No visual corruption.
**Desktop/Branch (please complete the following information):**
- OS: Windows
- Branch: 0.8.8-rc
**Additional context**
Does not happened in windowed mode.
See if `-d3d9ex` changes anything since I know that affects alt-tabbing in other source games to some degree.
| non_defect | rendering issues when alt tabbing in fullscreen describe the bug alt tabbing in momentum on chaos causes the game to flicker and or freeze for roughly half a second happens anywhere in the game under some circumstances it can show the setting panel for half a second even if it was not open prior to alt tabbing to reproduce set game to fullscreen alt tab out and in notice flickering freezes freezes seem to happen more often when in map compared to main menu expected behavior no visual corruption desktop branch please complete the following information os windows branch rc additional context does not happened in windowed mode see if changes anything since i know that affects alt tabbing in other source games to some degree | 0 |
353,370 | 10,552,050,208 | IssuesEvent | 2019-10-03 14:28:40 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | pass on ?to url to edit account profile page | component: firefox accounts priority: p3 state: pull request ready | Needed for mozilla/addons-frontend#8666
When a new user is redirected to the edit account profile page on frontend, we should pass along the intended destination url as a GET parameter in the url.
Note: currently not all login/registration requests from frontend result in a new user being redirected to the edit page - #12417 will change that but we can't do that until frontend supports the `?to=xxx` param. | 1.0 | pass on ?to url to edit account profile page - Needed for mozilla/addons-frontend#8666
When a new user is redirected to the edit account profile page on frontend, we should pass along the intended destination url as a GET parameter in the url.
Note: currently not all login/registration requests from frontend result in a new user being redirected to the edit page - #12417 will change that but we can't do that until frontend supports the `?to=xxx` param. | non_defect | pass on to url to edit account profile page needed for mozilla addons frontend when a new user is redirected to the edit account profile page on frontend we should pass along the intended destination url as a get parameter in the url note currently not all login registration requests from frontend result in a new user being redirected to the edit page will change that but we can t do that until frontend supports the to xxx param | 0 |
44,091 | 11,962,106,747 | IssuesEvent | 2020-04-05 11:06:50 | gitblit/gitblit | https://api.github.com/repos/gitblit/gitblit | closed | Manifest Change in 1.9.0 Requires Explicit Classpath | Catg-Defect Status-Fixed | First of all thank you very much for the development of Gitblit, it's a really nice software :)
I was just helping a colleage to update Gitblit GO on a FreeBSD server from 1.8.0 to 1.9.0. We ran into problems regarding a change in `MANIFEST.MF` in `gitblit.jar`. Since the issues are Java-related, they most likely apply for all operating systems.
The manifest from 1.8.0 looks like this:
```
Manifest-Version: 1.0
Ant-Version: Apache Ant 1.9.3
Created-By: Moxie v0.9.4
Build-Jdk: 1.8.0_91
Build-Date: 2016-06-22
Implementation-Title: Gitblit
Implementation-Vendor-Id: com.gitblit
Implementation-Vendor-URL: http://gitblit.com
Implementation-Version: 1.8.0
Bundle-Name: Gitblit
Bundle-SymbolicName: gitblit
Bundle-Version: 1.8.0
Maven-Url: http://gitblit.github.io/gitblit-maven
Main-Class: org.moxie.MxLauncher
mxMain-Class: com.gitblit.GitBlitServer
mxMain-Paths: ext
```
and the manifest from 1.9.0 looks like this:
```
Manifest-Version: 1.0
Ant-Version: Apache Ant 1.9.2
Created-By: Moxie v0.9.4
Build-Jdk: 1.7.0_252
Build-Date: 2020-02-01
Implementation-Title: Gitblit
Implementation-Vendor-Id: com.gitblit
Implementation-Vendor-URL: http://gitblit.com
Implementation-Version: 1.9.0
Bundle-Name: Gitblit
Bundle-SymbolicName: gitblit
Bundle-Version: 1.9.0
Maven-Url: http://gitblit.github.io/gitblit-maven
Main-Class: com.gitblit.GitBlitServer
```
The most notable changes are:
- Main class changed from `org.moxie.MxLauncher` to `com.gitblit.GitblitServer`
- `mxMain-Class: com.gitblit.GitBlitServer` and `mxMain-Paths: ext` was removed
When using 1.8.0, it was possible to start gitblit using a command line like this:
```
java -jar gitblit.jar --baseFolder data
```
This was possible because the Moxie launcher automatically added the jars in the `ext` directory to the classpath. Now that this was removed, we end up with `NoClassDefFoundErrors`.
To resolve this, we have to specify the classpath explicitly **and** replace `-jar gitblit.jar` with the main class:
```
java -cp "gitblit.jar:ext/*" com.gitblit.GitBlitServer --baseFolder data
```
Note: you **have to add the quotes around the classpath**, otherwise it does not work.
Is this behavior intended? If yes, then several parts of the documentation should be updated. Also, the service scripts do not work, because they use the `-jar` version of the command. If the behavior is not intended, `ext` should be added to the classpath of gitblit.jar in the manifest.
| 1.0 | Manifest Change in 1.9.0 Requires Explicit Classpath - First of all thank you very much for the development of Gitblit, it's a really nice software :)
I was just helping a colleage to update Gitblit GO on a FreeBSD server from 1.8.0 to 1.9.0. We ran into problems regarding a change in `MANIFEST.MF` in `gitblit.jar`. Since the issues are Java-related, they most likely apply for all operating systems.
The manifest from 1.8.0 looks like this:
```
Manifest-Version: 1.0
Ant-Version: Apache Ant 1.9.3
Created-By: Moxie v0.9.4
Build-Jdk: 1.8.0_91
Build-Date: 2016-06-22
Implementation-Title: Gitblit
Implementation-Vendor-Id: com.gitblit
Implementation-Vendor-URL: http://gitblit.com
Implementation-Version: 1.8.0
Bundle-Name: Gitblit
Bundle-SymbolicName: gitblit
Bundle-Version: 1.8.0
Maven-Url: http://gitblit.github.io/gitblit-maven
Main-Class: org.moxie.MxLauncher
mxMain-Class: com.gitblit.GitBlitServer
mxMain-Paths: ext
```
and the manifest from 1.9.0 looks like this:
```
Manifest-Version: 1.0
Ant-Version: Apache Ant 1.9.2
Created-By: Moxie v0.9.4
Build-Jdk: 1.7.0_252
Build-Date: 2020-02-01
Implementation-Title: Gitblit
Implementation-Vendor-Id: com.gitblit
Implementation-Vendor-URL: http://gitblit.com
Implementation-Version: 1.9.0
Bundle-Name: Gitblit
Bundle-SymbolicName: gitblit
Bundle-Version: 1.9.0
Maven-Url: http://gitblit.github.io/gitblit-maven
Main-Class: com.gitblit.GitBlitServer
```
The most notable changes are:
- Main class changed from `org.moxie.MxLauncher` to `com.gitblit.GitblitServer`
- `mxMain-Class: com.gitblit.GitBlitServer` and `mxMain-Paths: ext` was removed
When using 1.8.0, it was possible to start gitblit using a command line like this:
```
java -jar gitblit.jar --baseFolder data
```
This was possible because the Moxie launcher automatically added the jars in the `ext` directory to the classpath. Now that this was removed, we end up with `NoClassDefFoundErrors`.
To resolve this, we have to specify the classpath explicitly **and** replace `-jar gitblit.jar` with the main class:
```
java -cp "gitblit.jar:ext/*" com.gitblit.GitBlitServer --baseFolder data
```
Note: you **have to add the quotes around the classpath**, otherwise it does not work.
Is this behavior intended? If yes, then several parts of the documentation should be updated. Also, the service scripts do not work, because they use the `-jar` version of the command. If the behavior is not intended, `ext` should be added to the classpath of gitblit.jar in the manifest.
| defect | manifest change in requires explicit classpath first of all thank you very much for the development of gitblit it s a really nice software i was just helping a colleage to update gitblit go on a freebsd server from to we ran into problems regarding a change in manifest mf in gitblit jar since the issues are java related they most likely apply for all operating systems the manifest from looks like this manifest version ant version apache ant created by moxie build jdk build date implementation title gitblit implementation vendor id com gitblit implementation vendor url implementation version bundle name gitblit bundle symbolicname gitblit bundle version maven url main class org moxie mxlauncher mxmain class com gitblit gitblitserver mxmain paths ext and the manifest from looks like this manifest version ant version apache ant created by moxie build jdk build date implementation title gitblit implementation vendor id com gitblit implementation vendor url implementation version bundle name gitblit bundle symbolicname gitblit bundle version maven url main class com gitblit gitblitserver the most notable changes are main class changed from org moxie mxlauncher to com gitblit gitblitserver mxmain class com gitblit gitblitserver and mxmain paths ext was removed when using it was possible to start gitblit using a command line like this java jar gitblit jar basefolder data this was possible because the moxie launcher automatically added the jars in the ext directory to the classpath now that this was removed we end up with noclassdeffounderrors to resolve this we have to specify the classpath explicitly and replace jar gitblit jar with the main class java cp gitblit jar ext com gitblit gitblitserver basefolder data note you have to add the quotes around the classpath otherwise it does not work is this behavior intended if yes then several parts of the documentation should be updated also the service scripts do not work because they use the jar version of the command if the behavior is not intended ext should be added to the classpath of gitblit jar in the manifest | 1 |
62,704 | 17,155,327,467 | IssuesEvent | 2021-07-14 05:50:38 | martinrotter/rssguard | https://api.github.com/repos/martinrotter/rssguard | closed | [BUG]: inoreader login failed when used with VPN v2ray | Type-Defect | **Brief description of the issue.**
When I try to log in inoreader(and feedly), I always fail to log in.
**How to reproduce the bug?**
Try to use Inoreader and test the login.
Network: I am in China, but I have used v2rayN to hook up the proxy, and I can log in to the inoreader webpage.
Login process: I have applied for app id and app key and re(picture1), and when I click login in rssguard, the webpage jumps to the inoreader authorization page (picture 2), after clicking the authorization, it will jump to the prompt to close the window page (picture 3). But the software crystal surface prompts an unknown error(pic4).
**What was the expected result?**
For the login to report success.
**What actually happened?**
The login fails with the following message:
There is error. unknown error
Hope to reply,thanks a lot!
* OS: win10 20H2 and macos big sur
* RSS Guard version:3.92 and 4.0
<img width="122" alt="pic1" src="https://user-images.githubusercontent.com/54013763/123372407-12234e80-d5b6-11eb-8ad8-7684bba09a3e.png">
<img width="699" alt="pic2" src="https://user-images.githubusercontent.com/54013763/123372766-c1602580-d5b6-11eb-81a6-c831c8807d91.png">
<img width="518" alt="pic3" src="https://user-images.githubusercontent.com/54013763/123372430-1d767a00-d5b6-11eb-8470-8069e50549cd.png">
<img width="122" alt="pic4" src="https://user-images.githubusercontent.com/54013763/123372435-1f403d80-d5b6-11eb-8dfb-75785fe7e5d4.png">
| 1.0 | [BUG]: inoreader login failed when used with VPN v2ray - **Brief description of the issue.**
When I try to log in inoreader(and feedly), I always fail to log in.
**How to reproduce the bug?**
Try to use Inoreader and test the login.
Network: I am in China, but I have used v2rayN to hook up the proxy, and I can log in to the inoreader webpage.
Login process: I have applied for app id and app key and re(picture1), and when I click login in rssguard, the webpage jumps to the inoreader authorization page (picture 2), after clicking the authorization, it will jump to the prompt to close the window page (picture 3). But the software crystal surface prompts an unknown error(pic4).
**What was the expected result?**
For the login to report success.
**What actually happened?**
The login fails with the following message:
There is error. unknown error
Hope to reply,thanks a lot!
* OS: win10 20H2 and macos big sur
* RSS Guard version:3.92 and 4.0
<img width="122" alt="pic1" src="https://user-images.githubusercontent.com/54013763/123372407-12234e80-d5b6-11eb-8ad8-7684bba09a3e.png">
<img width="699" alt="pic2" src="https://user-images.githubusercontent.com/54013763/123372766-c1602580-d5b6-11eb-81a6-c831c8807d91.png">
<img width="518" alt="pic3" src="https://user-images.githubusercontent.com/54013763/123372430-1d767a00-d5b6-11eb-8470-8069e50549cd.png">
<img width="122" alt="pic4" src="https://user-images.githubusercontent.com/54013763/123372435-1f403d80-d5b6-11eb-8dfb-75785fe7e5d4.png">
| defect | inoreader login failed when used with vpn brief description of the issue when i try to log in inoreader(and feedly) i always fail to log in how to reproduce the bug try to use inoreader and test the login network i am in china but i have used to hook up the proxy and i can log in to the inoreader webpage login process i have applied for app id and app key and re( ) and when i click login in rssguard the webpage jumps to the inoreader authorization page picture after clicking the authorization it will jump to the prompt to close the window page picture but the software crystal surface prompts an unknown error what was the expected result for the login to report success what actually happened the login fails with the following message there is error unknown error hope to reply,thanks a lot os and macos big sur rss guard version and img width alt src img width alt src img width alt src img width alt src | 1 |
362,037 | 10,722,221,516 | IssuesEvent | 2019-10-27 10:17:29 | Jinnaix/nemesis-bug-reporting | https://api.github.com/repos/Jinnaix/nemesis-bug-reporting | closed | [Heroic and Normal Steamvault] Mekgineer Steamrigger [Boss] | C++ Script Creature Dungeon Fixed Medium Priority | You can pull the boss with a ranged weapon and the gnomes around him doesnt trigger combat
-------------------
Normal mode missing ability;
Gnome Repairers: A few times during the fight at 75%, 50% and 25% health, he will call upon 3-4 gnomes to heal him. These gnomes have 5k hp and deal 200-500 melee damage on hit. These are non-elite mobs and should receive all focus from DPS, to prevent boss from healing. The healing is a channeled spell which break on damage, and after it has been broken the gnome will start it again after a few seconds, so keep the DPS up.
----------
Heroic mode missing 2 abilities;
1- Steamrigger in heroic mode is almost identical to normal mode, except that an extra gnome
appears every 7 seconds or so
2- Enrage: Enrages after 5 minutes (heroic mode only)
--------
https://wowwiki.fandom.com/wiki/Mekgineer_Steamrigger | 1.0 | [Heroic and Normal Steamvault] Mekgineer Steamrigger [Boss] - You can pull the boss with a ranged weapon and the gnomes around him doesnt trigger combat
-------------------
Normal mode missing ability;
Gnome Repairers: A few times during the fight at 75%, 50% and 25% health, he will call upon 3-4 gnomes to heal him. These gnomes have 5k hp and deal 200-500 melee damage on hit. These are non-elite mobs and should receive all focus from DPS, to prevent boss from healing. The healing is a channeled spell which break on damage, and after it has been broken the gnome will start it again after a few seconds, so keep the DPS up.
----------
Heroic mode missing 2 abilities;
1- Steamrigger in heroic mode is almost identical to normal mode, except that an extra gnome
appears every 7 seconds or so
2- Enrage: Enrages after 5 minutes (heroic mode only)
--------
https://wowwiki.fandom.com/wiki/Mekgineer_Steamrigger | non_defect | mekgineer steamrigger you can pull the boss with a ranged weapon and the gnomes around him doesnt trigger combat normal mode missing ability gnome repairers a few times during the fight at and health he will call upon gnomes to heal him these gnomes have hp and deal melee damage on hit these are non elite mobs and should receive all focus from dps to prevent boss from healing the healing is a channeled spell which break on damage and after it has been broken the gnome will start it again after a few seconds so keep the dps up heroic mode missing abilities steamrigger in heroic mode is almost identical to normal mode except that an extra gnome appears every seconds or so enrage enrages after minutes heroic mode only | 0 |
77,848 | 27,192,301,272 | IssuesEvent | 2023-02-19 23:09:09 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Element Call video room PIP (always on top) avatars are not round | T-Defect | ### Steps to reproduce
1. join an element call video room
2. another call member has an avatar that is rectangular (height > width), which usually isn't visible as they're cropped to circles
3. click on another room, watch the call become a PIP floaty that's always on top
4. notice the kinder surprise shaped avatar in the PIP (screenshot)
### Outcome
#### What did you expect?
avatars are always perfect circles
#### What happened instead?

```html
<div class="_avatar_w2fgr_17 _avatar_7m54n_80" style="background-color: rgb(45, 194, 197); width: 110px; height: 110px; border-radius: 110px; font-size: 55px;">
<img src="https://matrix.example.org/_matrix/media/r0/thumbnail/example.org/abcdef1234567890?width=110&height=110&method=scale" >
</div>
```
### Operating system
arch
### Application version
Element Nightly version: 2023021901 Olm version: 3.2.12
### How did you install the app?
aur
### Homeserver
synapse 1.77.0"
### Will you send logs?
No | 1.0 | Element Call video room PIP (always on top) avatars are not round - ### Steps to reproduce
1. join an element call video room
2. another call member has an avatar that is rectangular (height > width), which usually isn't visible as they're cropped to circles
3. click on another room, watch the call become a PIP floaty that's always on top
4. notice the kinder surprise shaped avatar in the PIP (screenshot)
### Outcome
#### What did you expect?
avatars are always perfect circles
#### What happened instead?

```html
<div class="_avatar_w2fgr_17 _avatar_7m54n_80" style="background-color: rgb(45, 194, 197); width: 110px; height: 110px; border-radius: 110px; font-size: 55px;">
<img src="https://matrix.example.org/_matrix/media/r0/thumbnail/example.org/abcdef1234567890?width=110&height=110&method=scale" >
</div>
```
### Operating system
arch
### Application version
Element Nightly version: 2023021901 Olm version: 3.2.12
### How did you install the app?
aur
### Homeserver
synapse 1.77.0"
### Will you send logs?
No | defect | element call video room pip always on top avatars are not round steps to reproduce join an element call video room another call member has an avatar that is rectangular height width which usually isn t visible as they re cropped to circles click on another room watch the call become a pip floaty that s always on top notice the kinder surprise shaped avatar in the pip screenshot outcome what did you expect avatars are always perfect circles what happened instead html operating system arch application version element nightly version olm version how did you install the app aur homeserver synapse will you send logs no | 1 |
3,668 | 2,610,066,731 | IssuesEvent | 2015-02-26 18:19:36 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 路桥治前列腺炎一般需要多少钱 | auto-migrated Priority-Medium Type-Defect | ```
路桥治前列腺炎一般需要多少钱【台州五洲生殖医院】24小时
健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地�
��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1
04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、
105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:39 | 1.0 | 路桥治前列腺炎一般需要多少钱 - ```
路桥治前列腺炎一般需要多少钱【台州五洲生殖医院】24小时
健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地�
��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1
04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、
105、109、112、901、
902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:39 | defect | 路桥治前列腺炎一般需要多少钱 路桥治前列腺炎一般需要多少钱【台州五洲生殖医院】 健康咨询热线 微信号tzwzszyy 医院地� �� (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 1 |
75,608 | 25,943,666,033 | IssuesEvent | 2022-12-16 21:19:43 | bigbluebutton/bigbluebutton | https://api.github.com/repos/bigbluebutton/bigbluebutton | closed | Breakout rooms don't save room assigning | type: defect status: accepted module: client component: breakout rooms | <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
This issue tracker is only for bbb development related issues.-->
**Describe the bug**
When creating breakout rooms in a single session, only room names and amount of rooms are saved. Participants have to be assigned to rooms manually every time.
**To Reproduce**
Steps to reproduce the behavior:
1. Create breakout rooms as normal and assign participants
2. End breakout room session
3. Create new round of breakout rooms
4. Every student is not assigned to room
**Expected behavior**
Update notes BBB 2.5: "Breakout rooms now remember your previous rooms assignments within the current session. This means if you assign students to breakout rooms and then, later in the class want to reuse the breakout rooms, you don’t need to re-assign students."
**Actual behavior**
Assignment is not saved. Only amount and names of rooms.
**BBB version:**
BigBlueButton continually evolves. Providing the version/build helps us to pinpoint when an issue was introduced.
BigBlueButton Server 2.5.5
Issue also happens on 2.6 test servers
**Desktop (please complete the following information):**
- OS: [multiple (MacOS Big Sur, Windows 10)]
- Browser [multiple: Chrome, Safari, Edge, Firefox]
| 1.0 | Breakout rooms don't save room assigning - <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
This issue tracker is only for bbb development related issues.-->
**Describe the bug**
When creating breakout rooms in a single session, only room names and amount of rooms are saved. Participants have to be assigned to rooms manually every time.
**To Reproduce**
Steps to reproduce the behavior:
1. Create breakout rooms as normal and assign participants
2. End breakout room session
3. Create new round of breakout rooms
4. Every student is not assigned to room
**Expected behavior**
Update notes BBB 2.5: "Breakout rooms now remember your previous rooms assignments within the current session. This means if you assign students to breakout rooms and then, later in the class want to reuse the breakout rooms, you don’t need to re-assign students."
**Actual behavior**
Assignment is not saved. Only amount and names of rooms.
**BBB version:**
BigBlueButton continually evolves. Providing the version/build helps us to pinpoint when an issue was introduced.
BigBlueButton Server 2.5.5
Issue also happens on 2.6 test servers
**Desktop (please complete the following information):**
- OS: [multiple (MacOS Big Sur, Windows 10)]
- Browser [multiple: Chrome, Safari, Edge, Firefox]
| defect | breakout rooms don t save room assigning please do not file issues for general support questions this issue tracker is only for bbb development related issues describe the bug when creating breakout rooms in a single session only room names and amount of rooms are saved participants have to be assigned to rooms manually every time to reproduce steps to reproduce the behavior create breakout rooms as normal and assign participants end breakout room session create new round of breakout rooms every student is not assigned to room expected behavior update notes bbb breakout rooms now remember your previous rooms assignments within the current session this means if you assign students to breakout rooms and then later in the class want to reuse the breakout rooms you don’t need to re assign students actual behavior assignment is not saved only amount and names of rooms bbb version bigbluebutton continually evolves providing the version build helps us to pinpoint when an issue was introduced bigbluebutton server issue also happens on test servers desktop please complete the following information os browser | 1 |
58,215 | 16,439,562,163 | IssuesEvent | 2021-05-20 13:01:37 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | Test Failing: Duplicate Benefit Hub references should not be allowed | Automated testing Core Application Team Critical defect Drupal engineering Unplanned work | ```
Running: entity_reference_validation.feature (2 of 6)
Entity Reference Validation
1) Duplicate Benefit Hub references should not be allowed
0 passing (46s)
1) Entity Reference Validation
Duplicate Benefit Hub references should not be allowed:
AssertionError: Timed out retrying after 10000ms: Expected to find content: '1 error has been found' but never did.
at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:174838:28)
at Context.resolveAndRunStepDefinition (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23780:9)
at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23121:35)
```
_Originally posted by @ElijahLynn in https://github.com/department-of-veterans-affairs/va.gov-cms/issues/5363#issuecomment-842578031_ | 1.0 | Test Failing: Duplicate Benefit Hub references should not be allowed - ```
Running: entity_reference_validation.feature (2 of 6)
Entity Reference Validation
1) Duplicate Benefit Hub references should not be allowed
0 passing (46s)
1) Entity Reference Validation
Duplicate Benefit Hub references should not be allowed:
AssertionError: Timed out retrying after 10000ms: Expected to find content: '1 error has been found' but never did.
at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:174838:28)
at Context.resolveAndRunStepDefinition (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23780:9)
at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23121:35)
```
_Originally posted by @ElijahLynn in https://github.com/department-of-veterans-affairs/va.gov-cms/issues/5363#issuecomment-842578031_ | defect | test failing duplicate benefit hub references should not be allowed running entity reference validation feature of entity reference validation duplicate benefit hub references should not be allowed passing entity reference validation duplicate benefit hub references should not be allowed assertionerror timed out retrying after expected to find content error has been found but never did at context eval at context resolveandrunstepdefinition at context eval originally posted by elijahlynn in | 1 |
9,444 | 2,615,150,382 | IssuesEvent | 2015-03-01 06:27:40 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | Hung on [+] Waiting for beacon | auto-migrated Priority-Triage Type-Defect | ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
v1.3
1. What operating system are you using (Linux is the only supported OS)?
Windows 7 32bit with VMware 8 running Back Track 5 R1
2. Is your wireless card in monitor mode (yes/no)?
Yes under mon0
3. What is the signal strength of the Access Point you are trying to crack?
signal strength is -53
4. What is the manufacturer and model # of the device you are trying to
crack?
Manufacturer SMC Model # D3GN-RRR
http://www.smc.com/files/AJ/SMCD3GN-UserManual_Rogers.pdf
5. What is the entire command line string you are supplying to reaver?
# reaver -i mon0 -b 78:CD:8E:CD:18:99 -vv
6. Please describe what you think the issue is.
I have no idea of what the problem is :(
7. Paste the output from Reaver below.
[+] Waiting for beacon from 78:CD:8E:CD:18:99
Additional info
Modem/router has a WPS button so i assume is supports it.
CTRL + ALT produces the following
[+] Nothing done, nothing to save.
[+] Session saved.
Using Linksys WUSB100 Ralink RT2870/3070 rt2800usb
```
Original issue reported on code.google.com by `gordonJe...@gmail.com` on 18 Jan 2012 at 11:23 | 1.0 | Hung on [+] Waiting for beacon - ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
v1.3
1. What operating system are you using (Linux is the only supported OS)?
Windows 7 32bit with VMware 8 running Back Track 5 R1
2. Is your wireless card in monitor mode (yes/no)?
Yes under mon0
3. What is the signal strength of the Access Point you are trying to crack?
signal strength is -53
4. What is the manufacturer and model # of the device you are trying to
crack?
Manufacturer SMC Model # D3GN-RRR
http://www.smc.com/files/AJ/SMCD3GN-UserManual_Rogers.pdf
5. What is the entire command line string you are supplying to reaver?
# reaver -i mon0 -b 78:CD:8E:CD:18:99 -vv
6. Please describe what you think the issue is.
I have no idea of what the problem is :(
7. Paste the output from Reaver below.
[+] Waiting for beacon from 78:CD:8E:CD:18:99
Additional info
Modem/router has a WPS button so i assume is supports it.
CTRL + ALT produces the following
[+] Nothing done, nothing to save.
[+] Session saved.
Using Linksys WUSB100 Ralink RT2870/3070 rt2800usb
```
Original issue reported on code.google.com by `gordonJe...@gmail.com` on 18 Jan 2012 at 11:23 | defect | hung on waiting for beacon what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os windows with vmware running back track is your wireless card in monitor mode yes no yes under what is the signal strength of the access point you are trying to crack signal strength is what is the manufacturer and model of the device you are trying to crack manufacturer smc model rrr what is the entire command line string you are supplying to reaver reaver i b cd cd vv please describe what you think the issue is i have no idea of what the problem is paste the output from reaver below waiting for beacon from cd cd additional info modem router has a wps button so i assume is supports it ctrl alt produces the following nothing done nothing to save session saved using linksys ralink original issue reported on code google com by gordonje gmail com on jan at | 1 |
110,331 | 11,698,355,243 | IssuesEvent | 2020-03-06 13:44:32 | MeteoSwiss-APN/dawn | https://api.github.com/repos/MeteoSwiss-APN/dawn | closed | Clean up Atlas Utils Repo | Testing documentation enhancement infrastructure | The utilities in https://github.com/mroethlin/AtlasUtilities are poorly documented, there is no build instructions and testing is done manually. Address these issues. Properly document the utilities, provide a basic CMake, make a test script. | 1.0 | Clean up Atlas Utils Repo - The utilities in https://github.com/mroethlin/AtlasUtilities are poorly documented, there is no build instructions and testing is done manually. Address these issues. Properly document the utilities, provide a basic CMake, make a test script. | non_defect | clean up atlas utils repo the utilities in are poorly documented there is no build instructions and testing is done manually address these issues properly document the utilities provide a basic cmake make a test script | 0 |
583,506 | 17,390,573,199 | IssuesEvent | 2021-08-02 06:43:18 | codidact/qpixel | https://api.github.com/repos/codidact/qpixel | closed | I can see a removed thread in Linux Systems | area: html/css/js area: ruby complexity: unassessed priority: medium type: bug | I can see a removed thread in Linux Systems. https://linux.codidact.com/comments/thread/3720

The thread were created by Elgonzo. And, I guess, it was deleted by a moderator. | 1.0 | I can see a removed thread in Linux Systems - I can see a removed thread in Linux Systems. https://linux.codidact.com/comments/thread/3720

The thread were created by Elgonzo. And, I guess, it was deleted by a moderator. | non_defect | i can see a removed thread in linux systems i can see a removed thread in linux systems the thread were created by elgonzo and i guess it was deleted by a moderator | 0 |
563,968 | 16,706,707,153 | IssuesEvent | 2021-06-09 10:51:20 | googleapis/google-api-ruby-client | https://api.github.com/repos/googleapis/google-api-ruby-client | closed | Synthesis failed for language-v1beta1 | autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate language-v1beta1. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the language-v1beta1 API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
2021-06-08 03:17:53,723 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client
2021-06-08 03:17:54,515 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2021-06-08 03:17:54,518 autosynth [DEBUG] > Running: git config user.name yoshi-automation
2021-06-08 03:17:54,520 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com
2021-06-08 03:17:54,522 autosynth [DEBUG] > Running: git config push.default simple
2021-06-08 03:17:54,524 autosynth [DEBUG] > Running: git branch -f autosynth-language-v1beta1
2021-06-08 03:17:54,527 autosynth [DEBUG] > Running: git checkout autosynth-language-v1beta1
Switched to branch 'autosynth-language-v1beta1'
2021-06-08 03:17:54,726 autosynth [INFO] > Running synthtool
2021-06-08 03:17:54,726 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--']
2021-06-08 03:17:54,726 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/language/v1beta1/sponge_log.log
2021-06-08 03:17:54,728 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-language_v1beta1/synth.metadata synth.py -- language v1beta1
2021-06-08 03:17:54,919 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py.
On branch autosynth-language-v1beta1
nothing to commit, working tree clean
2021-06-08 03:17:54,977 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1
DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1
git clean -df
bundle install
Don't run Bundler as root. Bundler can ask for sudo if it is needed, and
installing your bundle as root will break this application for all non-root
users on this machine.
Fetching source index from https://rubygems.org/
Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
Could not fetch specs from https://rubygems.org/ due to underlying error <bad
response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
chown -R 1000:1000 /workspace/generated
2021-06-08 03:17:56,329 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1:
None
ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module>
shell.run(command, hide_output=False)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'language', 'v1beta1']' returned non-zero exit status 1.
2021-06-08 03:17:56,355 autosynth [ERROR] > Synthesis failed
2021-06-08 03:17:56,355 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--', 'language', 'v1beta1']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
| 1.0 | Synthesis failed for language-v1beta1 - Hello! Autosynth couldn't regenerate language-v1beta1. :broken_heart:
Please investigate and fix this issue within 5 business days. While it remains broken,
this library cannot be updated with changes to the language-v1beta1 API, and the library grows
stale.
See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md
for trouble shooting tips.
Here's the output from running `synth.py`:
```
2021-06-08 03:17:53,723 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client
2021-06-08 03:17:54,515 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore
2021-06-08 03:17:54,518 autosynth [DEBUG] > Running: git config user.name yoshi-automation
2021-06-08 03:17:54,520 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com
2021-06-08 03:17:54,522 autosynth [DEBUG] > Running: git config push.default simple
2021-06-08 03:17:54,524 autosynth [DEBUG] > Running: git branch -f autosynth-language-v1beta1
2021-06-08 03:17:54,527 autosynth [DEBUG] > Running: git checkout autosynth-language-v1beta1
Switched to branch 'autosynth-language-v1beta1'
2021-06-08 03:17:54,726 autosynth [INFO] > Running synthtool
2021-06-08 03:17:54,726 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--']
2021-06-08 03:17:54,726 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/language/v1beta1/sponge_log.log
2021-06-08 03:17:54,728 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-language_v1beta1/synth.metadata synth.py -- language v1beta1
2021-06-08 03:17:54,919 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py.
On branch autosynth-language-v1beta1
nothing to commit, working tree clean
2021-06-08 03:17:54,977 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1
DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1
git clean -df
bundle install
Don't run Bundler as root. Bundler can ask for sudo if it is needed, and
installing your bundle as root will break this application for all non-root
users on this machine.
Fetching source index from https://rubygems.org/
Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
Could not fetch specs from https://rubygems.org/ due to underlying error <bad
response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)>
chown -R 1000:1000 /workspace/generated
2021-06-08 03:17:56,329 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1:
None
ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1:
None
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke
return __callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module>
shell.run(command, hide_output=False)
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run
raise exc
File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'language', 'v1beta1']' returned non-zero exit status 1.
2021-06-08 03:17:56,355 autosynth [ERROR] > Synthesis failed
2021-06-08 03:17:56,355 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main
).synthesize(synth_log_path / "sponge_log.log")
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--', 'language', 'v1beta1']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
| non_defect | synthesis failed for language hello autosynth couldn t regenerate language broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the language api and the library grows stale see for trouble shooting tips here s the output from running synth py autosynth logs will be written to tmpfs src logs google api ruby client autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth language autosynth running git checkout autosynth language switched to branch autosynth language autosynth running synthtool autosynth autosynth log file path tmpfs src logs google api ruby client language sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata generated google apis language synth metadata synth py language synthtool executing home kbuilder cache synthtool google api ruby client synth py on branch autosynth language nothing to commit working tree clean synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language debug synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language git clean df bundle install don t run bundler as root bundler can ask for sudo if it is needed and installing your bundle as root will break this application for all non root users on this machine fetching source index from retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable could not fetch specs from due to underlying error bad response service unavailable chown r workspace generated synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language none error synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google api ruby client synth py line in shell run command hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 0 |
35,651 | 7,795,087,015 | IssuesEvent | 2018-06-08 06:43:32 | StrikeNP/trac_test | https://api.github.com/repos/StrikeNP/trac_test | closed | CLUBB can't read GrADS files with a 5 minute timestep (Trac #119) | Migrated from Trac clubb_src defect fasching@uwm.edu | Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce:
Edit ../input/case_setups/fire_model.in to use stats_tout = 300.
$ ./run_scm.bash fire
Edit ../input/case_setup/fire_model.in to use fname_prefix= "fire_tmp"
Edit ../input_misc/inputfields/fire_inputfields to use datafile = "../output/fire"
$ ./run_inputfields.bash fire
Output:
At line 390 of file ../src/inputgrads.F90 (Unit 15 "../output/fire_zt.dat")
Traceback: (Innermost first)
Called from line 246 of file ../src/inputfields.F90
Called from line 707 of file ../src/clubb_driver.F90
Called from line 170 of file ../src/clubb_inputfields.F90
Fortran runtime error: Record number must be positive
I suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/119
```json
{
"status": "closed",
"changetime": "2009-07-22T20:30:12",
"description": "Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce:\n\nEdit ../input/case_setups/fire_model.in to use stats_tout = 300.\n$ ./run_scm.bash fire\nEdit ../input/case_setup/fire_model.in to use fname_prefix= \"fire_tmp\"\nEdit ../input_misc/inputfields/fire_inputfields to use datafile = \"../output/fire\"\n$ ./run_inputfields.bash fire\n\nOutput:\nAt line 390 of file ../src/inputgrads.F90 (Unit 15 \"../output/fire_zt.dat\")\nTraceback: (Innermost first)\n Called from line 246 of file ../src/inputfields.F90\n Called from line 707 of file ../src/clubb_driver.F90\n Called from line 170 of file ../src/clubb_inputfields.F90\nFortran runtime error: Record number must be positive\nI suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug.",
"reporter": "dschanen@uwm.edu",
"cc": "senkbeil@uwm.edu, fasching@uwm.edu",
"resolution": "Verified by V. Larson",
"_ts": "1248294612000000",
"component": "clubb_src",
"summary": "CLUBB can't read GrADS files with a 5 minute timestep",
"priority": "major",
"keywords": "GrADS, inputfields",
"time": "2009-07-09T20:22:15",
"milestone": "",
"owner": "fasching@uwm.edu",
"type": "defect"
}
```
| 1.0 | CLUBB can't read GrADS files with a 5 minute timestep (Trac #119) - Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce:
Edit ../input/case_setups/fire_model.in to use stats_tout = 300.
$ ./run_scm.bash fire
Edit ../input/case_setup/fire_model.in to use fname_prefix= "fire_tmp"
Edit ../input_misc/inputfields/fire_inputfields to use datafile = "../output/fire"
$ ./run_inputfields.bash fire
Output:
At line 390 of file ../src/inputgrads.F90 (Unit 15 "../output/fire_zt.dat")
Traceback: (Innermost first)
Called from line 246 of file ../src/inputfields.F90
Called from line 707 of file ../src/clubb_driver.F90
Called from line 170 of file ../src/clubb_inputfields.F90
Fortran runtime error: Record number must be positive
I suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug.
Attachments:
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/119
```json
{
"status": "closed",
"changetime": "2009-07-22T20:30:12",
"description": "Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce:\n\nEdit ../input/case_setups/fire_model.in to use stats_tout = 300.\n$ ./run_scm.bash fire\nEdit ../input/case_setup/fire_model.in to use fname_prefix= \"fire_tmp\"\nEdit ../input_misc/inputfields/fire_inputfields to use datafile = \"../output/fire\"\n$ ./run_inputfields.bash fire\n\nOutput:\nAt line 390 of file ../src/inputgrads.F90 (Unit 15 \"../output/fire_zt.dat\")\nTraceback: (Innermost first)\n Called from line 246 of file ../src/inputfields.F90\n Called from line 707 of file ../src/clubb_driver.F90\n Called from line 170 of file ../src/clubb_inputfields.F90\nFortran runtime error: Record number must be positive\nI suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug.",
"reporter": "dschanen@uwm.edu",
"cc": "senkbeil@uwm.edu, fasching@uwm.edu",
"resolution": "Verified by V. Larson",
"_ts": "1248294612000000",
"component": "clubb_src",
"summary": "CLUBB can't read GrADS files with a 5 minute timestep",
"priority": "major",
"keywords": "GrADS, inputfields",
"time": "2009-07-09T20:22:15",
"milestone": "",
"owner": "fasching@uwm.edu",
"type": "defect"
}
```
| defect | clubb can t read grads files with a minute timestep trac ryan discovered when trying to do a run inputfields simulation that our inputgrads code for reading a variable does not work for grads data output every seconds to reproduce edit input case setups fire model in to use stats tout run scm bash fire edit input case setup fire model in to use fname prefix fire tmp edit input misc inputfields fire inputfields to use datafile output fire run inputfields bash fire output at line of file src inputgrads unit output fire zt dat traceback innermost first called from line of file src inputfields called from line of file src clubb driver called from line of file src clubb inputfields fortran runtime error record number must be positive i suspect the formula might have a bug attachments migrated from json status closed changetime description ryan discovered when trying to do a run inputfields simulation that our inputgrads code for reading a variable does not work for grads data output every seconds to reproduce n nedit input case setups fire model in to use stats tout n run scm bash fire nedit input case setup fire model in to use fname prefix fire tmp nedit input misc inputfields fire inputfields to use datafile output fire n run inputfields bash fire n noutput nat line of file src inputgrads unit output fire zt dat ntraceback innermost first n called from line of file src inputfields n called from line of file src clubb driver n called from line of file src clubb inputfields nfortran runtime error record number must be positive ni suspect the formula might have a bug reporter dschanen uwm edu cc senkbeil uwm edu fasching uwm edu resolution verified by v larson ts component clubb src summary clubb can t read grads files with a minute timestep priority major keywords grads inputfields time milestone owner fasching uwm edu type defect | 1 |
21,304 | 3,487,385,789 | IssuesEvent | 2016-01-01 21:11:11 | jOOQ/jOOL | https://api.github.com/repos/jOOQ/jOOL | closed | NoSuchElementException is thrown on Seq.cycle() for empty Seq | P: Medium R: Fixed T: Defect | The following code:
```java
Seq.empty().cycle().toList();
```
throws:
```
java.util.NoSuchElementException
at java.util.ArrayList$Itr.next(ArrayList.java:854)
at org.jooq.lambda.Seq$1Cycle.next(Seq.java:1910)
at java.util.Iterator.forEachRemaining(Iterator.java:116)
at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at org.jooq.lambda.SeqImpl.collect(SeqImpl.java:176)
at org.jooq.lambda.Seq.toList(Seq.java:4634)
at org.jooq.lambda.Seq.toList(Seq.java:1237)
at org.jooq.lambda.SeqTest.testCycle(SeqTest.java:1133)
``` | 1.0 | NoSuchElementException is thrown on Seq.cycle() for empty Seq - The following code:
```java
Seq.empty().cycle().toList();
```
throws:
```
java.util.NoSuchElementException
at java.util.ArrayList$Itr.next(ArrayList.java:854)
at org.jooq.lambda.Seq$1Cycle.next(Seq.java:1910)
at java.util.Iterator.forEachRemaining(Iterator.java:116)
at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
at org.jooq.lambda.SeqImpl.collect(SeqImpl.java:176)
at org.jooq.lambda.Seq.toList(Seq.java:4634)
at org.jooq.lambda.Seq.toList(Seq.java:1237)
at org.jooq.lambda.SeqTest.testCycle(SeqTest.java:1133)
``` | defect | nosuchelementexception is thrown on seq cycle for empty seq the following code java seq empty cycle tolist throws java util nosuchelementexception at java util arraylist itr next arraylist java at org jooq lambda seq next seq java at java util iterator foreachremaining iterator java at java util spliterators iteratorspliterator foreachremaining spliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at org jooq lambda seqimpl collect seqimpl java at org jooq lambda seq tolist seq java at org jooq lambda seq tolist seq java at org jooq lambda seqtest testcycle seqtest java | 1 |
259,256 | 22,423,011,530 | IssuesEvent | 2022-06-20 06:26:02 | opencurve/curve | https://api.github.com/repos/opencurve/curve | closed | after restart curvefs mds, the partition num become 0 in metaserver | bug need test | **Describe the bug (描述bug)**
1. curvefs create one fs, then metric shows metaserver has partition num.
> topology_metric_metaserver_id_1_partition_num 3
> topology_metric_metaserver_id_2_partition_num 3
> topology_metric_metaserver_id_3_partition_num 3
> topology_metric_pool_default_pool_partition_num 3
2. curvefs restart mds, then metric shows metaserver has no partition num
> topology_metric_metaserver_id_1_partition_num 0
> topology_metric_metaserver_id_2_partition_num 0
> topology_metric_metaserver_id_3_partition_num 0
> topology_metric_pool_default_pool_partition_num 3
**To Reproduce (复现方法)**
Mount 1 fs on curvefs, restart mds.
**Expected behavior (期望行为)**
**Versions (各种版本)**
OS:
Compiler:
branch: master
commit id: 941e593a4e8818266cc695cd20fe2956e5807910
**Additional context/screenshots (更多上下文/截图)**
| 1.0 | after restart curvefs mds, the partition num become 0 in metaserver - **Describe the bug (描述bug)**
1. curvefs create one fs, then metric shows metaserver has partition num.
> topology_metric_metaserver_id_1_partition_num 3
> topology_metric_metaserver_id_2_partition_num 3
> topology_metric_metaserver_id_3_partition_num 3
> topology_metric_pool_default_pool_partition_num 3
2. curvefs restart mds, then metric shows metaserver has no partition num
> topology_metric_metaserver_id_1_partition_num 0
> topology_metric_metaserver_id_2_partition_num 0
> topology_metric_metaserver_id_3_partition_num 0
> topology_metric_pool_default_pool_partition_num 3
**To Reproduce (复现方法)**
Mount 1 fs on curvefs, restart mds.
**Expected behavior (期望行为)**
**Versions (各种版本)**
OS:
Compiler:
branch: master
commit id: 941e593a4e8818266cc695cd20fe2956e5807910
**Additional context/screenshots (更多上下文/截图)**
| non_defect | after restart curvefs mds the partition num become in metaserver describe the bug 描述bug curvefs create one fs then metric shows metaserver has partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric pool default pool partition num curvefs restart mds then metric shows metaserver has no partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric pool default pool partition num to reproduce 复现方法 mount fs on curvefs restart mds expected behavior 期望行为 versions 各种版本 os compiler branch master commit id additional context screenshots 更多上下文 截图 | 0 |
55,332 | 6,470,512,135 | IssuesEvent | 2017-08-17 09:13:09 | elastic/logstash | https://api.github.com/repos/elastic/logstash | opened | Module Tests are Broken in Part due to Removing Max Bucket Size Setting | test failure | Obvious issue :) the fix for #8010 broke two tests it seems:
```sh
1) LogStash::Modules::Scaffold kibana operation provides a list of importable files
Failure/Error: expect(resource1.content.size).to eq(2)
expected: 2
got: 1
(compared using ==)
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method
Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}})
expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}
got: {"changes"=>{"defaultIndex"=>"tester-*"}}
(compared using ==)
Diff:
@@ -1,2 +1,2 @@
-"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"},
+"changes" => {"defaultIndex"=>"tester-*"},
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
Finished in 6 minutes 54 seconds (files took 21.47 seconds to load)
2462 examples, 2 failures, 3 pending
1) LogStash::Modules::Scaffold kibana operation provides a list of importable files
Failure/Error: expect(resource1.content.size).to eq(2)
expected: 2
got: 1
(compared using ==)
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method
Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}})
expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}
got: {"changes"=>{"defaultIndex"=>"tester-*"}}
(compared using ==)
Diff:
@@ -1,2 +1,2 @@
-"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"},
+"changes" => {"defaultIndex"=>"tester-*"},
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
Finished in 6 minutes 54 seconds (files took 21.47 seconds to load)
2462 examples, 2 failures, 3 pending
``` | 1.0 | Module Tests are Broken in Part due to Removing Max Bucket Size Setting - Obvious issue :) the fix for #8010 broke two tests it seems:
```sh
1) LogStash::Modules::Scaffold kibana operation provides a list of importable files
Failure/Error: expect(resource1.content.size).to eq(2)
expected: 2
got: 1
(compared using ==)
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method
Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}})
expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}
got: {"changes"=>{"defaultIndex"=>"tester-*"}}
(compared using ==)
Diff:
@@ -1,2 +1,2 @@
-"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"},
+"changes" => {"defaultIndex"=>"tester-*"},
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
Finished in 6 minutes 54 seconds (files took 21.47 seconds to load)
2462 examples, 2 failures, 3 pending
1) LogStash::Modules::Scaffold kibana operation provides a list of importable files
Failure/Error: expect(resource1.content.size).to eq(2)
expected: 2
got: 1
(compared using ==)
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method
Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}})
expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}
got: {"changes"=>{"defaultIndex"=>"tester-*"}}
(compared using ==)
Diff:
@@ -1,2 +1,2 @@
-"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"},
+"changes" => {"defaultIndex"=>"tester-*"},
# ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)'
# ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb'
# ./lib/bootstrap/rspec.rb:13:in `<main>'
Finished in 6 minutes 54 seconds (files took 21.47 seconds to load)
2462 examples, 2 failures, 3 pending
``` | non_defect | module tests are broken in part due to removing max bucket size setting obvious issue the fix for broke two tests it seems sh logstash modules scaffold kibana operation provides a list of importable files failure error expect content size to eq expected got compared using logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in logstash modules scaffold importing to elasticsearch stubbed client calls the import method failure error expect contents to eq changes defaultindex tester metrics max buckets expected changes defaultindex tester metrics max buckets got changes defaultindex tester compared using diff changes defaultindex tester metrics max buckets changes defaultindex tester logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in finished in minutes seconds files took seconds to load examples failures pending logstash modules scaffold kibana operation provides a list of importable files failure error expect content size to eq expected got compared using logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in logstash modules scaffold importing to elasticsearch stubbed client calls the import method failure error expect contents to eq changes defaultindex tester metrics max buckets expected changes defaultindex tester metrics max buckets got changes defaultindex tester compared using diff changes defaultindex tester metrics max buckets changes defaultindex tester logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in finished in minutes seconds files took seconds to load examples failures pending | 0 |
44,092 | 5,732,291,974 | IssuesEvent | 2017-04-21 14:32:02 | gudell/bwardp | https://api.github.com/repos/gudell/bwardp | opened | Update HTML on Website for New Pricing | design open development | Implement changes to website to reflect new pricing and the features that would go under each subscription type | 1.0 | Update HTML on Website for New Pricing - Implement changes to website to reflect new pricing and the features that would go under each subscription type | non_defect | update html on website for new pricing implement changes to website to reflect new pricing and the features that would go under each subscription type | 0 |
30,202 | 6,041,037,933 | IssuesEvent | 2017-06-10 20:05:32 | Advanced-Post-List/advanced-post-list | https://api.github.com/repos/Advanced-Post-List/advanced-post-list | closed | [Show Preview] Not Correctly Working On Admin Side | P3 - Major T-Defect X-Fixed | The show preview checkbox next to the save button does not appear to do anything.

| 1.0 | [Show Preview] Not Correctly Working On Admin Side - The show preview checkbox next to the save button does not appear to do anything.

| defect | not correctly working on admin side the show preview checkbox next to the save button does not appear to do anything | 1 |
29,617 | 11,763,768,999 | IssuesEvent | 2020-03-14 09:04:00 | uniquelyparticular/shipengine-request | https://api.github.com/repos/uniquelyparticular/shipengine-request | opened | WS-2020-0042 (High) detected in acorn-5.7.3.tgz, acorn-6.1.1.tgz | security vulnerability | ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>acorn-5.7.3.tgz</b>, <b>acorn-6.1.1.tgz</b></p></summary>
<p>
<details><summary><b>acorn-5.7.3.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.8.0.tgz (Root Library)
- jest-cli-24.8.0.tgz
- jest-config-24.8.0.tgz
- jest-environment-jsdom-24.8.0.tgz
- jsdom-11.12.0.tgz
- :x: **acorn-5.7.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>acorn-6.1.1.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz">https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn-globals/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.8.0.tgz (Root Library)
- jest-cli-24.8.0.tgz
- jest-config-24.8.0.tgz
- jest-environment-jsdom-24.8.0.tgz
- jsdom-11.12.0.tgz
- acorn-globals-4.3.2.tgz
- :x: **acorn-6.1.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/shipengine-request/commit/f53abdebfdb119b33faab5e4131106b799eb636c">f53abdebfdb119b33faab5e4131106b799eb636c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-08
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-08</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2020-0042 (High) detected in acorn-5.7.3.tgz, acorn-6.1.1.tgz - ## WS-2020-0042 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>acorn-5.7.3.tgz</b>, <b>acorn-6.1.1.tgz</b></p></summary>
<p>
<details><summary><b>acorn-5.7.3.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.8.0.tgz (Root Library)
- jest-cli-24.8.0.tgz
- jest-config-24.8.0.tgz
- jest-environment-jsdom-24.8.0.tgz
- jsdom-11.12.0.tgz
- :x: **acorn-5.7.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>acorn-6.1.1.tgz</b></p></summary>
<p>ECMAScript parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz">https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn-globals/node_modules/acorn/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.8.0.tgz (Root Library)
- jest-cli-24.8.0.tgz
- jest-config-24.8.0.tgz
- jest-environment-jsdom-24.8.0.tgz
- jsdom-11.12.0.tgz
- acorn-globals-4.3.2.tgz
- :x: **acorn-6.1.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/shipengine-request/commit/f53abdebfdb119b33faab5e4131106b799eb636c">f53abdebfdb119b33faab5e4131106b799eb636c</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser.
<p>Publish Date: 2020-03-08
<p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p>
<p>Release Date: 2020-03-08</p>
<p>Fix Resolution: 7.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | ws high detected in acorn tgz acorn tgz ws high severity vulnerability vulnerable libraries acorn tgz acorn tgz acorn tgz ecmascript parser library home page a href path to dependency file tmp ws scm shipengine request package json path to vulnerable library tmp ws scm shipengine request node modules acorn package json dependency hierarchy jest tgz root library jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz x acorn tgz vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file tmp ws scm shipengine request package json path to vulnerable library tmp ws scm shipengine request node modules acorn globals node modules acorn package json dependency hierarchy jest tgz root library jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz acorn globals tgz x acorn tgz vulnerable library found in head commit a href vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
62,224 | 8,582,249,667 | IssuesEvent | 2018-11-13 16:31:52 | godaddy/wp-primer-theme | https://api.github.com/repos/godaddy/wp-primer-theme | closed | Update documentation for latest release | bug documentation | The documentation hasn't been updated since v1.8.2. Investigate why this hasn't been updating in our build & release process and fix it.
https://godaddy.github.io/wp-primer-theme/index.html | 1.0 | Update documentation for latest release - The documentation hasn't been updated since v1.8.2. Investigate why this hasn't been updating in our build & release process and fix it.
https://godaddy.github.io/wp-primer-theme/index.html | non_defect | update documentation for latest release the documentation hasn t been updated since investigate why this hasn t been updating in our build release process and fix it | 0 |
55,238 | 14,289,208,147 | IssuesEvent | 2020-11-23 18:53:22 | idaholab/raven | https://api.github.com/repos/idaholab/raven | closed | [DEFECT] statsmodels error | defect priority_normal | --------
Defect Description
--------
**Describe the defect**
When using the internal parallel as True, and sampling the ARMA, it might cause the error of `statsmodels.tsa.statespace._initialization.dInitialization.__init__`
##### What did you expect to see happen?
##### What did you see instead?
```
( 33.51 sec) STEP MULTIRUN : Message -> *** Beginning run ***
(pid=57717) Unable to load "timingProfile" decorator; replacing with passthrough ...
(pid=57717) Unable to load "memoryProfile" decorator; replacing with passthrough ...
(pid=57718) Unable to load "timingProfile" decorator; replacing with passthrough ...
(pid=57718) Unable to load "memoryProfile" decorator; replacing with passthrough ...
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1
(pid=57718) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working
(pid=57718) from collections import Mapping
(pid=57717) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working
(pid=57717) from collections import Mapping
Traceback (most recent call last):
File "/home/zhoujia/projects/raven/framework/Driver.py", line 305, in <module>
raven()
File "/home/zhoujia/projects/raven/framework/Driver.py", line 258, in raven
simulation.run()
File "/home/zhoujia/projects/raven/framework/Simulation.py", line 762, in run
stepInstance.takeAstep(stepInputDict)
File "/home/zhoujia/projects/raven/framework/Steps.py", line 346, in takeAstep
self._localTakeAstepRun(inDictionary)
File "/home/zhoujia/projects/raven/framework/Steps.py", line 690, in _localTakeAstepRun
myLambda([finishedJob,outputs[outIndex]])
File "/home/zhoujia/projects/raven/framework/Steps.py", line 637, in <lambda>
self._outputCollectionLambda.append( (lambda x: inDictionary['Model'].collectOutput(x[0],x[1]), outIndex) )
File "/home/zhoujia/projects/raven/framework/Models/EnsembleModel.py", line 453, in collectOutput
evaluation = finishedJob.getEvaluation()
File "/home/zhoujia/projects/raven/framework/Runners/InternalRunner.py", line 124, in getEvaluation
self._collectRunnerResponse()
File "/home/zhoujia/projects/raven/framework/Runners/DistributedMemoryRunner.py", line 109, in _collectRunnerResponse
self.runReturn = ray.get(self.thread) if im.isLibAvail("ray") else self.thread()
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/worker.py", line 1428, in get
raise value.as_instanceof_cause()
ray.exceptions.RayTaskError(ValueError): ray::evaluateSample() (pid=57718, ip=141.221.84.10)
File "python/ray/_raylet.pyx", line 448, in ray._raylet.execute_task
File "python/ray/_raylet.pyx", line 465, in ray._raylet.execute_task
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 308, in deserialize_objects
self._deserialize_object(data, metadata, object_ref))
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 247, in _deserialize_object
return self._deserialize_msgpack_data(data, metadata)
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 226, in _deserialize_msgpack_data
python_objects = self._deserialize_pickle5_data(pickle5_data)
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 214, in _deserialize_pickle5_data
obj = pickle.loads(in_band, buffers=buffers)
File "statsmodels/tsa/statespace/_initialization.pyx", line 227, in statsmodels.tsa.statespace._initialization.dInitialization.__init__
File "stringsource", line 658, in View.MemoryView.memoryview_cwrapper
File "stringsource", line 349, in View.MemoryView.memoryview.__cinit__
ValueError: buffer source array is read-only
```
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| 1.0 | [DEFECT] statsmodels error - --------
Defect Description
--------
**Describe the defect**
When using the internal parallel as True, and sampling the ARMA, it might cause the error of `statsmodels.tsa.statespace._initialization.dInitialization.__init__`
##### What did you expect to see happen?
##### What did you see instead?
```
( 33.51 sec) STEP MULTIRUN : Message -> *** Beginning run ***
(pid=57717) Unable to load "timingProfile" decorator; replacing with passthrough ...
(pid=57717) Unable to load "memoryProfile" decorator; replacing with passthrough ...
(pid=57718) Unable to load "timingProfile" decorator; replacing with passthrough ...
(pid=57718) Unable to load "memoryProfile" decorator; replacing with passthrough ...
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1
(pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1
(pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1
(pid=57718) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working
(pid=57718) from collections import Mapping
(pid=57717) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working
(pid=57717) from collections import Mapping
Traceback (most recent call last):
File "/home/zhoujia/projects/raven/framework/Driver.py", line 305, in <module>
raven()
File "/home/zhoujia/projects/raven/framework/Driver.py", line 258, in raven
simulation.run()
File "/home/zhoujia/projects/raven/framework/Simulation.py", line 762, in run
stepInstance.takeAstep(stepInputDict)
File "/home/zhoujia/projects/raven/framework/Steps.py", line 346, in takeAstep
self._localTakeAstepRun(inDictionary)
File "/home/zhoujia/projects/raven/framework/Steps.py", line 690, in _localTakeAstepRun
myLambda([finishedJob,outputs[outIndex]])
File "/home/zhoujia/projects/raven/framework/Steps.py", line 637, in <lambda>
self._outputCollectionLambda.append( (lambda x: inDictionary['Model'].collectOutput(x[0],x[1]), outIndex) )
File "/home/zhoujia/projects/raven/framework/Models/EnsembleModel.py", line 453, in collectOutput
evaluation = finishedJob.getEvaluation()
File "/home/zhoujia/projects/raven/framework/Runners/InternalRunner.py", line 124, in getEvaluation
self._collectRunnerResponse()
File "/home/zhoujia/projects/raven/framework/Runners/DistributedMemoryRunner.py", line 109, in _collectRunnerResponse
self.runReturn = ray.get(self.thread) if im.isLibAvail("ray") else self.thread()
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/worker.py", line 1428, in get
raise value.as_instanceof_cause()
ray.exceptions.RayTaskError(ValueError): ray::evaluateSample() (pid=57718, ip=141.221.84.10)
File "python/ray/_raylet.pyx", line 448, in ray._raylet.execute_task
File "python/ray/_raylet.pyx", line 465, in ray._raylet.execute_task
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 308, in deserialize_objects
self._deserialize_object(data, metadata, object_ref))
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 247, in _deserialize_object
return self._deserialize_msgpack_data(data, metadata)
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 226, in _deserialize_msgpack_data
python_objects = self._deserialize_pickle5_data(pickle5_data)
File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 214, in _deserialize_pickle5_data
obj = pickle.loads(in_band, buffers=buffers)
File "statsmodels/tsa/statespace/_initialization.pyx", line 227, in statsmodels.tsa.statespace._initialization.dInitialization.__init__
File "stringsource", line 658, in View.MemoryView.memoryview_cwrapper
File "stringsource", line 349, in View.MemoryView.memoryview.__cinit__
ValueError: buffer source array is read-only
```
##### Do you have a suggested fix for the development team?
**Describe how to Reproduce**
Steps to reproduce the behavior:
1.
2.
3.
4.
**Screenshots and Input Files**
Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
**Platform (please complete the following information):**
- OS: [e.g. iOS]
- Version: [e.g. 22]
- Dependencies Installation: [CONDA or PIP]
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or task?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)?
- [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
| defect | statsmodels error defect description describe the defect when using the internal parallel as true and sampling the arma it might cause the error of statsmodels tsa statespace initialization dinitialization init what did you expect to see happen what did you see instead sec step multirun message beginning run pid unable to load timingprofile decorator replacing with passthrough pid unable to load memoryprofile decorator replacing with passthrough pid unable to load timingprofile decorator replacing with passthrough pid unable to load memoryprofile decorator replacing with passthrough pid input spec error adding checked and unchecked to lowerbound in uniformdiscrete len pid input spec error adding checked and unchecked to upperbound in uniformdiscrete len pid input spec error adding checked and unchecked to lowerbound in loguniform len pid input spec error adding checked and unchecked to upperbound in loguniform len pid input spec error adding checked and unchecked to lowerbound in uniformdiscrete len pid input spec error adding checked and unchecked to upperbound in uniformdiscrete len pid input spec error adding checked and unchecked to lowerbound in loguniform len pid input spec error adding checked and unchecked to upperbound in loguniform len pid home zhoujia conda envs raven libraries lib site packages patsy constraint py deprecationwarning using or importing the abcs from collections instead of from collections abc is deprecated since python and in it will stop working pid from collections import mapping pid home zhoujia conda envs raven libraries lib site packages patsy constraint py deprecationwarning using or importing the abcs from collections instead of from collections abc is deprecated since python and in it will stop working pid from collections import mapping traceback most recent call last file home zhoujia projects raven framework driver py line in raven file home zhoujia projects raven framework driver py line in raven simulation run file home zhoujia projects raven framework simulation py line in run stepinstance takeastep stepinputdict file home zhoujia projects raven framework steps py line in takeastep self localtakeasteprun indictionary file home zhoujia projects raven framework steps py line in localtakeasteprun mylambda file home zhoujia projects raven framework steps py line in self outputcollectionlambda append lambda x indictionary collectoutput x x outindex file home zhoujia projects raven framework models ensemblemodel py line in collectoutput evaluation finishedjob getevaluation file home zhoujia projects raven framework runners internalrunner py line in getevaluation self collectrunnerresponse file home zhoujia projects raven framework runners distributedmemoryrunner py line in collectrunnerresponse self runreturn ray get self thread if im islibavail ray else self thread file home zhoujia conda envs raven libraries lib site packages ray worker py line in get raise value as instanceof cause ray exceptions raytaskerror valueerror ray evaluatesample pid ip file python ray raylet pyx line in ray raylet execute task file python ray raylet pyx line in ray raylet execute task file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize objects self deserialize object data metadata object ref file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize object return self deserialize msgpack data data metadata file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize msgpack data python objects self deserialize data data file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize data obj pickle loads in band buffers buffers file statsmodels tsa statespace initialization pyx line in statsmodels tsa statespace initialization dinitialization init file stringsource line in view memoryview memoryview cwrapper file stringsource line in view memoryview memoryview cinit valueerror buffer source array is read only do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided | 1 |
53,643 | 13,262,039,387 | IssuesEvent | 2020-08-20 20:59:43 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [mue] Memory Leak (Trac #1801) | Migrated from Trac combo reconstruction defect | found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1801">https://code.icecube.wisc.edu/projects/icecube/ticket/1801</a>, reported by kjmeagherand owned by dima</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-08-01T18:41:59",
"_ts": "1470076919611236",
"description": "found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2016-07-27T08:06:31",
"component": "combo reconstruction",
"summary": "[mue] Memory Leak",
"priority": "normal",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "dima",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [mue] Memory Leak (Trac #1801) - found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1801">https://code.icecube.wisc.edu/projects/icecube/ticket/1801</a>, reported by kjmeagherand owned by dima</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-08-01T18:41:59",
"_ts": "1470076919611236",
"description": "found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath",
"reporter": "kjmeagher",
"cc": "",
"resolution": "fixed",
"time": "2016-07-27T08:06:31",
"component": "combo reconstruction",
"summary": "[mue] Memory Leak",
"priority": "normal",
"keywords": "",
"milestone": "Long-Term Future",
"owner": "dima",
"type": "defect"
}
```
</p>
</details>
| defect | memory leak trac found by static analyser migrated from json status closed changetime ts description found by static analyser reporter kjmeagher cc resolution fixed time component combo reconstruction summary memory leak priority normal keywords milestone long term future owner dima type defect | 1 |
1,153 | 2,598,004,852 | IssuesEvent | 2015-02-22 01:41:55 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | opened | Can't build/land/etc on far right and/or bottom edge of the map | auto-migrated Component-Logic Priority-High Type-Defect Usability | ```
What steps will reproduce the problem?
1. Call canBuildHere(), placing a building at a clear location whose edge is
against the far right and/or bottom edge of the map.
What is the expected output? What do you see instead?
It should return true, but it always returns false.
What version of the product are you using? On what operating system?
HEAD of branch bwapi4 of the src
Please provide any additional information below.
I think Shared\Templates.h:canBuildHere is returning false because the rb
variable is initialised to an invalid location (i.e. off the edge of the map)
during this statement:
TilePosition rb = position + TilePosition(width, height);
It would probably be simple to fix this function, but I expect that similar
bugs exist in several other places in the BWAPI src, and there may be a lot of
src in BWAPI (and other libraries & bots) that is only called if this function
returns true and manipulates position-related objects without doing map array
bounds safety checks.
My concern is that if just this function is fixed then unless all such src is
fixed too then there may be crashes or strange problems caused by memory
corruption, and these bugs may be hard to reproduce/debug because they would
probably only occur in rare scenarios, and may not even be consistently
reproducable.
So, risk may not be worth the effort.
```
-----
Original issue reported on code.google.com by `chris.c...@gmail.com` on 23 May 2013 at 2:01 | 1.0 | Can't build/land/etc on far right and/or bottom edge of the map - ```
What steps will reproduce the problem?
1. Call canBuildHere(), placing a building at a clear location whose edge is
against the far right and/or bottom edge of the map.
What is the expected output? What do you see instead?
It should return true, but it always returns false.
What version of the product are you using? On what operating system?
HEAD of branch bwapi4 of the src
Please provide any additional information below.
I think Shared\Templates.h:canBuildHere is returning false because the rb
variable is initialised to an invalid location (i.e. off the edge of the map)
during this statement:
TilePosition rb = position + TilePosition(width, height);
It would probably be simple to fix this function, but I expect that similar
bugs exist in several other places in the BWAPI src, and there may be a lot of
src in BWAPI (and other libraries & bots) that is only called if this function
returns true and manipulates position-related objects without doing map array
bounds safety checks.
My concern is that if just this function is fixed then unless all such src is
fixed too then there may be crashes or strange problems caused by memory
corruption, and these bugs may be hard to reproduce/debug because they would
probably only occur in rare scenarios, and may not even be consistently
reproducable.
So, risk may not be worth the effort.
```
-----
Original issue reported on code.google.com by `chris.c...@gmail.com` on 23 May 2013 at 2:01 | defect | can t build land etc on far right and or bottom edge of the map what steps will reproduce the problem call canbuildhere placing a building at a clear location whose edge is against the far right and or bottom edge of the map what is the expected output what do you see instead it should return true but it always returns false what version of the product are you using on what operating system head of branch of the src please provide any additional information below i think shared templates h canbuildhere is returning false because the rb variable is initialised to an invalid location i e off the edge of the map during this statement tileposition rb position tileposition width height it would probably be simple to fix this function but i expect that similar bugs exist in several other places in the bwapi src and there may be a lot of src in bwapi and other libraries bots that is only called if this function returns true and manipulates position related objects without doing map array bounds safety checks my concern is that if just this function is fixed then unless all such src is fixed too then there may be crashes or strange problems caused by memory corruption and these bugs may be hard to reproduce debug because they would probably only occur in rare scenarios and may not even be consistently reproducable so risk may not be worth the effort original issue reported on code google com by chris c gmail com on may at | 1 |
52,231 | 7,754,237,826 | IssuesEvent | 2018-05-31 05:39:08 | dustinRDF/food-roulette | https://api.github.com/repos/dustinRDF/food-roulette | closed | Create Cards for Each Section | Documentation | - [x] Application Information #11
- [x] Application Design #12
- [x] Development and Testing #13
- [x] Deployment #14 | 1.0 | Create Cards for Each Section - - [x] Application Information #11
- [x] Application Design #12
- [x] Development and Testing #13
- [x] Deployment #14 | non_defect | create cards for each section application information application design development and testing deployment | 0 |
802,968 | 29,086,615,502 | IssuesEvent | 2023-05-16 00:50:19 | googleapis/google-cloud-go | https://api.github.com/repos/googleapis/google-cloud-go | closed | pubsublite/internal/wire: TestRoutingPublisherUnloadIdlePublisher failed | type: bug priority: p1 api: pubsublite flakybot: issue | Note: #7854 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 85d084cc9d22f03ec60356d9f875c143d92a4aba
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/5cb50424-759c-407f-867d-e2264dd350dc), [Sponge](http://sponge2/5cb50424-759c-407f-867d-e2264dd350dc)
status: failed
<details><summary>Test output</summary><br><pre> verifier.go:226: call(2): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"2"} first_sequence_number:1}
verifier.go:226: call(3): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"3"} first_sequence_number:2}
verifier.go:226: call(4): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"4"} first_sequence_number:3}
verifier.go:226: call(5): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"5"} first_sequence_number:4}
verifier.go:226: call(6): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"6"} first_sequence_number:5}
verifier.go:257: unexpected stream index 1 for key publish:projects/123456/locations/us-central1-b/topics/my-topic:0
publish_batcher_test.go:71: Publish("2") error: (rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection), want: partition=0,offset=11
publish_batcher_test.go:71: Publish("3") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=12
publish_batcher_test.go:71: Publish("4") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=13
publish_batcher_test.go:71: Publish("5") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=14
publish_batcher_test.go:71: Publish("6") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=15
publish_batcher_test.go:71: Publish("b") error: (pubsublite: service has stopped or is stopping), want: partition=1,offset=22
publisher_test.go:619: Stop() got err: (publisher(projects/123456/locations/us-central1-b/topics/my-topic/partitions/0): rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection)
verifier.go:226: call(0): did not receive expected request:
[*pubsublitepb.PublishRequest] initial_request:{topic:"projects/123456/locations/us-central1-b/topics/my-topic" partition:1 client_id:"publisher"}
verifier.go:226: call(1): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"bar" data:"b"} first_sequence_number:1}</pre></details> | 1.0 | pubsublite/internal/wire: TestRoutingPublisherUnloadIdlePublisher failed - Note: #7854 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 85d084cc9d22f03ec60356d9f875c143d92a4aba
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/5cb50424-759c-407f-867d-e2264dd350dc), [Sponge](http://sponge2/5cb50424-759c-407f-867d-e2264dd350dc)
status: failed
<details><summary>Test output</summary><br><pre> verifier.go:226: call(2): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"2"} first_sequence_number:1}
verifier.go:226: call(3): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"3"} first_sequence_number:2}
verifier.go:226: call(4): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"4"} first_sequence_number:3}
verifier.go:226: call(5): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"5"} first_sequence_number:4}
verifier.go:226: call(6): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"6"} first_sequence_number:5}
verifier.go:257: unexpected stream index 1 for key publish:projects/123456/locations/us-central1-b/topics/my-topic:0
publish_batcher_test.go:71: Publish("2") error: (rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection), want: partition=0,offset=11
publish_batcher_test.go:71: Publish("3") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=12
publish_batcher_test.go:71: Publish("4") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=13
publish_batcher_test.go:71: Publish("5") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=14
publish_batcher_test.go:71: Publish("6") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=15
publish_batcher_test.go:71: Publish("b") error: (pubsublite: service has stopped or is stopping), want: partition=1,offset=22
publisher_test.go:619: Stop() got err: (publisher(projects/123456/locations/us-central1-b/topics/my-topic/partitions/0): rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection)
verifier.go:226: call(0): did not receive expected request:
[*pubsublitepb.PublishRequest] initial_request:{topic:"projects/123456/locations/us-central1-b/topics/my-topic" partition:1 client_id:"publisher"}
verifier.go:226: call(1): did not receive expected request:
[*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"bar" data:"b"} first_sequence_number:1}</pre></details> | non_defect | pubsublite internal wire testroutingpublisherunloadidlepublisher failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go unexpected stream index for key publish projects locations us b topics my topic publish batcher test go publish error rpc error code failedprecondition desc mockserver got unexpected stream connection want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish b error pubsublite service has stopped or is stopping want partition offset publisher test go stop got err publisher projects locations us b topics my topic partitions rpc error code failedprecondition desc mockserver got unexpected stream connection verifier go call did not receive expected request initial request topic projects locations us b topics my topic partition client id publisher verifier go call did not receive expected request message publish request messages key bar data b first sequence number | 0 |
21,445 | 11,660,236,226 | IssuesEvent | 2020-03-03 02:39:09 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | opened | Signal Timing and Phasing Dashboard | Product: Data & Performance Hub Service: Dev Type: New Application Workgroup: AMD migrated | Description: See also: SR19-104475 .
Brian has requested that we display signal timing and phasing information from D4 on a map/dashboard so that it can be viewed by TMC staff and field technicians.
Definitely need to clarify use case.
Request Date: 2019-03-15 17:01:00
Request ID: DTS19-104480
Status: Backlog
Customer Priority: 4
Level of Effort: Small Project
DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c8c20a2a49e7c3051e72d52
*Migrated from [atd-data-and-performance #268](https://github.com/cityofaustin/atd-data-and-performance/issues/268)* | 1.0 | Signal Timing and Phasing Dashboard - Description: See also: SR19-104475 .
Brian has requested that we display signal timing and phasing information from D4 on a map/dashboard so that it can be viewed by TMC staff and field technicians.
Definitely need to clarify use case.
Request Date: 2019-03-15 17:01:00
Request ID: DTS19-104480
Status: Backlog
Customer Priority: 4
Level of Effort: Small Project
DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c8c20a2a49e7c3051e72d52
*Migrated from [atd-data-and-performance #268](https://github.com/cityofaustin/atd-data-and-performance/issues/268)* | non_defect | signal timing and phasing dashboard description see also brian has requested that we display signal timing and phasing information from on a map dashboard so that it can be viewed by tmc staff and field technicians definitely need to clarify use case request date request id status backlog customer priority level of effort small project dts url migrated from | 0 |
19,348 | 3,193,184,581 | IssuesEvent | 2015-09-30 02:32:00 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | dart2js: assert_with_message_test fails in minified code | Area-Dart2JS Priority-High Triaged Type-Defect | The minified interceptor name for JSNull is being rejected, leading to the evaluation of null.constructor.
This happens in the formatting of the TypeError for a failed boolean conversion for the case when the assertion expression evaluates to non-bool.
The null deref is trapped an converted to a NullError.
Seems related to: https://github.com/dart-lang/sdk/issues/19137
----
@@@BUILD_STEP dart2js_extra-drt tests --minified --dart2js-batch --csp --checked failures@@@
FAILED: dart2js-drt-checked release_ia32 dart2js_extra/assert_with_message_test
Expected: Pass
Actual: RuntimeError
CommandOutput[dart2js]:
CommandOutput[content_shell]:
stdout:
#READY
CONSOLE MESSAGE: line 78: {
"type": "debug",
"value": "test_controller.js started",
"timestamp": 0
}
CONSOLE MESSAGE: line 78: {
"type": "print",
"value": "dart-calling-main",
"timestamp": 0.064
}
CONSOLE MESSAGE: line 78: {
"type": "sync_exception",
"value": "Exception: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\nStack: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\n at dart.d (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:682:3)\n at dart.aB (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1510:24)\n at dart.aC (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1509:3)\n at ak.dart.ak.$2 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1295:3)\n at dart.c5 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1242:3)\n at dart.bw (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1266:3)\n at dartMainRunner (http://127.0.0.1:41436/root_dart/tools/testing/dart/test_controller.js:311:5)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1638:55\n at init.currentScript (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1634:55)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1637:84",
"timestamp": 0.092
} | 1.0 | dart2js: assert_with_message_test fails in minified code - The minified interceptor name for JSNull is being rejected, leading to the evaluation of null.constructor.
This happens in the formatting of the TypeError for a failed boolean conversion for the case when the assertion expression evaluates to non-bool.
The null deref is trapped an converted to a NullError.
Seems related to: https://github.com/dart-lang/sdk/issues/19137
----
@@@BUILD_STEP dart2js_extra-drt tests --minified --dart2js-batch --csp --checked failures@@@
FAILED: dart2js-drt-checked release_ia32 dart2js_extra/assert_with_message_test
Expected: Pass
Actual: RuntimeError
CommandOutput[dart2js]:
CommandOutput[content_shell]:
stdout:
#READY
CONSOLE MESSAGE: line 78: {
"type": "debug",
"value": "test_controller.js started",
"timestamp": 0
}
CONSOLE MESSAGE: line 78: {
"type": "print",
"value": "dart-calling-main",
"timestamp": 0.064
}
CONSOLE MESSAGE: line 78: {
"type": "sync_exception",
"value": "Exception: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\nStack: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\n at dart.d (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:682:3)\n at dart.aB (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1510:24)\n at dart.aC (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1509:3)\n at ak.dart.ak.$2 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1295:3)\n at dart.c5 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1242:3)\n at dart.bw (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1266:3)\n at dartMainRunner (http://127.0.0.1:41436/root_dart/tools/testing/dart/test_controller.js:311:5)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1638:55\n at init.currentScript (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1634:55)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1637:84",
"timestamp": 0.092
} | defect | assert with message test fails in minified code the minified interceptor name for jsnull is being rejected leading to the evaluation of null constructor this happens in the formatting of the typeerror for a failed boolean conversion for the case when the assertion expression evaluates to non bool the null deref is trapped an converted to a nullerror seems related to build step extra drt tests minified batch csp checked failures failed drt checked release extra assert with message test expected pass actual runtimeerror commandoutput commandoutput stdout ready console message line type debug value test controller js started timestamp console message line type print value dart calling main timestamp console message line type sync exception value exception expect istrue false name f is typeerror fails nstack expect istrue false name f is typeerror fails n at dart d at dart ab at dart ac at ak dart ak at dart at dart bw at dartmainrunner at at init currentscript at timestamp | 1 |
59,583 | 17,023,168,320 | IssuesEvent | 2021-07-03 00:40:50 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | cannot access account settings | Component: website Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 3.38pm, Saturday, 26th May 2007]**
Whenever I try to access my account settings at http://www.openstreetmap.org/user/Marc+Sch%FCtz/account, I get the following error:
Application error
Change this error message for exceptions thrown outside of an action (like in Dispatcher setups or broken Ruby code) in public/500.html
Maybe this is related to the %FC in the URL? | 1.0 | cannot access account settings - **[Submitted to the original trac issue database at 3.38pm, Saturday, 26th May 2007]**
Whenever I try to access my account settings at http://www.openstreetmap.org/user/Marc+Sch%FCtz/account, I get the following error:
Application error
Change this error message for exceptions thrown outside of an action (like in Dispatcher setups or broken Ruby code) in public/500.html
Maybe this is related to the %FC in the URL? | defect | cannot access account settings whenever i try to access my account settings at i get the following error application error change this error message for exceptions thrown outside of an action like in dispatcher setups or broken ruby code in public html maybe this is related to the fc in the url | 1 |
335,118 | 24,454,563,436 | IssuesEvent | 2022-10-07 04:53:52 | justicehub-in/justice-hub-docs | https://api.github.com/repos/justicehub-in/justice-hub-docs | closed | Data points to curate | documentation data-request | - [ ] Parliament session wise questions for Law and Justice
- [ ] High Court data released by Daksh | 1.0 | Data points to curate - - [ ] Parliament session wise questions for Law and Justice
- [ ] High Court data released by Daksh | non_defect | data points to curate parliament session wise questions for law and justice high court data released by daksh | 0 |
69,135 | 22,199,980,864 | IssuesEvent | 2022-06-07 10:16:55 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Mismatch between reactions on thread root between main timeline and thread view | T-Defect S-Major A-Reactions O-Occasional A-Threads | ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
Reactions to be the same
#### What happened instead?
Fewer reactions on the main list

### Operating system
Ubuntu
### Browser information
Chrome
### URL for webapp
develop.element.io
### Application version
_No response_
### Homeserver
jki.re
### Will you send logs?
No | 1.0 | Mismatch between reactions on thread root between main timeline and thread view - ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
Reactions to be the same
#### What happened instead?
Fewer reactions on the main list

### Operating system
Ubuntu
### Browser information
Chrome
### URL for webapp
develop.element.io
### Application version
_No response_
### Homeserver
jki.re
### Will you send logs?
No | defect | mismatch between reactions on thread root between main timeline and thread view steps to reproduce where are you starting what can you see what do you click more steps… outcome what did you expect reactions to be the same what happened instead fewer reactions on the main list operating system ubuntu browser information chrome url for webapp develop element io application version no response homeserver jki re will you send logs no | 1 |
5,721 | 2,610,214,096 | IssuesEvent | 2015-02-26 19:08:21 | chrsmith/somefinders | https://api.github.com/repos/chrsmith/somefinders | opened | yoksel dll | auto-migrated Priority-Medium Type-Defect | ```
'''Владелин Лапин'''
День добрый никак не могу найти .yoksel dll. где
то видел уже
'''Гарибальди Анисимов'''
Качай тут http://bit.ly/16sJ2P3
'''Афиноген Рожков'''
Спасибо вроде то но просит телефон вводить
'''Гелеон Соколов'''
Не это не влияет на баланс
'''Аркадий Русаков'''
Неа все ок у меня ничего не списало
Информация о файле: yoksel dll
Загружен: В этом месяце
Скачан раз: 234
Рейтинг: 1475
Средняя скорость скачивания: 181
Похожих файлов: 15
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 12:25 | 1.0 | yoksel dll - ```
'''Владелин Лапин'''
День добрый никак не могу найти .yoksel dll. где
то видел уже
'''Гарибальди Анисимов'''
Качай тут http://bit.ly/16sJ2P3
'''Афиноген Рожков'''
Спасибо вроде то но просит телефон вводить
'''Гелеон Соколов'''
Не это не влияет на баланс
'''Аркадий Русаков'''
Неа все ок у меня ничего не списало
Информация о файле: yoksel dll
Загружен: В этом месяце
Скачан раз: 234
Рейтинг: 1475
Средняя скорость скачивания: 181
Похожих файлов: 15
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 12:25 | defect | yoksel dll владелин лапин день добрый никак не могу найти yoksel dll где то видел уже гарибальди анисимов качай тут афиноген рожков спасибо вроде то но просит телефон вводить гелеон соколов не это не влияет на баланс аркадий русаков неа все ок у меня ничего не списало информация о файле yoksel dll загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at | 1 |
12,095 | 5,145,208,685 | IssuesEvent | 2017-01-12 20:54:32 | dotnet/cli | https://api.github.com/repos/dotnet/cli | closed | Razor not copied to folder on change | msbuild-mitigated | When I change only a razor (cshtml) page and build/rebuild the solution I do not see the changes applied, files dont get copied to build folder. need to remove bin folder or change a controller to copy all new razor content.
## Steps to reproduce
modify a razor page (.cshtml)
## Expected behaviour
always copy content files even if they are not changed
## Actual behaviour
Does not copy Views unless a .cs file has been modified.
## Environment data
`dotnet --info` output:
Product Information:
Version: 1.0.0-preview1-002700
Commit Sha: 0336f6bb34
Runtime Environment:
OS Name: Mac OS X
OS Version: 10.11
OS Platform: Darwin
RID: osx.10.11-x64
| 1.0 | Razor not copied to folder on change - When I change only a razor (cshtml) page and build/rebuild the solution I do not see the changes applied, files dont get copied to build folder. need to remove bin folder or change a controller to copy all new razor content.
## Steps to reproduce
modify a razor page (.cshtml)
## Expected behaviour
always copy content files even if they are not changed
## Actual behaviour
Does not copy Views unless a .cs file has been modified.
## Environment data
`dotnet --info` output:
Product Information:
Version: 1.0.0-preview1-002700
Commit Sha: 0336f6bb34
Runtime Environment:
OS Name: Mac OS X
OS Version: 10.11
OS Platform: Darwin
RID: osx.10.11-x64
| non_defect | razor not copied to folder on change when i change only a razor cshtml page and build rebuild the solution i do not see the changes applied files dont get copied to build folder need to remove bin folder or change a controller to copy all new razor content steps to reproduce modify a razor page cshtml expected behaviour always copy content files even if they are not changed actual behaviour does not copy views unless a cs file has been modified environment data dotnet info output product information version commit sha runtime environment os name mac os x os version os platform darwin rid osx | 0 |
59,436 | 24,767,496,805 | IssuesEvent | 2022-10-22 18:16:22 | openkfw/TruBudget | https://api.github.com/repos/openkfw/TruBudget | closed | Improve notification link to item | bug wontfix blockchain api frontend storage-service | - [ ] I have searched the [issues](https://github.com/openkfw/TruBudget/issues) of this repository and believe that this is not a duplicate.
## Description 😯
Problem: when a user clicks on a notification linked to an WFI in trubudget, he is directed to the "project" level. He does not know therefore in some component is the WFI.
Need: When users click on a notification, trubudget should take them to the relevant level in the user interface. Thus, a notification concerning an WFI must lead the user who clicks on this notification to the WFI concerned.
### How to reproduce 🕹
## Your Environment 🌎
| Tech | Version |
| --------------- | ------- |
| Frontend | v1.?.? |
| API | v1.?.? |
| Blockchain | v1.?.? |
| Export-Service | v1.?.? |
| Email-Service | v1.?.? |
| Storage-Service | v1.?.? |
| 1.0 | Improve notification link to item - - [ ] I have searched the [issues](https://github.com/openkfw/TruBudget/issues) of this repository and believe that this is not a duplicate.
## Description 😯
Problem: when a user clicks on a notification linked to an WFI in trubudget, he is directed to the "project" level. He does not know therefore in some component is the WFI.
Need: When users click on a notification, trubudget should take them to the relevant level in the user interface. Thus, a notification concerning an WFI must lead the user who clicks on this notification to the WFI concerned.
### How to reproduce 🕹
## Your Environment 🌎
| Tech | Version |
| --------------- | ------- |
| Frontend | v1.?.? |
| API | v1.?.? |
| Blockchain | v1.?.? |
| Export-Service | v1.?.? |
| Email-Service | v1.?.? |
| Storage-Service | v1.?.? |
| non_defect | improve notification link to item i have searched the of this repository and believe that this is not a duplicate description 😯 problem when a user clicks on a notification linked to an wfi in trubudget he is directed to the project level he does not know therefore in some component is the wfi need when users click on a notification trubudget should take them to the relevant level in the user interface thus a notification concerning an wfi must lead the user who clicks on this notification to the wfi concerned how to reproduce 🕹 your environment 🌎 tech version frontend api blockchain export service email service storage service | 0 |
814,543 | 30,511,454,531 | IssuesEvent | 2023-07-18 21:12:58 | DiscoTrayStudios/hendrix-today | https://api.github.com/repos/DiscoTrayStudios/hendrix-today | closed | Hendrix Today Uploader - CI/CD with GitHub Actions | category: backend priority: medium size: medium type: incomplete | **Describe the incomplete feature**
Building and deployment of the Hendrix Today Uploader is not automated.
**Describe the solution you'd like**
The uploader should be rebuilt and re-deployed on each push to `main`.
| 1.0 | Hendrix Today Uploader - CI/CD with GitHub Actions - **Describe the incomplete feature**
Building and deployment of the Hendrix Today Uploader is not automated.
**Describe the solution you'd like**
The uploader should be rebuilt and re-deployed on each push to `main`.
| non_defect | hendrix today uploader ci cd with github actions describe the incomplete feature building and deployment of the hendrix today uploader is not automated describe the solution you d like the uploader should be rebuilt and re deployed on each push to main | 0 |
76,669 | 26,545,312,014 | IssuesEvent | 2023-01-19 23:20:51 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | opened | Preview button not working for some Lovell nodes | Defect ⭐️ Facilities Needs refining | ## Describe the defect
As a Lovell editor, I want to be able to preview content that is in the Lovell-VA or Lovell-TRICARE sections.
This is currently broken for VAMC Detail Pages. The Preview button works if they are in the Lovell Federal section, but not if they are in the Lovell-VA or Lovell-TRICARE sections. I created a test page to investigate the cases in which this is true. I have not investigated other content types.
## To Reproduce
Steps to reproduce the behavior:
1. Go to [Revision log for this test page](https://prod.cms.va.gov/node/52805/revisions)
2. Revert to Revision #743400 or Revision #743399 (Section = Lovell - VA or TRICARE)
3. Confirm that Preview button leads to Internal Server Error
4. Revert to Revision #743398 (Section = Lovell Federal health care)
4. Confirm that Preview button magically works again
## AC / Expected behavior
The Preview button should work consistently for content types regardless of which Lovell section they are in.
## Screenshots



| 1.0 | Preview button not working for some Lovell nodes - ## Describe the defect
As a Lovell editor, I want to be able to preview content that is in the Lovell-VA or Lovell-TRICARE sections.
This is currently broken for VAMC Detail Pages. The Preview button works if they are in the Lovell Federal section, but not if they are in the Lovell-VA or Lovell-TRICARE sections. I created a test page to investigate the cases in which this is true. I have not investigated other content types.
## To Reproduce
Steps to reproduce the behavior:
1. Go to [Revision log for this test page](https://prod.cms.va.gov/node/52805/revisions)
2. Revert to Revision #743400 or Revision #743399 (Section = Lovell - VA or TRICARE)
3. Confirm that Preview button leads to Internal Server Error
4. Revert to Revision #743398 (Section = Lovell Federal health care)
4. Confirm that Preview button magically works again
## AC / Expected behavior
The Preview button should work consistently for content types regardless of which Lovell section they are in.
## Screenshots



| defect | preview button not working for some lovell nodes describe the defect as a lovell editor i want to be able to preview content that is in the lovell va or lovell tricare sections this is currently broken for vamc detail pages the preview button works if they are in the lovell federal section but not if they are in the lovell va or lovell tricare sections i created a test page to investigate the cases in which this is true i have not investigated other content types to reproduce steps to reproduce the behavior go to revert to revision or revision section lovell va or tricare confirm that preview button leads to internal server error revert to revision section lovell federal health care confirm that preview button magically works again ac expected behavior the preview button should work consistently for content types regardless of which lovell section they are in screenshots | 1 |
34,234 | 7,431,748,273 | IssuesEvent | 2018-03-25 17:43:29 | Yahkal/replicaisland | https://api.github.com/repos/Yahkal/replicaisland | closed | "Google Code will be turning read-only on August 25th. See this post for more information." | Priority-Medium Type-Defect auto-migrated | ```
https://code.google.com/p/support/wiki/ReadOnlyTransition
I hope you're moving this repository someplace active.…
```
Original issue reported on code.google.com by `TPSam...@gmail.com` on 2 Aug 2015 at 4:10
| 1.0 | "Google Code will be turning read-only on August 25th. See this post for more information." - ```
https://code.google.com/p/support/wiki/ReadOnlyTransition
I hope you're moving this repository someplace active.…
```
Original issue reported on code.google.com by `TPSam...@gmail.com` on 2 Aug 2015 at 4:10
| defect | google code will be turning read only on august see this post for more information i hope you re moving this repository someplace active … original issue reported on code google com by tpsam gmail com on aug at | 1 |
701,144 | 24,088,097,498 | IssuesEvent | 2022-09-19 12:44:53 | AY2223S1-CS2103T-T09-4/tp | https://api.github.com/repos/AY2223S1-CS2103T-T09-4/tp | closed | Add student and Next of Kin phone number fields | priority.HIGH type.Task type.Task.Add | Phone number will be of 8 digits in length, starting with 8 or 9. | 1.0 | Add student and Next of Kin phone number fields - Phone number will be of 8 digits in length, starting with 8 or 9. | non_defect | add student and next of kin phone number fields phone number will be of digits in length starting with or | 0 |
60,365 | 17,023,406,747 | IssuesEvent | 2021-07-03 01:52:05 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Potlatch not saving comments from offline save | Component: potlatch (flash editor) Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 6.15pm, Saturday, 23rd May 2009]**
When doing a save in offline mode, Potlatch prompts for a comment, but this doesn't get used as the comment on the changeset (and the changeset doesn't get closed out, but perhaps that is expected behaviour). | 1.0 | Potlatch not saving comments from offline save - **[Submitted to the original trac issue database at 6.15pm, Saturday, 23rd May 2009]**
When doing a save in offline mode, Potlatch prompts for a comment, but this doesn't get used as the comment on the changeset (and the changeset doesn't get closed out, but perhaps that is expected behaviour). | defect | potlatch not saving comments from offline save when doing a save in offline mode potlatch prompts for a comment but this doesn t get used as the comment on the changeset and the changeset doesn t get closed out but perhaps that is expected behaviour | 1 |
596,958 | 18,151,489,119 | IssuesEvent | 2021-09-26 10:44:22 | GoogleCloudPlatform/python-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples | opened | people-and-planet-ai.image-classification.e2e_test: test_predict failed | priority: p1 type: bug flakybot: issue | Note: #6463 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 35873ab69eaf0f2a6dd4a7007a12917265e2b848
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8), [Sponge](http://sponge2/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 67, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "{"created":"@1632652201.879730366","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Deadline Exceeded","grpc_status":4}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/people-and-planet-ai/image-classification/e2e_test.py", line 112, in model_endpoint_id
PROJECT, REGION, MODEL_PATH, MODEL_ENDPOINT, endpoint_id
File "/workspace/people-and-planet-ai/image-classification/deploy_model.py", line 80, in deploy_model
deployed_model = response.result()
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 130, in result
self._blocking_poll(timeout=timeout, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 108, in _blocking_poll
retry_(self._done_or_raise)(**kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func
on_error=on_error,
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target
return target()
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 86, in _done_or_raise
if not self.done(**kwargs):
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 170, in done
self._refresh_and_update(retry)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 158, in _refresh_and_update
self._operation = self._refresh(retry=retry)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operations_v1/operations_client.py", line 143, in get_operation
request, retry=retry, timeout=timeout, metadata=metadata
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func
on_error=on_error,
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target
return target()
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout
return func(*args, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 69, in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
File "<string>", line 3, in raise_from
google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded</pre></details> | 1.0 | people-and-planet-ai.image-classification.e2e_test: test_predict failed - Note: #6463 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 35873ab69eaf0f2a6dd4a7007a12917265e2b848
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8), [Sponge](http://sponge2/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 67, in error_remapped_callable
return callable_(*args, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__
return _end_unary_response_blocking(state, call, False, None)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking
raise _InactiveRpcError(state)
grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
status = StatusCode.DEADLINE_EXCEEDED
details = "Deadline Exceeded"
debug_error_string = "{"created":"@1632652201.879730366","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Deadline Exceeded","grpc_status":4}"
>
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/workspace/people-and-planet-ai/image-classification/e2e_test.py", line 112, in model_endpoint_id
PROJECT, REGION, MODEL_PATH, MODEL_ENDPOINT, endpoint_id
File "/workspace/people-and-planet-ai/image-classification/deploy_model.py", line 80, in deploy_model
deployed_model = response.result()
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 130, in result
self._blocking_poll(timeout=timeout, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 108, in _blocking_poll
retry_(self._done_or_raise)(**kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func
on_error=on_error,
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target
return target()
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 86, in _done_or_raise
if not self.done(**kwargs):
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 170, in done
self._refresh_and_update(retry)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 158, in _refresh_and_update
self._operation = self._refresh(retry=retry)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operations_v1/operations_client.py", line 143, in get_operation
request, retry=retry, timeout=timeout, metadata=metadata
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__
return wrapped_func(*args, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func
on_error=on_error,
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target
return target()
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout
return func(*args, **kwargs)
File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 69, in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
File "<string>", line 3, in raise_from
google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded</pre></details> | non_defect | people and planet ai image classification test test predict failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace people and planet ai image classification nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace people and planet ai image classification nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace people and planet ai image classification nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status the above exception was the direct cause of the following exception traceback most recent call last file workspace people and planet ai image classification test py line in model endpoint id project region model path model endpoint endpoint id file workspace people and planet ai image classification deploy model py line in deploy model deployed model response result file workspace people and planet ai image classification nox py lib site packages google api core future polling py line in result self blocking poll timeout timeout kwargs file workspace people and planet ai image classification nox py lib site packages google api core future polling py line in blocking poll retry self done or raise kwargs file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry target return target file workspace people and planet ai image classification nox py lib site packages google api core future polling py line in done or raise if not self done kwargs file workspace people and planet ai image classification nox py lib site packages google api core operation py line in done self refresh and update retry file workspace people and planet ai image classification nox py lib site packages google api core operation py line in refresh and update self operation self refresh retry retry file workspace people and planet ai image classification nox py lib site packages google api core operations operations client py line in get operation request retry retry timeout timeout metadata metadata file workspace people and planet ai image classification nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry target return target file workspace people and planet ai image classification nox py lib site packages google api core timeout py line in func with timeout return func args kwargs file workspace people and planet ai image classification nox py lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions deadlineexceeded deadline exceeded | 0 |
23,063 | 3,756,128,559 | IssuesEvent | 2016-03-13 04:20:35 | StarsOCV/eve-tf2hud | https://api.github.com/repos/StarsOCV/eve-tf2hud | closed | I broke it didnt I. TF2 Update. | auto-migrated OpSys-All OpSys-Windows Priority-Medium Type-Defect update Usability | ```
Okay so here it is. Tf2 update. It updates i try to start game, i get disk
write error and it doesnt start. I try opening up eve hud updater to uninstall
the hud to see if that's the problem and i get this. See attached picture.
```
Original issue reported on code.google.com by `byron.ar...@gmail.com` on 20 Dec 2013 at 6:38
Attachments:
* [i broke it.png](https://storage.googleapis.com/google-code-attachments/eve-tf2hud/issue-101/comment-0/i broke it.png)
| 1.0 | I broke it didnt I. TF2 Update. - ```
Okay so here it is. Tf2 update. It updates i try to start game, i get disk
write error and it doesnt start. I try opening up eve hud updater to uninstall
the hud to see if that's the problem and i get this. See attached picture.
```
Original issue reported on code.google.com by `byron.ar...@gmail.com` on 20 Dec 2013 at 6:38
Attachments:
* [i broke it.png](https://storage.googleapis.com/google-code-attachments/eve-tf2hud/issue-101/comment-0/i broke it.png)
| defect | i broke it didnt i update okay so here it is update it updates i try to start game i get disk write error and it doesnt start i try opening up eve hud updater to uninstall the hud to see if that s the problem and i get this see attached picture original issue reported on code google com by byron ar gmail com on dec at attachments broke it png | 1 |
46,286 | 13,055,885,660 | IssuesEvent | 2020-07-30 03:01:25 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting (Trac #926) | Incomplete Migration Migrated from Trac combo reconstruction defect | Migrated from https://code.icecube.wisc.edu/ticket/926
```json
{
"status": "closed",
"changetime": "2015-05-12T14:06:42",
"description": "there is no check of the particle-shape in steamshovel/trunk/private/shovelart/artists/MCTreeArtist.cpp\n\nneeds be fixed, so that neutrinos are actually plotted as incoming particles!",
"reporter": "mzoll",
"cc": "",
"resolution": "fixed",
"_ts": "1431439602511916",
"component": "combo reconstruction",
"summary": "Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting",
"priority": "normal",
"keywords": "",
"time": "2015-04-10T16:15:03",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
| 1.0 | Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting (Trac #926) - Migrated from https://code.icecube.wisc.edu/ticket/926
```json
{
"status": "closed",
"changetime": "2015-05-12T14:06:42",
"description": "there is no check of the particle-shape in steamshovel/trunk/private/shovelart/artists/MCTreeArtist.cpp\n\nneeds be fixed, so that neutrinos are actually plotted as incoming particles!",
"reporter": "mzoll",
"cc": "",
"resolution": "fixed",
"_ts": "1431439602511916",
"component": "combo reconstruction",
"summary": "Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting",
"priority": "normal",
"keywords": "",
"time": "2015-04-10T16:15:03",
"milestone": "",
"owner": "hdembinski",
"type": "defect"
}
```
| defect | steamshovel wronly plots stopping tracks as starting trac migrated from json status closed changetime description there is no check of the particle shape in steamshovel trunk private shovelart artists mctreeartist cpp n nneeds be fixed so that neutrinos are actually plotted as incoming particles reporter mzoll cc resolution fixed ts component combo reconstruction summary steamshovel wronly plots stopping tracks as starting priority normal keywords time milestone owner hdembinski type defect | 1 |
61,380 | 17,023,680,032 | IssuesEvent | 2021-07-03 03:15:59 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | addr:street is ignored | Component: nominatim Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 9.25am, Monday, 7th February 2011]**
I perform search:
For "18, . , " (sorry for Russian text).
The result is http://goo.gl/LB3yG
It returns wrong building: http://www.openstreetmap.org/browse/way/38535341
addr:housenumber: 18
addr:street: .
building: yes
Expected result is:
http://www.openstreetmap.org/browse/way/38323699
addr:housenumber: 18
addr:street: .
building: yes
It seems the nearest calculation is used instead of addr:street. | 1.0 | addr:street is ignored - **[Submitted to the original trac issue database at 9.25am, Monday, 7th February 2011]**
I perform search:
For "18, . , " (sorry for Russian text).
The result is http://goo.gl/LB3yG
It returns wrong building: http://www.openstreetmap.org/browse/way/38535341
addr:housenumber: 18
addr:street: .
building: yes
Expected result is:
http://www.openstreetmap.org/browse/way/38323699
addr:housenumber: 18
addr:street: .
building: yes
It seems the nearest calculation is used instead of addr:street. | defect | addr street is ignored i perform search for sorry for russian text the result is it returns wrong building addr housenumber addr street building yes expected result is addr housenumber addr street building yes it seems the nearest calculation is used instead of addr street | 1 |
4,939 | 2,891,810,781 | IssuesEvent | 2015-06-15 08:49:23 | LibreCat/Catmandu | https://api.github.com/repos/LibreCat/Catmandu | closed | Extend help command to importers, exporters and fixes | documentation enhancement | The `help` command should also show help about importers, exporter, fixes etc.:
catmandu help export JSON
catmandu help RDF importer
catmandu help add_field
Some general help would also be nice, for instance a short introduction into fix language:
catmandu help fix
Help messages can be generated from module documentation. See https://github.com/LibreCat/Catmandu/tree/dev_help fo a start. | 1.0 | Extend help command to importers, exporters and fixes - The `help` command should also show help about importers, exporter, fixes etc.:
catmandu help export JSON
catmandu help RDF importer
catmandu help add_field
Some general help would also be nice, for instance a short introduction into fix language:
catmandu help fix
Help messages can be generated from module documentation. See https://github.com/LibreCat/Catmandu/tree/dev_help fo a start. | non_defect | extend help command to importers exporters and fixes the help command should also show help about importers exporter fixes etc catmandu help export json catmandu help rdf importer catmandu help add field some general help would also be nice for instance a short introduction into fix language catmandu help fix help messages can be generated from module documentation see fo a start | 0 |
276,436 | 20,983,776,922 | IssuesEvent | 2022-03-28 23:18:10 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | closed | Identifying Profile Page Requirements under User Requirements | priority-high type-documentation status-completed | ## Description
- Profile page requirements for both types of users (students and teachers) will be identified under the title of User Requirements as a part of the Functional Requirements for this project.
- It will include the details for but not limited to
* Profile page creation for students
* Profile page creation for teachers
* Profile page edit for students
* Profile page edit for teachers
* Experience tracking
* Achievements and badges
- This issue is a sub-issue of #16 . It will be linked there to easily access but specific details will be under this issue.
- After the initial phase for the profile page requirements is finished, this issue will be closed and reviewing process will continue under the main issue mentioned above.
**Deadline:** 13.03.2022 - Sunday - 20:00
**Reviewer:** Will be reviewed under the main issue after all user requirements are specified. | 1.0 | Identifying Profile Page Requirements under User Requirements - ## Description
- Profile page requirements for both types of users (students and teachers) will be identified under the title of User Requirements as a part of the Functional Requirements for this project.
- It will include the details for but not limited to
* Profile page creation for students
* Profile page creation for teachers
* Profile page edit for students
* Profile page edit for teachers
* Experience tracking
* Achievements and badges
- This issue is a sub-issue of #16 . It will be linked there to easily access but specific details will be under this issue.
- After the initial phase for the profile page requirements is finished, this issue will be closed and reviewing process will continue under the main issue mentioned above.
**Deadline:** 13.03.2022 - Sunday - 20:00
**Reviewer:** Will be reviewed under the main issue after all user requirements are specified. | non_defect | identifying profile page requirements under user requirements description profile page requirements for both types of users students and teachers will be identified under the title of user requirements as a part of the functional requirements for this project it will include the details for but not limited to profile page creation for students profile page creation for teachers profile page edit for students profile page edit for teachers experience tracking achievements and badges this issue is a sub issue of it will be linked there to easily access but specific details will be under this issue after the initial phase for the profile page requirements is finished this issue will be closed and reviewing process will continue under the main issue mentioned above deadline sunday reviewer will be reviewed under the main issue after all user requirements are specified | 0 |
26,647 | 4,775,714,828 | IssuesEvent | 2016-10-27 11:25:00 | wheeler-microfluidics/microdrop | https://api.github.com/repos/wheeler-microfluidics/microdrop | opened | Add mechanism to add scheduling constraints when running emit_signal (Trac #68) | defect Incomplete Migration microdrop Migrated from Trac | Migrated from http://microfluidics.utoronto.ca/ticket/68
```json
{
"status": "closed",
"changetime": "2014-04-17T19:39:01",
"description": "This ticket takes over the second issue in ticket #21.\n\nThe idea is to provide a way to ensure plugins are processed in the correct order upon an emit_signal() call. ",
"reporter": "cfobel",
"cc": "",
"resolution": "fixed",
"_ts": "1397763541728826",
"component": "microdrop",
"summary": "Add mechanism to add scheduling constraints when running emit_signal",
"priority": "major",
"keywords": "",
"version": "0.1",
"time": "2012-03-14T15:41:14",
"milestone": "Microdrop 1.0",
"owner": "cfobel",
"type": "defect"
}
```
| 1.0 | Add mechanism to add scheduling constraints when running emit_signal (Trac #68) - Migrated from http://microfluidics.utoronto.ca/ticket/68
```json
{
"status": "closed",
"changetime": "2014-04-17T19:39:01",
"description": "This ticket takes over the second issue in ticket #21.\n\nThe idea is to provide a way to ensure plugins are processed in the correct order upon an emit_signal() call. ",
"reporter": "cfobel",
"cc": "",
"resolution": "fixed",
"_ts": "1397763541728826",
"component": "microdrop",
"summary": "Add mechanism to add scheduling constraints when running emit_signal",
"priority": "major",
"keywords": "",
"version": "0.1",
"time": "2012-03-14T15:41:14",
"milestone": "Microdrop 1.0",
"owner": "cfobel",
"type": "defect"
}
```
| defect | add mechanism to add scheduling constraints when running emit signal trac migrated from json status closed changetime description this ticket takes over the second issue in ticket n nthe idea is to provide a way to ensure plugins are processed in the correct order upon an emit signal call reporter cfobel cc resolution fixed ts component microdrop summary add mechanism to add scheduling constraints when running emit signal priority major keywords version time milestone microdrop owner cfobel type defect | 1 |
411,273 | 12,016,238,631 | IssuesEvent | 2020-04-10 15:40:57 | yedhink/covid19-kerala-api | https://api.github.com/repos/yedhink/covid19-kerala-api | closed | Major bug - the pdf has another inconsistency where "on today" in annex1 has prefix | bug high-priority python | Probably changing the regex to just an "in" check should fix this. | 1.0 | Major bug - the pdf has another inconsistency where "on today" in annex1 has prefix - Probably changing the regex to just an "in" check should fix this. | non_defect | major bug the pdf has another inconsistency where on today in has prefix probably changing the regex to just an in check should fix this | 0 |
60,539 | 17,023,451,285 | IssuesEvent | 2021-07-03 02:05:56 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | Changeset Not Being Rendered | Component: mapnik Priority: major Resolution: fixed Type: defect | **[Submitted to the original trac issue database at 4.35pm, Friday, 31st July 2009]**
Take a look at this boundary where a forest and national park meet:
http://osm.org/go/TwUljNo--
Notice that the boundaries don't line up. This is because the national park is in slightly the wrong place. The national park is this changeset uploaded yesterday:
http://www.openstreetmap.org/browse/changeset/1980439
Today I moved the national park into the correct position. The changeset was closed at 31 Jul 00:09:
http://www.openstreetmap.org/browse/changeset/1989864
I then marked the tile you are looking at as dirty. It was apparently rendered by Mapnik on 31 Jul 03:21:
http://a.tile.openstreetmap.org/12/772/1608.png/status
As you can see the data from my new changeset has not been used.
On 31 Jul 01:33 I added a new changeset with some trails:
http://www.openstreetmap.org/browse/changeset/1990063
This was rendered with trails at 31 Jul 03:13:
http://a.tile.openstreetmap.org/13/1567/3318.png/status
If data I uploaded at 01:33 was rendered at 3:13, how come data I uploaded at 00:09 has not been rendered at the time of writing this? (03:21)?
One clue might be that the trails are new data but the movement of nodes was not. Also JOSM gave me an error of "unexpected end of file" when the changeset was closing, but the changeset is listed in my edits as being closed anyway. It also has all 23573 nodes.
I have cleared my browser cache and tried two browsers. | 1.0 | Changeset Not Being Rendered - **[Submitted to the original trac issue database at 4.35pm, Friday, 31st July 2009]**
Take a look at this boundary where a forest and national park meet:
http://osm.org/go/TwUljNo--
Notice that the boundaries don't line up. This is because the national park is in slightly the wrong place. The national park is this changeset uploaded yesterday:
http://www.openstreetmap.org/browse/changeset/1980439
Today I moved the national park into the correct position. The changeset was closed at 31 Jul 00:09:
http://www.openstreetmap.org/browse/changeset/1989864
I then marked the tile you are looking at as dirty. It was apparently rendered by Mapnik on 31 Jul 03:21:
http://a.tile.openstreetmap.org/12/772/1608.png/status
As you can see the data from my new changeset has not been used.
On 31 Jul 01:33 I added a new changeset with some trails:
http://www.openstreetmap.org/browse/changeset/1990063
This was rendered with trails at 31 Jul 03:13:
http://a.tile.openstreetmap.org/13/1567/3318.png/status
If data I uploaded at 01:33 was rendered at 3:13, how come data I uploaded at 00:09 has not been rendered at the time of writing this? (03:21)?
One clue might be that the trails are new data but the movement of nodes was not. Also JOSM gave me an error of "unexpected end of file" when the changeset was closing, but the changeset is listed in my edits as being closed anyway. It also has all 23573 nodes.
I have cleared my browser cache and tried two browsers. | defect | changeset not being rendered take a look at this boundary where a forest and national park meet notice that the boundaries don t line up this is because the national park is in slightly the wrong place the national park is this changeset uploaded yesterday today i moved the national park into the correct position the changeset was closed at jul i then marked the tile you are looking at as dirty it was apparently rendered by mapnik on jul as you can see the data from my new changeset has not been used on jul i added a new changeset with some trails this was rendered with trails at jul if data i uploaded at was rendered at how come data i uploaded at has not been rendered at the time of writing this one clue might be that the trails are new data but the movement of nodes was not also josm gave me an error of unexpected end of file when the changeset was closing but the changeset is listed in my edits as being closed anyway it also has all nodes i have cleared my browser cache and tried two browsers | 1 |
278,064 | 21,058,031,895 | IssuesEvent | 2022-04-01 06:38:51 | medajet/ped | https://api.github.com/repos/medajet/ped | opened | "upt" command has incorrect format in UG | severity.Medium type.DocumentationBug | For the updt command, both fields are listed as optional even though at least 1 is mandatory.

<!--session: 1648792880809-a0d404c0-a5ce-4319-b3e0-6ef579fd4865-->
<!--Version: Web v3.4.2--> | 1.0 | "upt" command has incorrect format in UG - For the updt command, both fields are listed as optional even though at least 1 is mandatory.

<!--session: 1648792880809-a0d404c0-a5ce-4319-b3e0-6ef579fd4865-->
<!--Version: Web v3.4.2--> | non_defect | upt command has incorrect format in ug for the updt command both fields are listed as optional even though at least is mandatory | 0 |
52,837 | 13,225,122,571 | IssuesEvent | 2020-08-17 20:32:11 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | Healpix port does not build shared libraries (Trac #438) | Migrated from Trac defect other | I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/438">https://code.icecube.wisc.edu/projects/icecube/ticket/438</a>, reported by gladstoneand owned by dladieu</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:59:20",
"_ts": "1351706360000000",
"description": "I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray. ",
"reporter": "gladstone",
"cc": "gladstone@icecube.wisc.edu",
"resolution": "fixed",
"time": "2012-08-15T23:13:51",
"component": "other",
"summary": "Healpix port does not build shared libraries",
"priority": "normal",
"keywords": "ports",
"milestone": "",
"owner": "dladieu",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Healpix port does not build shared libraries (Trac #438) - I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/438">https://code.icecube.wisc.edu/projects/icecube/ticket/438</a>, reported by gladstoneand owned by dladieu</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:59:20",
"_ts": "1351706360000000",
"description": "I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray. ",
"reporter": "gladstone",
"cc": "gladstone@icecube.wisc.edu",
"resolution": "fixed",
"time": "2012-08-15T23:13:51",
"component": "other",
"summary": "Healpix port does not build shared libraries",
"priority": "normal",
"keywords": "ports",
"milestone": "",
"owner": "dladieu",
"type": "defect"
}
```
</p>
</details>
| defect | healpix port does not build shared libraries trac i m trying to use the healpix port instead of the healpix cxx project within icerec and the port hasn t built properly i m working on net local in madison it s only built static libraries so i can t use it to link to shared libraries in icetray migrated from json status closed changetime ts description i m trying to use the healpix port instead of the healpix cxx project within icerec and the port hasn t built properly i m working on net local in madison it s only built static libraries so i can t use it to link to shared libraries in icetray reporter gladstone cc gladstone icecube wisc edu resolution fixed time component other summary healpix port does not build shared libraries priority normal keywords ports milestone owner dladieu type defect | 1 |
72,628 | 24,207,431,154 | IssuesEvent | 2022-09-25 12:38:23 | bcpierce00/unison | https://api.github.com/repos/bcpierce00/unison | opened | Backups locations are mixed for backupcurr and backup, and shared across replicas (central) | defect wontfix impact-low effort-medium | As discussed on theunison-users list in 2022-09, `backupcurr` (which is not really backup, but merge pre-images) and `backup` store data in the same place. However, they have different rules.
Related, central storage blurs copies across replicas, but each of these is a per-replica concept.
To resolve:
1) Central storage should somehow be per-replica
2) backupcurr and backup data should not collide
3) The manual should advise maintaining this separation.
| 1.0 | Backups locations are mixed for backupcurr and backup, and shared across replicas (central) - As discussed on theunison-users list in 2022-09, `backupcurr` (which is not really backup, but merge pre-images) and `backup` store data in the same place. However, they have different rules.
Related, central storage blurs copies across replicas, but each of these is a per-replica concept.
To resolve:
1) Central storage should somehow be per-replica
2) backupcurr and backup data should not collide
3) The manual should advise maintaining this separation.
| defect | backups locations are mixed for backupcurr and backup and shared across replicas central as discussed on theunison users list in backupcurr which is not really backup but merge pre images and backup store data in the same place however they have different rules related central storage blurs copies across replicas but each of these is a per replica concept to resolve central storage should somehow be per replica backupcurr and backup data should not collide the manual should advise maintaining this separation | 1 |
54,996 | 14,117,132,710 | IssuesEvent | 2020-11-08 07:32:49 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | Oops, Unable to handle kernel paging request in scan_prefetch_queue_compare | Status: Triage Needed Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu (armbian)
Distribution Version | Focal
Linux Kernel | 5.8.17
Architecture | amd64
ZFS Version | 2.0.0-rc5
SPL Version | 2.0.0-rc5
### Describe the problem you're observing
zpool + zfs hang, seemingly uninterruptible, high load, Oops in dmesg
### Describe how to reproduce the problem
`zpool import tank0`
### Include any warning/errors/backtraces from the system logs
```
[ 506.332804] Unable to handle kernel paging request at virtual address fffe0000c4f5f316
[ 506.333567] Mem abort info:
[ 506.333815] ESR = 0x96000004
[ 506.334107] EC = 0x25: DABT (current EL), IL = 32 bits
[ 506.334581] SET = 0, FnV = 0
[ 506.334921] EA = 0, S1PTW = 0
[ 506.335198] Data abort info:
[ 506.335476] ISV = 0, ISS = 0x00000004
[ 506.335813] CM = 0, WnR = 0
[ 506.336076] [fffe0000c4f5f316] address between user and kernel address ranges
[ 506.336702] Internal error: Oops: 96000004 [#1] PREEMPT SMP
[ 506.337190] Modules linked in: xt_conntrack xt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo xt_addrtype iptable_filter iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 bpfilter br_netfilter bridge rfkill governor_performance zfs(POE) zunicode(POE) zzstd(OE) zlua(OE) zcommon(POE) znvpair(POE) zavl(POE) icp(POE) spl(OE) r8152 snd_soc_hdmi_codec leds_pwm pwm_fan gpio_charger panfrost gpu_sched snd_soc_rockchip_i2s snd_soc_core snd_pcm_dmaengine snd_pcm snd_timer hantro_vpu(C) snd videobuf2_vmalloc rockchip_vdec(C) fusb30x(C) rockchip_rga v4l2_h264 videobuf2_dma_contig soundcore v4l2_mem2mem videobuf2_dma_sg videobuf2_memops videobuf2_v4l2 videobuf2_common videodev mc zstd sg gpio_beeper cpufreq_dt zram sch_fq_codel lm75 ip_tables x_tables autofs4 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx raid1 raid0 multipath linear md_mod realtek rockchipdrm analogix_dp dw_hdmi dw_mipi_dsi drm_kms_helper cec rc_core dwmac_rk stmmac_platform drm
[ 506.337260] stmmac mdio_xpcs drm_panel_orientation_quirks adc_keys
[ 506.345397] CPU: 5 PID: 3192 Comm: z_rd_int Tainted: P C OE 5.8.17-rockchip64 #20.08.21
[ 506.346186] Hardware name: Helios64 (DT)
[ 506.346533] pstate: 80000005 (Nzcv daif -PAN -UAO BTYPE=--)
[ 506.347191] pc : scan_prefetch_queue_compare+0x20/0x40 [zfs]
[ 506.347694] lr : avl_find+0x68/0xc8 [zavl]
[ 506.348054] sp : ffff80001546baa0
[ 506.348347] x29: ffff80001546baa0 x28: ffff0000c207afa8
[ 506.348814] x27: ffff0000c316c9a8 x26: ffff0000c316c9c8
[ 506.349280] x25: ffff0000c30e5090 x24: 0000000000000000
[ 506.349746] x23: ffff80001546bb60 x22: ffff8000092b10d8
[ 506.350212] x21: ffff0000be677000 x20: ffff0000be57f200
[ 506.350677] x19: ffff0000be57f200 x18: 001c050ea761f9c0
[ 506.351143] x17: 00000445581d3e8e x16: 00001510366b591d
[ 506.351609] x15: 0000000000000000 x14: 0000000000000000
[ 506.352074] x13: 0000000000000008 x12: ffff0000f6fa9328
[ 506.352541] x11: 0000000000000020 x10: 0000000000000000
[ 506.353006] x9 : 000000000001e067 x8 : 00000000003c0ce0
[ 506.353472] x7 : 00000000003e2880 x6 : 0000000080000000
[ 506.353938] x5 : ffff0000be57f2a0 x4 : ffff0000be6770a0
[ 506.354404] x3 : 0000000000000011 x2 : ffff0000c4f5f300
[ 506.354870] x1 : ffff0000be57f200 x0 : ffff0000c4f5f300
[ 506.355337] Call trace:
[ 506.355661] scan_prefetch_queue_compare+0x20/0x40 [zfs]
[ 506.356132] avl_find+0x68/0xc8 [zavl]
[ 506.356553] dsl_scan_prefetch+0x1a8/0x228 [zfs]
[ 506.357042] dsl_scan_prefetch_cb+0x108/0x268 [zfs]
[ 506.357553] arc_read_done+0x20c/0x3f8 [zfs]
[ 506.358011] zio_done+0x254/0xd40 [zfs]
[ 506.358433] zio_execute+0xac/0x110 [zfs]
[ 506.358798] taskq_thread+0x298/0x440 [spl]
[ 506.359175] kthread+0x118/0x150
[ 506.359462] ret_from_fork+0x10/0x34
[ 506.359782] Code: f9400c22 91028025 f9400c00 39405043 (79402c42)
[ 506.360318] ---[ end trace 5477bbfb085279e5 ]---
``` | 1.0 | Oops, Unable to handle kernel paging request in scan_prefetch_queue_compare - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
Type | Version/Name
--- | ---
Distribution Name | Ubuntu (armbian)
Distribution Version | Focal
Linux Kernel | 5.8.17
Architecture | amd64
ZFS Version | 2.0.0-rc5
SPL Version | 2.0.0-rc5
### Describe the problem you're observing
zpool + zfs hang, seemingly uninterruptible, high load, Oops in dmesg
### Describe how to reproduce the problem
`zpool import tank0`
### Include any warning/errors/backtraces from the system logs
```
[ 506.332804] Unable to handle kernel paging request at virtual address fffe0000c4f5f316
[ 506.333567] Mem abort info:
[ 506.333815] ESR = 0x96000004
[ 506.334107] EC = 0x25: DABT (current EL), IL = 32 bits
[ 506.334581] SET = 0, FnV = 0
[ 506.334921] EA = 0, S1PTW = 0
[ 506.335198] Data abort info:
[ 506.335476] ISV = 0, ISS = 0x00000004
[ 506.335813] CM = 0, WnR = 0
[ 506.336076] [fffe0000c4f5f316] address between user and kernel address ranges
[ 506.336702] Internal error: Oops: 96000004 [#1] PREEMPT SMP
[ 506.337190] Modules linked in: xt_conntrack xt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo xt_addrtype iptable_filter iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 bpfilter br_netfilter bridge rfkill governor_performance zfs(POE) zunicode(POE) zzstd(OE) zlua(OE) zcommon(POE) znvpair(POE) zavl(POE) icp(POE) spl(OE) r8152 snd_soc_hdmi_codec leds_pwm pwm_fan gpio_charger panfrost gpu_sched snd_soc_rockchip_i2s snd_soc_core snd_pcm_dmaengine snd_pcm snd_timer hantro_vpu(C) snd videobuf2_vmalloc rockchip_vdec(C) fusb30x(C) rockchip_rga v4l2_h264 videobuf2_dma_contig soundcore v4l2_mem2mem videobuf2_dma_sg videobuf2_memops videobuf2_v4l2 videobuf2_common videodev mc zstd sg gpio_beeper cpufreq_dt zram sch_fq_codel lm75 ip_tables x_tables autofs4 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx raid1 raid0 multipath linear md_mod realtek rockchipdrm analogix_dp dw_hdmi dw_mipi_dsi drm_kms_helper cec rc_core dwmac_rk stmmac_platform drm
[ 506.337260] stmmac mdio_xpcs drm_panel_orientation_quirks adc_keys
[ 506.345397] CPU: 5 PID: 3192 Comm: z_rd_int Tainted: P C OE 5.8.17-rockchip64 #20.08.21
[ 506.346186] Hardware name: Helios64 (DT)
[ 506.346533] pstate: 80000005 (Nzcv daif -PAN -UAO BTYPE=--)
[ 506.347191] pc : scan_prefetch_queue_compare+0x20/0x40 [zfs]
[ 506.347694] lr : avl_find+0x68/0xc8 [zavl]
[ 506.348054] sp : ffff80001546baa0
[ 506.348347] x29: ffff80001546baa0 x28: ffff0000c207afa8
[ 506.348814] x27: ffff0000c316c9a8 x26: ffff0000c316c9c8
[ 506.349280] x25: ffff0000c30e5090 x24: 0000000000000000
[ 506.349746] x23: ffff80001546bb60 x22: ffff8000092b10d8
[ 506.350212] x21: ffff0000be677000 x20: ffff0000be57f200
[ 506.350677] x19: ffff0000be57f200 x18: 001c050ea761f9c0
[ 506.351143] x17: 00000445581d3e8e x16: 00001510366b591d
[ 506.351609] x15: 0000000000000000 x14: 0000000000000000
[ 506.352074] x13: 0000000000000008 x12: ffff0000f6fa9328
[ 506.352541] x11: 0000000000000020 x10: 0000000000000000
[ 506.353006] x9 : 000000000001e067 x8 : 00000000003c0ce0
[ 506.353472] x7 : 00000000003e2880 x6 : 0000000080000000
[ 506.353938] x5 : ffff0000be57f2a0 x4 : ffff0000be6770a0
[ 506.354404] x3 : 0000000000000011 x2 : ffff0000c4f5f300
[ 506.354870] x1 : ffff0000be57f200 x0 : ffff0000c4f5f300
[ 506.355337] Call trace:
[ 506.355661] scan_prefetch_queue_compare+0x20/0x40 [zfs]
[ 506.356132] avl_find+0x68/0xc8 [zavl]
[ 506.356553] dsl_scan_prefetch+0x1a8/0x228 [zfs]
[ 506.357042] dsl_scan_prefetch_cb+0x108/0x268 [zfs]
[ 506.357553] arc_read_done+0x20c/0x3f8 [zfs]
[ 506.358011] zio_done+0x254/0xd40 [zfs]
[ 506.358433] zio_execute+0xac/0x110 [zfs]
[ 506.358798] taskq_thread+0x298/0x440 [spl]
[ 506.359175] kthread+0x118/0x150
[ 506.359462] ret_from_fork+0x10/0x34
[ 506.359782] Code: f9400c22 91028025 f9400c00 39405043 (79402c42)
[ 506.360318] ---[ end trace 5477bbfb085279e5 ]---
``` | defect | oops unable to handle kernel paging request in scan prefetch queue compare thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu armbian distribution version focal linux kernel architecture zfs version spl version describe the problem you re observing zpool zfs hang seemingly uninterruptible high load oops in dmesg describe how to reproduce the problem zpool import include any warning errors backtraces from the system logs unable to handle kernel paging request at virtual address mem abort info esr ec dabt current el il bits set fnv ea data abort info isv iss cm wnr address between user and kernel address ranges internal error oops preempt smp modules linked in xt conntrack xt masquerade nf conntrack netlink nfnetlink xfrm user xfrm algo xt addrtype iptable filter iptable nat nf nat nf conntrack nf defrag nf defrag bpfilter br netfilter bridge rfkill governor performance zfs poe zunicode poe zzstd oe zlua oe zcommon poe znvpair poe zavl poe icp poe spl oe snd soc hdmi codec leds pwm pwm fan gpio charger panfrost gpu sched snd soc rockchip snd soc core snd pcm dmaengine snd pcm snd timer hantro vpu c snd vmalloc rockchip vdec c c rockchip rga dma contig soundcore dma sg memops common videodev mc zstd sg gpio beeper cpufreq dt zram sch fq codel ip tables x tables async recov async memcpy async pq async xor async tx multipath linear md mod realtek rockchipdrm analogix dp dw hdmi dw mipi dsi drm kms helper cec rc core dwmac rk stmmac platform drm stmmac mdio xpcs drm panel orientation quirks adc keys cpu pid comm z rd int tainted p c oe hardware name dt pstate nzcv daif pan uao btype pc scan prefetch queue compare lr avl find sp call trace scan prefetch queue compare avl find dsl scan prefetch dsl scan prefetch cb arc read done zio done zio execute taskq thread kthread ret from fork code | 1 |
17,087 | 2,974,593,191 | IssuesEvent | 2015-07-15 02:10:24 | Reimashi/jotai | https://api.github.com/repos/Reimashi/jotai | closed | Inclusion of Network Download and Upload Speeds | auto-migrated Priority-Medium Type-Defect wontfix | ```
What is the expected output?
Download and Upload speeds for each network adapter.
What version of the product are you using? On what operating system?
0.5.1 Beta on Windows 7 x64
Please provide any additional information below.
This would be a great feature and definitely one that is most likely in demand.
Thanks
```
Original issue reported on code.google.com by `sandeep....@gmail.com` on 7 Aug 2012 at 11:55 | 1.0 | Inclusion of Network Download and Upload Speeds - ```
What is the expected output?
Download and Upload speeds for each network adapter.
What version of the product are you using? On what operating system?
0.5.1 Beta on Windows 7 x64
Please provide any additional information below.
This would be a great feature and definitely one that is most likely in demand.
Thanks
```
Original issue reported on code.google.com by `sandeep....@gmail.com` on 7 Aug 2012 at 11:55 | defect | inclusion of network download and upload speeds what is the expected output download and upload speeds for each network adapter what version of the product are you using on what operating system beta on windows please provide any additional information below this would be a great feature and definitely one that is most likely in demand thanks original issue reported on code google com by sandeep gmail com on aug at | 1 |
22,484 | 15,217,301,666 | IssuesEvent | 2021-02-17 16:27:30 | airyhq/airy | https://api.github.com/repos/airyhq/airy | opened | Wait for resources to be created on airy create | cli infrastructure | Instead of waiting for services to start with the bash `wait-for-service` function, we need to implement this logic inside the CLI, for the full workflow of the `airy create` command. | 1.0 | Wait for resources to be created on airy create - Instead of waiting for services to start with the bash `wait-for-service` function, we need to implement this logic inside the CLI, for the full workflow of the `airy create` command. | non_defect | wait for resources to be created on airy create instead of waiting for services to start with the bash wait for service function we need to implement this logic inside the cli for the full workflow of the airy create command | 0 |
104,521 | 13,095,925,180 | IssuesEvent | 2020-08-03 14:52:53 | dusk-network/plonk | https://api.github.com/repos/dusk-network/plonk | closed | Make composer.public_inputs fn public again with `trace` feature | API-design constraint_system type:question | For some reason, this was removed with #255.
It is useful when you test gadgets to have access to this feature in order to avoid making by yourself the `public_inputs` vector as we agreed in the past.
Is there any reason why this was removed @kevaundray ?
If not, we should probably implement `public_inputs()` for `Prover` which would get the public inputs from the `StandardComposer` and give them to the end user. | 1.0 | Make composer.public_inputs fn public again with `trace` feature - For some reason, this was removed with #255.
It is useful when you test gadgets to have access to this feature in order to avoid making by yourself the `public_inputs` vector as we agreed in the past.
Is there any reason why this was removed @kevaundray ?
If not, we should probably implement `public_inputs()` for `Prover` which would get the public inputs from the `StandardComposer` and give them to the end user. | non_defect | make composer public inputs fn public again with trace feature for some reason this was removed with it is useful when you test gadgets to have access to this feature in order to avoid making by yourself the public inputs vector as we agreed in the past is there any reason why this was removed kevaundray if not we should probably implement public inputs for prover which would get the public inputs from the standardcomposer and give them to the end user | 0 |
559,958 | 16,581,872,711 | IssuesEvent | 2021-05-31 12:58:10 | gardener/landscaper | https://api.github.com/repos/gardener/landscaper | closed | Use new EnvTest Binaries Setup Tool | area/dev-productivity kind/enhancement priority/4 | **How to categorize this issue?**
<!--
Please select area, kind, and priority for this issue. This helps the community categorizing it.
Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion.
If multiple identifiers make sense you can also state the commands multiple times, e.g.
/area control-plane
/area auto-scaling
...
"/area" identifiers: backup|certification|cost|delivery|deployers|manifest-deployer|helm-deployer|container-deployer|dev-productivity|documentation|high-availability|logging|monitoring|oci|open-source|operations|ops-productivity|performance|quality|robustness|scalability|security|storage|testing|usability|user-management
"/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test
"/priority" identifiers (numerical value): 1 (blocker)|2 (critical)|3 (normal)|4 (low priority)|5 (nice to have)
-->
/area dev-productivity
/kind enhancement
/priority 4
**Description**:
Currently we install the test tools (kubebuilder apiserver + etcd) directly from the kubebuilder bucket using curl.
The controller-runtime has introduced a new tool manage these binaries also to have multiple versions available.
So lets switch our install requirements script to that new tool.
https://github.com/kubernetes-sigs/controller-runtime/pull/1488
| 1.0 | Use new EnvTest Binaries Setup Tool - **How to categorize this issue?**
<!--
Please select area, kind, and priority for this issue. This helps the community categorizing it.
Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion.
If multiple identifiers make sense you can also state the commands multiple times, e.g.
/area control-plane
/area auto-scaling
...
"/area" identifiers: backup|certification|cost|delivery|deployers|manifest-deployer|helm-deployer|container-deployer|dev-productivity|documentation|high-availability|logging|monitoring|oci|open-source|operations|ops-productivity|performance|quality|robustness|scalability|security|storage|testing|usability|user-management
"/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test
"/priority" identifiers (numerical value): 1 (blocker)|2 (critical)|3 (normal)|4 (low priority)|5 (nice to have)
-->
/area dev-productivity
/kind enhancement
/priority 4
**Description**:
Currently we install the test tools (kubebuilder apiserver + etcd) directly from the kubebuilder bucket using curl.
The controller-runtime has introduced a new tool manage these binaries also to have multiple versions available.
So lets switch our install requirements script to that new tool.
https://github.com/kubernetes-sigs/controller-runtime/pull/1488
| non_defect | use new envtest binaries setup tool how to categorize this issue please select area kind and priority for this issue this helps the community categorizing it replace below todos or exchange the existing identifiers with those that fit best in your opinion if multiple identifiers make sense you can also state the commands multiple times e g area control plane area auto scaling area identifiers backup certification cost delivery deployers manifest deployer helm deployer container deployer dev productivity documentation high availability logging monitoring oci open source operations ops productivity performance quality robustness scalability security storage testing usability user management kind identifiers api change bug cleanup discussion enhancement epic impediment poc post mortem question regression task technical debt test priority identifiers numerical value blocker critical normal low priority nice to have area dev productivity kind enhancement priority description currently we install the test tools kubebuilder apiserver etcd directly from the kubebuilder bucket using curl the controller runtime has introduced a new tool manage these binaries also to have multiple versions available so lets switch our install requirements script to that new tool | 0 |
710,477 | 24,419,968,151 | IssuesEvent | 2022-10-05 19:22:44 | bireme/fi-admin | https://api.github.com/repos/bireme/fi-admin | reopened | Pesquisar por novos índices | enhancement priority 2 tesauro | **Novos índices** em "Campos":
1- "TR - Descritores relacionados" (mostra todas as ocorrências em que o string pesquisado ocorre no campo "descritores relacionados". Essa funcionalidade existe no sistema antigo http://decsweb.bireme.br)

2- "Termos em notas de escopo e de indexação" (no primeiro resultado, em lista, deve mostrar "Nota de indexação | [o descritor contendo a nota] | [a nota]" ou "Nota de escopo | [o descritor contendo a nota] | [a nota]") O segundo resultado é aquele onde o usuário seleciona a linha desejada e clica para abrir o registro.
Os mais importantes são esses dois primeiros novos índices.
3- "Termos em notas históricas" (no primeiro resultado, em lista, deve mostrar "Nota histórica | [o descritor contendo essa nota] | [a nota]") (obs.: importante para podermos conhecer a história do termo da base. Hoje só conseguimos encontrar esses termos usando o webservice trigramas e pesquisando com o índice **DeCS em português com escopo**, pois esse usa o campo histórico do DeCS de consulta, que inclui esses termos.
Por exemplo, hoje percebi que o termo Aparelhos Sanitários (no plural) aparecia na última versão do DeCS no [decsweb](http://decsweb.bireme.br/) e no DeCS 2020 de consulta usando o webservice [trigramas](http://trigramas.bireme.br/cgi-bin/mx/cgi=@1?collection=DeCSp&lang=p&minsim=0.30&maxrel=10&text=Aparelho+Sanit%E1rio&resubmit=DeCSxp&x=41&y=5) mas não foi para a nota histórica de nenhum termo no descritor Aparelho Sanitário. Se houvesse esse índice seria mais fácil, além do que não sabemos se o serviço de trigramas vai continuar a existir muito mais tempo. Acabo de recriar o termo no plural. Hoje, 7/4/2021, editei esse issue e ao tentar pesquisar qq termo desse descritor no sistema de consulta atual, o sistema não conseguiu encontrar o registro. Parece que ele estava ou está com algum problema. Favor verificar.
4- "Termos ativos, históricos e em notas" (obs.: históricos aqui inclui status histórico e em notas históricas) (agendar reunião para confirmar de forma consensual se este deveria ser o novo default).
**Outros índices novos em termos ativos**:
5- "Descritores MeSH (ID do descritor iniciado com D0)", (obs.: pesquisa com string em qq idioma)
6- "Descritores exclusivos (ID do descritor iniciado com DD)", (obs.: pesquisa com string em qq idioma)
7- "Termos MeSH" (ID do termo iniciado com T) (obs.: pesquisa com string só em inglês, é claro)
8- "Termos exclusivos" (ID do termo iniciado com eng, spa, por, fre) | 1.0 | Pesquisar por novos índices - **Novos índices** em "Campos":
1- "TR - Descritores relacionados" (mostra todas as ocorrências em que o string pesquisado ocorre no campo "descritores relacionados". Essa funcionalidade existe no sistema antigo http://decsweb.bireme.br)

2- "Termos em notas de escopo e de indexação" (no primeiro resultado, em lista, deve mostrar "Nota de indexação | [o descritor contendo a nota] | [a nota]" ou "Nota de escopo | [o descritor contendo a nota] | [a nota]") O segundo resultado é aquele onde o usuário seleciona a linha desejada e clica para abrir o registro.
Os mais importantes são esses dois primeiros novos índices.
3- "Termos em notas históricas" (no primeiro resultado, em lista, deve mostrar "Nota histórica | [o descritor contendo essa nota] | [a nota]") (obs.: importante para podermos conhecer a história do termo da base. Hoje só conseguimos encontrar esses termos usando o webservice trigramas e pesquisando com o índice **DeCS em português com escopo**, pois esse usa o campo histórico do DeCS de consulta, que inclui esses termos.
Por exemplo, hoje percebi que o termo Aparelhos Sanitários (no plural) aparecia na última versão do DeCS no [decsweb](http://decsweb.bireme.br/) e no DeCS 2020 de consulta usando o webservice [trigramas](http://trigramas.bireme.br/cgi-bin/mx/cgi=@1?collection=DeCSp&lang=p&minsim=0.30&maxrel=10&text=Aparelho+Sanit%E1rio&resubmit=DeCSxp&x=41&y=5) mas não foi para a nota histórica de nenhum termo no descritor Aparelho Sanitário. Se houvesse esse índice seria mais fácil, além do que não sabemos se o serviço de trigramas vai continuar a existir muito mais tempo. Acabo de recriar o termo no plural. Hoje, 7/4/2021, editei esse issue e ao tentar pesquisar qq termo desse descritor no sistema de consulta atual, o sistema não conseguiu encontrar o registro. Parece que ele estava ou está com algum problema. Favor verificar.
4- "Termos ativos, históricos e em notas" (obs.: históricos aqui inclui status histórico e em notas históricas) (agendar reunião para confirmar de forma consensual se este deveria ser o novo default).
**Outros índices novos em termos ativos**:
5- "Descritores MeSH (ID do descritor iniciado com D0)", (obs.: pesquisa com string em qq idioma)
6- "Descritores exclusivos (ID do descritor iniciado com DD)", (obs.: pesquisa com string em qq idioma)
7- "Termos MeSH" (ID do termo iniciado com T) (obs.: pesquisa com string só em inglês, é claro)
8- "Termos exclusivos" (ID do termo iniciado com eng, spa, por, fre) | non_defect | pesquisar por novos índices novos índices em campos tr descritores relacionados mostra todas as ocorrências em que o string pesquisado ocorre no campo descritores relacionados essa funcionalidade existe no sistema antigo termos em notas de escopo e de indexação no primeiro resultado em lista deve mostrar nota de indexação ou nota de escopo o segundo resultado é aquele onde o usuário seleciona a linha desejada e clica para abrir o registro os mais importantes são esses dois primeiros novos índices termos em notas históricas no primeiro resultado em lista deve mostrar nota histórica obs importante para podermos conhecer a história do termo da base hoje só conseguimos encontrar esses termos usando o webservice trigramas e pesquisando com o índice decs em português com escopo pois esse usa o campo histórico do decs de consulta que inclui esses termos por exemplo hoje percebi que o termo aparelhos sanitários no plural aparecia na última versão do decs no e no decs de consulta usando o webservice mas não foi para a nota histórica de nenhum termo no descritor aparelho sanitário se houvesse esse índice seria mais fácil além do que não sabemos se o serviço de trigramas vai continuar a existir muito mais tempo acabo de recriar o termo no plural hoje editei esse issue e ao tentar pesquisar qq termo desse descritor no sistema de consulta atual o sistema não conseguiu encontrar o registro parece que ele estava ou está com algum problema favor verificar termos ativos históricos e em notas obs históricos aqui inclui status histórico e em notas históricas agendar reunião para confirmar de forma consensual se este deveria ser o novo default outros índices novos em termos ativos descritores mesh id do descritor iniciado com obs pesquisa com string em qq idioma descritores exclusivos id do descritor iniciado com dd obs pesquisa com string em qq idioma termos mesh id do termo iniciado com t obs pesquisa com string só em inglês é claro termos exclusivos id do termo iniciado com eng spa por fre | 0 |
58,259 | 16,453,935,780 | IssuesEvent | 2021-05-21 09:50:11 | Questie/Questie | https://api.github.com/repos/Questie/Questie | closed | "[DND] Tar Pedestal" at lots of places | Type - Defect | <!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
Lots of blacksmith icons on the map, named "dnd tar pedestal", only reinstall fix this for me, but they reappear after a while.
I don't have any profession trainers selected with a tick in questie, or in the default tracker
## Screenshots
[WoWScrnShot_052121_103040](https://user-images.githubusercontent.com/84509345/119108259-4ee0b080-ba20-11eb-821d-c212c2aea89d.jpg)
## Questie version
v6.3.11
| 1.0 | "[DND] Tar Pedestal" at lots of places - <!-- READ THIS FIRST
Hello, thanks for taking the time to report a bug!
Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie
Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is:
* @AeroScripts / Aero#1357 (Discord)
* @BreakBB / TheCrux#1702 (Discord)
* @drejjmit / Drejjmit#8241 (Discord)
* @Dyaxler / Dyaxler#0086 (Discord)
* @gogo1951 / Gogo#0298 (Discord)
If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs¤cy_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted
You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7
-->
## Bug description
Lots of blacksmith icons on the map, named "dnd tar pedestal", only reinstall fix this for me, but they reappear after a while.
I don't have any profession trainers selected with a tick in questie, or in the default tracker
## Screenshots
[WoWScrnShot_052121_103040](https://user-images.githubusercontent.com/84509345/119108259-4ee0b080-ba20-11eb-821d-c212c2aea89d.jpg)
## Questie version
v6.3.11
| defect | tar pedestal at lots of places read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description lots of blacksmith icons on the map named dnd tar pedestal only reinstall fix this for me but they reappear after a while i don t have any profession trainers selected with a tick in questie or in the default tracker screenshots questie version | 1 |
28,085 | 5,185,297,643 | IssuesEvent | 2017-01-20 09:57:44 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Oracle 12c support for OFFSET .. FETCH causes ORA-00918 regression on queries with ambiguous column names | C: DB: Oracle C: Functionality P: Urgent R: Fixed T: Defect | jOOQ 3.4 introduced a fix (https://github.com/jOOQ/jOOQ/issues/2335) that works around a limitation in SQL where subqueries are not allowed to produce "ambiguous" column names.
jOOQ users may write top-level queries as such:
``` java
DSL.using(configuration)
.select()
.from(BOOK)
.join(AUTHOR).on(BOOK.AUTHOR_ID.eq(AUTHOR.ID))
.limit(1)
.fetch();
```
There are ambiguous column names `BOOK.ID` and `AUTHOR.ID`, although from a user perspective, this doesn't matter because the two columns are referenced from a top level select. jOOQ, however, emulates `LIMIT` for Oracle using derived tables and `ROWNUM` filtering. #2335 Renames all columns in the nested select, then applies `ROWNUM` filtering, then renames columns back again to their original names.
With support for Oracle 12c `OFFSET .. FETCH` (#2607), there is now a regression as Oracle 12c doesn't support "ambiguous" column names in top level selects that contain this clause:

We'll have to re-implement the same workaround again
| 1.0 | Oracle 12c support for OFFSET .. FETCH causes ORA-00918 regression on queries with ambiguous column names - jOOQ 3.4 introduced a fix (https://github.com/jOOQ/jOOQ/issues/2335) that works around a limitation in SQL where subqueries are not allowed to produce "ambiguous" column names.
jOOQ users may write top-level queries as such:
``` java
DSL.using(configuration)
.select()
.from(BOOK)
.join(AUTHOR).on(BOOK.AUTHOR_ID.eq(AUTHOR.ID))
.limit(1)
.fetch();
```
There are ambiguous column names `BOOK.ID` and `AUTHOR.ID`, although from a user perspective, this doesn't matter because the two columns are referenced from a top level select. jOOQ, however, emulates `LIMIT` for Oracle using derived tables and `ROWNUM` filtering. #2335 Renames all columns in the nested select, then applies `ROWNUM` filtering, then renames columns back again to their original names.
With support for Oracle 12c `OFFSET .. FETCH` (#2607), there is now a regression as Oracle 12c doesn't support "ambiguous" column names in top level selects that contain this clause:

We'll have to re-implement the same workaround again
| defect | oracle support for offset fetch causes ora regression on queries with ambiguous column names jooq introduced a fix that works around a limitation in sql where subqueries are not allowed to produce ambiguous column names jooq users may write top level queries as such java dsl using configuration select from book join author on book author id eq author id limit fetch there are ambiguous column names book id and author id although from a user perspective this doesn t matter because the two columns are referenced from a top level select jooq however emulates limit for oracle using derived tables and rownum filtering renames all columns in the nested select then applies rownum filtering then renames columns back again to their original names with support for oracle offset fetch there is now a regression as oracle doesn t support ambiguous column names in top level selects that contain this clause we ll have to re implement the same workaround again | 1 |
151,461 | 23,830,452,223 | IssuesEvent | 2022-09-05 20:00:10 | SynBioHub/synbiohub3 | https://api.github.com/repos/SynBioHub/synbiohub3 | opened | Redo User Design | Design | - Move off H2 database for user storage to a more standard relational database (mySql most likely)
- Add data access layers
- Add more tables to describe the connections between classes (refer to UML class diagram schematic)
- Integrate with proper authentication layers | 1.0 | Redo User Design - - Move off H2 database for user storage to a more standard relational database (mySql most likely)
- Add data access layers
- Add more tables to describe the connections between classes (refer to UML class diagram schematic)
- Integrate with proper authentication layers | non_defect | redo user design move off database for user storage to a more standard relational database mysql most likely add data access layers add more tables to describe the connections between classes refer to uml class diagram schematic integrate with proper authentication layers | 0 |
186,428 | 21,933,228,494 | IssuesEvent | 2022-05-23 11:39:33 | onokatio/blog.katio.net | https://api.github.com/repos/onokatio/blog.katio.net | closed | CVE-2021-23368 (Medium) detected in postcss-7.0.32.tgz | security vulnerability | ## CVE-2021-23368 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- css-loader-3.6.0.tgz (Root Library)
- :x: **postcss-7.0.32.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/onokatio/blog.katio.net/commit/936580315e62ac99730c0ed7a501c46359f1c0ed">936580315e62ac99730c0ed7a501c46359f1c0ed</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (css-loader): 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23368 (Medium) detected in postcss-7.0.32.tgz - ## CVE-2021-23368 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- css-loader-3.6.0.tgz (Root Library)
- :x: **postcss-7.0.32.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/onokatio/blog.katio.net/commit/936580315e62ac99730c0ed7a501c46359f1c0ed">936580315e62ac99730c0ed7a501c46359f1c0ed</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (css-loader): 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in postcss tgz cve medium severity vulnerability vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy css loader tgz root library x postcss tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package postcss from and before are vulnerable to regular expression denial of service redos during source map parsing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss direct dependency fix resolution css loader step up your open source security game with whitesource | 0 |
53,787 | 13,262,299,331 | IssuesEvent | 2020-08-20 21:29:52 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [mue] remove `compat` option from muex (Trac #2094) | Migrated from Trac combo reconstruction defect | it causes more confusion than it solves
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2094">https://code.icecube.wisc.edu/projects/icecube/ticket/2094</a>, reported by kjmeagherand owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"_ts": "1550067295757382",
"description": "it causes more confusion than it solves",
"reporter": "kjmeagher",
"cc": "desiati",
"resolution": "invalid",
"time": "2017-09-20T12:55:48",
"component": "combo reconstruction",
"summary": "[mue] remove `compat` option from muex",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [mue] remove `compat` option from muex (Trac #2094) - it causes more confusion than it solves
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2094">https://code.icecube.wisc.edu/projects/icecube/ticket/2094</a>, reported by kjmeagherand owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"_ts": "1550067295757382",
"description": "it causes more confusion than it solves",
"reporter": "kjmeagher",
"cc": "desiati",
"resolution": "invalid",
"time": "2017-09-20T12:55:48",
"component": "combo reconstruction",
"summary": "[mue] remove `compat` option from muex",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| defect | remove compat option from muex trac it causes more confusion than it solves migrated from json status closed changetime ts description it causes more confusion than it solves reporter kjmeagher cc desiati resolution invalid time component combo reconstruction summary remove compat option from muex priority normal keywords milestone owner kjmeagher type defect | 1 |
5,050 | 2,610,165,940 | IssuesEvent | 2015-02-26 18:52:36 | chrsmith/republic-at-war | https://api.github.com/repos/chrsmith/republic-at-war | closed | SFX | auto-migrated Priority-Medium Type-Defect | ```
Obi Wan's lock s-foils uses Lukes voice
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 4 May 2011 at 9:41 | 1.0 | SFX - ```
Obi Wan's lock s-foils uses Lukes voice
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 4 May 2011 at 9:41 | defect | sfx obi wan s lock s foils uses lukes voice original issue reported on code google com by gmail com on may at | 1 |
282,914 | 30,889,475,951 | IssuesEvent | 2023-08-04 02:46:56 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | reopened | CVE-2019-19768 (High) detected in linux-stable-rtv4.1.33 | Mend: dependency security vulnerability | ## CVE-2019-19768 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel 5.4.0-rc2, there is a use-after-free (read) in the __blk_add_trace function in kernel/trace/blktrace.c (which is used to fill out a blk_io_trace structure and place it in a per-cpu sub-buffer).
<p>Publish Date: 2019-12-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19768>CVE-2019-19768</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-19768">https://nvd.nist.gov/vuln/detail/CVE-2019-19768</a></p>
<p>Release Date: 2020-06-10</p>
<p>Fix Resolution: kernel-doc - 3.10.0-514.76.1,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-327.88.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1;kernel-rt-core - 4.18.0-193.rt13.51;kernel-rt-debug-debuginfo - 4.18.0-193.rt13.51;kernel-abi-whitelists - 3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1;kernel-zfcpdump-modules - 4.18.0-193,4.18.0-147.13.2;kernel-rt-trace-devel - 3.10.0-1127.8.2.rt56.1103;kernel-debug-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;kernel-bootwrapper - 3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1;kernel-rt-debuginfo - 4.18.0-193.rt13.51;kernel-rt-debug-modules - 4.18.0-193.rt13.51;kernel-zfcpdump-devel - 4.18.0-193,4.18.0-147.13.2;perf - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,4.18.0-193,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1;kernel-zfcpdump-modules-extra - 4.18.0-193,4.18.0-147.13.2;kernel-debuginfo - 3.10.0-514.76.1,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1;kernel-debug-devel - 3.10.0-514.76.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-327.88.1,4.18.0-193,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-80.18.1;bpftool - 3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-debug-core - 4.18.0-193.rt13.51;kernel-tools-libs - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;perf-debuginfo - 3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-327.88.1;kernel-cross-headers - 4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-debug-debuginfo - 3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;kernel-debug - 3.10.0-514.76.1,3.10.0-327.88.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-957.54.1,4.18.0-193,4.18.0-193,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2;kernel-devel - 4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-327.88.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2;bpftool-debuginfo - 4.18.0-193,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-80.18.1;kpatch-patch-3_10_0-1062_12_1 - 1-2,1-2;kernel-zfcpdump-core - 4.18.0-147.13.2,4.18.0-193;kernel-debug-core - 4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193;kernel-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;python-perf - 3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-core - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-debug - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-rt-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-debuginfo-common-ppc64 - 3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1062.26.1;python3-perf - 4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2;kernel-tools - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel-debug-modules - 4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-trace-kvm - 3.10.0-1127.8.2.rt56.1103;kernel-rt-debuginfo-common-x86_64 - 4.18.0-193.rt13.51;kernel-tools-libs-devel - 3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-modules - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193;kernel-tools-debuginfo - 3.10.0-1062.26.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-80.18.1,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-693.67.1;kernel-rt-modules - 4.18.0-193.rt13.51;kernel-rt-doc - 3.10.0-1127.8.2.rt56.1103;kernel-rt-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;python-perf-debuginfo - 3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1;kernel-headers - 3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-trace - 3.10.0-1127.8.2.rt56.1103;kernel-debuginfo-common-x86_64 - 3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-514.76.1,4.18.0-193,3.10.0-957.54.1;kernel-rt - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-zfcpdump - 4.18.0-147.13.2,4.18.0-193;kernel-rt-debug-modules-extra - 4.18.0-193.rt13.51;python3-perf-debuginfo - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193;kernel-rt-modules-extra - 4.18.0-193.rt13.51</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19768 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2019-19768 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel 5.4.0-rc2, there is a use-after-free (read) in the __blk_add_trace function in kernel/trace/blktrace.c (which is used to fill out a blk_io_trace structure and place it in a per-cpu sub-buffer).
<p>Publish Date: 2019-12-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19768>CVE-2019-19768</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-19768">https://nvd.nist.gov/vuln/detail/CVE-2019-19768</a></p>
<p>Release Date: 2020-06-10</p>
<p>Fix Resolution: kernel-doc - 3.10.0-514.76.1,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-327.88.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1;kernel-rt-core - 4.18.0-193.rt13.51;kernel-rt-debug-debuginfo - 4.18.0-193.rt13.51;kernel-abi-whitelists - 3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1;kernel-zfcpdump-modules - 4.18.0-193,4.18.0-147.13.2;kernel-rt-trace-devel - 3.10.0-1127.8.2.rt56.1103;kernel-debug-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;kernel-bootwrapper - 3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1;kernel-rt-debuginfo - 4.18.0-193.rt13.51;kernel-rt-debug-modules - 4.18.0-193.rt13.51;kernel-zfcpdump-devel - 4.18.0-193,4.18.0-147.13.2;perf - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,4.18.0-193,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1;kernel-zfcpdump-modules-extra - 4.18.0-193,4.18.0-147.13.2;kernel-debuginfo - 3.10.0-514.76.1,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1;kernel-debug-devel - 3.10.0-514.76.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-327.88.1,4.18.0-193,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-80.18.1;bpftool - 3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-debug-core - 4.18.0-193.rt13.51;kernel-tools-libs - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;perf-debuginfo - 3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-327.88.1;kernel-cross-headers - 4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-debug-debuginfo - 3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;kernel-debug - 3.10.0-514.76.1,3.10.0-327.88.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-957.54.1,4.18.0-193,4.18.0-193,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2;kernel-devel - 4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-327.88.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2;bpftool-debuginfo - 4.18.0-193,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-80.18.1;kpatch-patch-3_10_0-1062_12_1 - 1-2,1-2;kernel-zfcpdump-core - 4.18.0-147.13.2,4.18.0-193;kernel-debug-core - 4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193;kernel-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;python-perf - 3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-core - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-debug - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-rt-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-debuginfo-common-ppc64 - 3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1062.26.1;python3-perf - 4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2;kernel-tools - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel-debug-modules - 4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-trace-kvm - 3.10.0-1127.8.2.rt56.1103;kernel-rt-debuginfo-common-x86_64 - 4.18.0-193.rt13.51;kernel-tools-libs-devel - 3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-modules - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193;kernel-tools-debuginfo - 3.10.0-1062.26.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-80.18.1,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-693.67.1;kernel-rt-modules - 4.18.0-193.rt13.51;kernel-rt-doc - 3.10.0-1127.8.2.rt56.1103;kernel-rt-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;python-perf-debuginfo - 3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1;kernel-headers - 3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-trace - 3.10.0-1127.8.2.rt56.1103;kernel-debuginfo-common-x86_64 - 3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-514.76.1,4.18.0-193,3.10.0-957.54.1;kernel-rt - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-zfcpdump - 4.18.0-147.13.2,4.18.0-193;kernel-rt-debug-modules-extra - 4.18.0-193.rt13.51;python3-perf-debuginfo - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193;kernel-rt-modules-extra - 4.18.0-193.rt13.51</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files kernel trace blktrace c kernel trace blktrace c vulnerability details in the linux kernel there is a use after free read in the blk add trace function in kernel trace blktrace c which is used to fill out a blk io trace structure and place it in a per cpu sub buffer publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution kernel doc kernel rt core kernel rt debug debuginfo kernel abi whitelists kernel zfcpdump modules kernel rt trace devel kernel debug modules extra kernel rt debug kvm kernel bootwrapper kernel rt debuginfo kernel rt debug modules kernel zfcpdump devel perf kernel zfcpdump modules extra kernel debuginfo kernel debug devel bpftool kernel rt debug core kernel tools libs perf debuginfo kernel cross headers kernel debug debuginfo kernel debug kernel devel kernel bpftool debuginfo kpatch patch kernel zfcpdump core kernel debug core kernel modules extra kernel rt debug devel python perf kernel core kernel rt debug kernel rt devel kernel debuginfo common perf kernel tools kernel debug modules kernel rt trace kvm kernel rt debuginfo common kernel tools libs devel kernel modules kernel tools debuginfo kernel rt modules kernel rt doc kernel rt kvm python perf debuginfo kernel headers kernel rt trace kernel debuginfo common kernel rt kernel zfcpdump kernel rt debug modules extra perf debuginfo kernel rt modules extra step up your open source security game with mend | 0 |
100,974 | 21,562,551,057 | IssuesEvent | 2022-05-01 11:36:46 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.1.x] Cassiopea Registration page Privacy/Terms alignment | New Feature No Code Attached Yet J4 Frontend Template | Hi guys,
about the [Cassiopea Registration page Privacy/Terms alignment](https://photos.app.goo.gl/6VZuH7Ja7RQ8yfKt5 "Registration Privacy/Terms"),
Should not be better to add by default:
.required.radio {
display: inline-flex;
gap: 1rem;
}
to align them horizontally and don't waste precious space ? | 1.0 | [4.1.x] Cassiopea Registration page Privacy/Terms alignment - Hi guys,
about the [Cassiopea Registration page Privacy/Terms alignment](https://photos.app.goo.gl/6VZuH7Ja7RQ8yfKt5 "Registration Privacy/Terms"),
Should not be better to add by default:
.required.radio {
display: inline-flex;
gap: 1rem;
}
to align them horizontally and don't waste precious space ? | non_defect | cassiopea registration page privacy terms alignment hi guys about the registration privacy terms should not be better to add by default required radio display inline flex gap to align them horizontally and don t waste precious space | 0 |
75,148 | 25,557,257,859 | IssuesEvent | 2022-11-30 07:58:32 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Unread message/mention comes back after closing the browser and re-opening | T-Defect S-Minor A-Notifications O-Occasional | ### Steps to reproduce
When a message is last sent by a contact with a mention/response after logging in or after closing the client/browser and reopening it, it will still show up as an unread messages even if the message was read before.
### Outcome
#### What did you expect?
Normal use.
#### What happened instead?
There is a problem with the display.
### Operating system
Ubuntu (Actually all clients have this problem)
### Browser information
Edge 107.0
### URL for webapp
_No response_
### Application version
Element Web/Desktop 1.11.15 Element iOS 1.9.12
### Homeserver
Synapse 1.72.0
### Will you send logs?
No | 1.0 | Unread message/mention comes back after closing the browser and re-opening - ### Steps to reproduce
When a message is last sent by a contact with a mention/response after logging in or after closing the client/browser and reopening it, it will still show up as an unread messages even if the message was read before.
### Outcome
#### What did you expect?
Normal use.
#### What happened instead?
There is a problem with the display.
### Operating system
Ubuntu (Actually all clients have this problem)
### Browser information
Edge 107.0
### URL for webapp
_No response_
### Application version
Element Web/Desktop 1.11.15 Element iOS 1.9.12
### Homeserver
Synapse 1.72.0
### Will you send logs?
No | defect | unread message mention comes back after closing the browser and re opening steps to reproduce when a message is last sent by a contact with a mention response after logging in or after closing the client browser and reopening it it will still show up as an unread messages even if the message was read before outcome what did you expect normal use what happened instead there is a problem with the display operating system ubuntu actually all clients have this problem browser information edge url for webapp no response application version element web desktop element ios homeserver synapse will you send logs no | 1 |
10,079 | 2,618,934,929 | IssuesEvent | 2015-03-03 00:01:42 | chrsmith/open-ig | https://api.github.com/repos/chrsmith/open-ig | closed | pre-hyperdrive bugs | auto-migrated Component-Logic Priority-Medium Type-Defect | ```
Game version: 0.95.141
Operating System: xp 32
Java runtime version: jdk1.7.0_04
Installed using the Launcher? yes
Game language de
1) Produced destroyers possess a hyperdrive and they have according speed.
2) The cruiser given at the start of level 2 has a hyperdrive and according
speed.
3) The destroyer left over from level 1 is shown to have a hyperdrive in fleet
equipment screen (box in the middle, actual:1/max:1) but is shown with speed 6
as it should.
It looks as if fleets/ships do not always move with the speed they are shown
with now since you made the traders "catchable", this confuses me. (Am I faster
or is the trader slower than shown ?) The cruiser really moves faster than the
leftover destroyer at least. Same for produced destroyers.
Attached the start save for level 2 and the last one I still have from before.
```
Original issue reported on code.google.com by `anonymis...@arcor.de` on 27 Aug 2013 at 3:15
Attachments:
* [save-2013-08-27-01-32-55-406.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-32-55-406.xml.gz)
* [save-2013-08-27-01-55-32-718.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-55-32-718.xml.gz)
| 1.0 | pre-hyperdrive bugs - ```
Game version: 0.95.141
Operating System: xp 32
Java runtime version: jdk1.7.0_04
Installed using the Launcher? yes
Game language de
1) Produced destroyers possess a hyperdrive and they have according speed.
2) The cruiser given at the start of level 2 has a hyperdrive and according
speed.
3) The destroyer left over from level 1 is shown to have a hyperdrive in fleet
equipment screen (box in the middle, actual:1/max:1) but is shown with speed 6
as it should.
It looks as if fleets/ships do not always move with the speed they are shown
with now since you made the traders "catchable", this confuses me. (Am I faster
or is the trader slower than shown ?) The cruiser really moves faster than the
leftover destroyer at least. Same for produced destroyers.
Attached the start save for level 2 and the last one I still have from before.
```
Original issue reported on code.google.com by `anonymis...@arcor.de` on 27 Aug 2013 at 3:15
Attachments:
* [save-2013-08-27-01-32-55-406.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-32-55-406.xml.gz)
* [save-2013-08-27-01-55-32-718.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-55-32-718.xml.gz)
| defect | pre hyperdrive bugs game version operating system xp java runtime version installed using the launcher yes game language de produced destroyers possess a hyperdrive and they have according speed the cruiser given at the start of level has a hyperdrive and according speed the destroyer left over from level is shown to have a hyperdrive in fleet equipment screen box in the middle actual max but is shown with speed as it should it looks as if fleets ships do not always move with the speed they are shown with now since you made the traders catchable this confuses me am i faster or is the trader slower than shown the cruiser really moves faster than the leftover destroyer at least same for produced destroyers attached the start save for level and the last one i still have from before original issue reported on code google com by anonymis arcor de on aug at attachments | 1 |
15,913 | 2,869,093,867 | IssuesEvent | 2015-06-05 23:17:04 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Span.getLocationMessage returns incorrect output for lines without trailing newlines | Area-Pkg Pkg-SourceMaps Priority-Unassigned Triaged Type-Defect | If a span points to the final line in a file, and that line doesn't end in a newline, the arrows in getLocationMessage will be rendered on the same line as the source text. For example:
Error on line 12, column 28 of pubspec.yaml: Error loading transformer: I hate these settings!
- app/src/html_rewriter: {} ^^
| 1.0 | Span.getLocationMessage returns incorrect output for lines without trailing newlines - If a span points to the final line in a file, and that line doesn't end in a newline, the arrows in getLocationMessage will be rendered on the same line as the source text. For example:
Error on line 12, column 28 of pubspec.yaml: Error loading transformer: I hate these settings!
- app/src/html_rewriter: {} ^^
| defect | span getlocationmessage returns incorrect output for lines without trailing newlines if a span points to the final line in a file and that line doesn t end in a newline the arrows in getlocationmessage will be rendered on the same line as the source text for example nbsp nbsp nbsp nbsp error on line column of pubspec yaml error loading transformer i hate these settings nbsp nbsp nbsp nbsp nbsp nbsp app src html rewriter | 1 |
118,302 | 9,981,595,018 | IssuesEvent | 2019-07-10 07:54:03 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | reopened | Update 'URI' to 'URL' on the first page of the Connect dialog | :gear: attach :gear: sas 🧪 testing | **Storage Explorer Version:** rel/1.9.0_20190626.4
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Commit:** 0afd7438
**Regression From:** Not a regression
**Steps to reproduce:**
1. Launch storage explorer -> Open Connect dialog.
2. Check the descriptions on the dialog.
**Expected Experience:**
Show 'Use a shared access signature (SAS) **URL**'
**Actual Experience:**
Show 'Use a shared access signature (SAS) **URI**'

**More Info:**
Show URL when generating share access signature.

| 1.0 | Update 'URI' to 'URL' on the first page of the Connect dialog - **Storage Explorer Version:** rel/1.9.0_20190626.4
**Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10
**Architecture:** ia32/x64
**Commit:** 0afd7438
**Regression From:** Not a regression
**Steps to reproduce:**
1. Launch storage explorer -> Open Connect dialog.
2. Check the descriptions on the dialog.
**Expected Experience:**
Show 'Use a shared access signature (SAS) **URL**'
**Actual Experience:**
Show 'Use a shared access signature (SAS) **URI**'

**More Info:**
Show URL when generating share access signature.

| non_defect | update uri to url on the first page of the connect dialog storage explorer version rel platform os linux ubuntu macos high sierra windows architecture commit regression from not a regression steps to reproduce launch storage explorer open connect dialog check the descriptions on the dialog expected experience show use a shared access signature sas url actual experience show use a shared access signature sas uri more info show url when generating share access signature | 0 |
6,732 | 2,610,274,129 | IssuesEvent | 2015-02-26 19:27:44 | chrsmith/scribefire-chrome | https://api.github.com/repos/chrsmith/scribefire-chrome | closed | Window always opens narrow | auto-migrated Priority-Medium Type-Defect | ```
What's the problem?
every time I click the scribefire icon it opens the scribefire "window" very
narrow and tall (unusable) -and I must adjust every single time I open the
window again
What browser are you using?
Firefox - several versions 3.6, 3.6.13 and 4 ... both ob XP, Vista and
Windows 7
What version of ScribeFire are you running?
How nice it would be if you had an "about" link inside scribefire to make it
easier running Scribefire Next 1.4.3.0 (it does not happen in regular
Scribefire 3.5.3.3
```
-----
Original issue reported on code.google.com by `cjcornell2` on 23 Feb 2011 at 9:47 | 1.0 | Window always opens narrow - ```
What's the problem?
every time I click the scribefire icon it opens the scribefire "window" very
narrow and tall (unusable) -and I must adjust every single time I open the
window again
What browser are you using?
Firefox - several versions 3.6, 3.6.13 and 4 ... both ob XP, Vista and
Windows 7
What version of ScribeFire are you running?
How nice it would be if you had an "about" link inside scribefire to make it
easier running Scribefire Next 1.4.3.0 (it does not happen in regular
Scribefire 3.5.3.3
```
-----
Original issue reported on code.google.com by `cjcornell2` on 23 Feb 2011 at 9:47 | defect | window always opens narrow what s the problem every time i click the scribefire icon it opens the scribefire window very narrow and tall unusable and i must adjust every single time i open the window again what browser are you using firefox several versions and both ob xp vista and windows what version of scribefire are you running how nice it would be if you had an about link inside scribefire to make it easier running scribefire next it does not happen in regular scribefire original issue reported on code google com by on feb at | 1 |
16,971 | 3,587,039,131 | IssuesEvent | 2016-01-30 02:17:08 | dylanaraps/fetch | https://api.github.com/repos/dylanaraps/fetch | closed | GPU support. | Needs Testing | ### An experimental version of GPU support is currently in master.
Here's the status for GPU support:
- [X] Linux
- [X] Mac OS X
- [X] FreeBSD
- [ ] OpenBSD
- We could've used `pciutils` to get lspci functionality on OpenBSD but it turns out that it requires root.
- [ ] NetBSD
- No clue on this one, I haven't been able to find a way to get gpu.
- [X] Windows
Adding support for OpenBSD and NetBSD is impossible at this point as I can't find a way to get the gpu name without needing root privileges.
### Troubleshooting
#### All OS:
I'll need to see a screenshot of fetch.
#### Linux:
I'll need to see the output of this command:
```sh
lspci | grep "VGA"
```
If the command above prints something vague instead of your gpu, your pci.ids files may be outdated. See this part of the readme on how to update them.
https://github.com/dylanaraps/fetch#getgpu-doesnt-show-my-exact-video-card-name
#### Mac OS X:
I'll need to see the output of this command:
```sh
system_profiler SPDisplaysDataType | awk -F': ' '/^\ *Chipset Model:/ {printf $2}'
```
#### FreeBSD:
I'll need to see the output of this command:
```sh
pciconf -lv 2>/dev/null | grep -B 4 "VGA"
```
#### Windows:
I'll need to see the output of this command:
```sh
wmic path Win32_VideoController get caption
```
| 1.0 | GPU support. - ### An experimental version of GPU support is currently in master.
Here's the status for GPU support:
- [X] Linux
- [X] Mac OS X
- [X] FreeBSD
- [ ] OpenBSD
- We could've used `pciutils` to get lspci functionality on OpenBSD but it turns out that it requires root.
- [ ] NetBSD
- No clue on this one, I haven't been able to find a way to get gpu.
- [X] Windows
Adding support for OpenBSD and NetBSD is impossible at this point as I can't find a way to get the gpu name without needing root privileges.
### Troubleshooting
#### All OS:
I'll need to see a screenshot of fetch.
#### Linux:
I'll need to see the output of this command:
```sh
lspci | grep "VGA"
```
If the command above prints something vague instead of your gpu, your pci.ids files may be outdated. See this part of the readme on how to update them.
https://github.com/dylanaraps/fetch#getgpu-doesnt-show-my-exact-video-card-name
#### Mac OS X:
I'll need to see the output of this command:
```sh
system_profiler SPDisplaysDataType | awk -F': ' '/^\ *Chipset Model:/ {printf $2}'
```
#### FreeBSD:
I'll need to see the output of this command:
```sh
pciconf -lv 2>/dev/null | grep -B 4 "VGA"
```
#### Windows:
I'll need to see the output of this command:
```sh
wmic path Win32_VideoController get caption
```
| non_defect | gpu support an experimental version of gpu support is currently in master here s the status for gpu support linux mac os x freebsd openbsd we could ve used pciutils to get lspci functionality on openbsd but it turns out that it requires root netbsd no clue on this one i haven t been able to find a way to get gpu windows adding support for openbsd and netbsd is impossible at this point as i can t find a way to get the gpu name without needing root privileges troubleshooting all os i ll need to see a screenshot of fetch linux i ll need to see the output of this command sh lspci grep vga if the command above prints something vague instead of your gpu your pci ids files may be outdated see this part of the readme on how to update them mac os x i ll need to see the output of this command sh system profiler spdisplaysdatatype awk f chipset model printf freebsd i ll need to see the output of this command sh pciconf lv dev null grep b vga windows i ll need to see the output of this command sh wmic path videocontroller get caption | 0 |
10,115 | 2,618,937,001 | IssuesEvent | 2015-03-03 00:02:28 | chrsmith/open-ig | https://api.github.com/repos/chrsmith/open-ig | closed | Units not attacks space stations in vinicity | auto-migrated Priority-Medium Type-Defect | ```
Game version: 0.95.153
Operating System: Linux x64
Java runtime version: 1.7.0_51
Installed using the Launcher? yes
Game language (en, hu, de): hu
What steps will reproduce the problem?
1. Attack an enemy planet which has space stations.
2. Move your crusiers or destroyers near a space station, just enough close for
the shot range of heavy weapons reach it.
What is the expected output? What do you see instead?
They should attack the station with the heavy weapons (particle, meson).
Instead they let them shot by the station. And when i tell them to attack, they
do not shot from they are, they move closer and now the other stations can shot
the player units.
Please provide any additional information below.
Please upload any save before and/or after the problem happened. Please
attach the open-ig.log file found in the
application's directory.
Can give, but any planet attack will do.
```
Original issue reported on code.google.com by `kli...@gmail.com` on 17 Jan 2014 at 11:18
* Merged into: #803 | 1.0 | Units not attacks space stations in vinicity - ```
Game version: 0.95.153
Operating System: Linux x64
Java runtime version: 1.7.0_51
Installed using the Launcher? yes
Game language (en, hu, de): hu
What steps will reproduce the problem?
1. Attack an enemy planet which has space stations.
2. Move your crusiers or destroyers near a space station, just enough close for
the shot range of heavy weapons reach it.
What is the expected output? What do you see instead?
They should attack the station with the heavy weapons (particle, meson).
Instead they let them shot by the station. And when i tell them to attack, they
do not shot from they are, they move closer and now the other stations can shot
the player units.
Please provide any additional information below.
Please upload any save before and/or after the problem happened. Please
attach the open-ig.log file found in the
application's directory.
Can give, but any planet attack will do.
```
Original issue reported on code.google.com by `kli...@gmail.com` on 17 Jan 2014 at 11:18
* Merged into: #803 | defect | units not attacks space stations in vinicity game version operating system linux java runtime version installed using the launcher yes game language en hu de hu what steps will reproduce the problem attack an enemy planet which has space stations move your crusiers or destroyers near a space station just enough close for the shot range of heavy weapons reach it what is the expected output what do you see instead they should attack the station with the heavy weapons particle meson instead they let them shot by the station and when i tell them to attack they do not shot from they are they move closer and now the other stations can shot the player units please provide any additional information below please upload any save before and or after the problem happened please attach the open ig log file found in the application s directory can give but any planet attack will do original issue reported on code google com by kli gmail com on jan at merged into | 1 |
6,929 | 3,933,560,306 | IssuesEvent | 2016-04-25 19:33:39 | ghutchis/avogadro | https://api.github.com/repos/ghutchis/avogadro | closed | Require #include <GL/glu.h> on Ubuntu 12.04 | auto-migrated Building / Install v_1.1.0 | Compiling avogadro 1.0.3 on Ubuntu 12.04 requires the insertion of
#include <GL/glu.h>
in the following files:
libavogadro/src/glpainter_p.cpp
libavogadro/src/glwidget.cpp
libavogadro/src/tools/selectrotatetool.cpp
libavogadro/src/camera.cpp
This was not required on Ubuntu 10.04, but I guess the GL headers have changed substantially in that period.
Reported by: *anonymous | 1.0 | Require #include <GL/glu.h> on Ubuntu 12.04 - Compiling avogadro 1.0.3 on Ubuntu 12.04 requires the insertion of
#include <GL/glu.h>
in the following files:
libavogadro/src/glpainter_p.cpp
libavogadro/src/glwidget.cpp
libavogadro/src/tools/selectrotatetool.cpp
libavogadro/src/camera.cpp
This was not required on Ubuntu 10.04, but I guess the GL headers have changed substantially in that period.
Reported by: *anonymous | non_defect | require include on ubuntu compiling avogadro on ubuntu requires the insertion of include in the following files libavogadro src glpainter p cpp libavogadro src glwidget cpp libavogadro src tools selectrotatetool cpp libavogadro src camera cpp this was not required on ubuntu but i guess the gl headers have changed substantially in that period reported by anonymous | 0 |
9,878 | 2,616,005,523 | IssuesEvent | 2015-03-02 00:49:56 | jasonhall/bwapi | https://api.github.com/repos/jasonhall/bwapi | closed | isIdle and build queue | auto-migrated Priority-Low Type-Defect | ```
isIdle does not consider a building that is training a unit.
```
Original issue reported on code.google.com by `AHeinerm` on 29 Jul 2009 at 8:18 | 1.0 | isIdle and build queue - ```
isIdle does not consider a building that is training a unit.
```
Original issue reported on code.google.com by `AHeinerm` on 29 Jul 2009 at 8:18 | defect | isidle and build queue isidle does not consider a building that is training a unit original issue reported on code google com by aheinerm on jul at | 1 |
593,518 | 18,010,100,434 | IssuesEvent | 2021-09-16 07:34:49 | vrchatapi/specification | https://api.github.com/repos/vrchatapi/specification | opened | Missing Select Fallback Avatar endpoint | Priority: Medium Status: Help Wanted Status: Accepted Type: Undocumented Endpoint | PUT `/avatars/<avatarId>/selectfallback`
Big :heart: to Natsumi for pointing this out | 1.0 | Missing Select Fallback Avatar endpoint - PUT `/avatars/<avatarId>/selectfallback`
Big :heart: to Natsumi for pointing this out | non_defect | missing select fallback avatar endpoint put avatars selectfallback big heart to natsumi for pointing this out | 0 |
16,754 | 2,941,602,420 | IssuesEvent | 2015-07-02 09:04:12 | schwabe/ics-openvpn | https://api.github.com/repos/schwabe/ics-openvpn | closed | Add a blacklist of network OpenVPN should not try to connect the VPN. | auto-migrated duplicate Priority-Medium Type-Defect | ```
I use OpenVPN to keep my phone connected to my home network while connected to
my work WiFi or my friend's Wifi or using my mobile network.
ics-openvpn works quite reliably.
The problems is when I arrive home: the phone connects to one my home Wifi
networks (I have two) and ics-openvpn tries to connect to the VPN but it does
not succeed (but, due to the way things are configured it can't connect, so
this is ok). The problem is that it keeps trying.
I think the solution should be having a blacklist so when connected to my wifi
it does not try to connect the VPN.
Thanks
A.
```
Original issue reported on code.google.com by `besang...@grumulla.net` on 1 Nov 2012 at 3:39 | 1.0 | Add a blacklist of network OpenVPN should not try to connect the VPN. - ```
I use OpenVPN to keep my phone connected to my home network while connected to
my work WiFi or my friend's Wifi or using my mobile network.
ics-openvpn works quite reliably.
The problems is when I arrive home: the phone connects to one my home Wifi
networks (I have two) and ics-openvpn tries to connect to the VPN but it does
not succeed (but, due to the way things are configured it can't connect, so
this is ok). The problem is that it keeps trying.
I think the solution should be having a blacklist so when connected to my wifi
it does not try to connect the VPN.
Thanks
A.
```
Original issue reported on code.google.com by `besang...@grumulla.net` on 1 Nov 2012 at 3:39 | defect | add a blacklist of network openvpn should not try to connect the vpn i use openvpn to keep my phone connected to my home network while connected to my work wifi or my friend s wifi or using my mobile network ics openvpn works quite reliably the problems is when i arrive home the phone connects to one my home wifi networks i have two and ics openvpn tries to connect to the vpn but it does not succeed but due to the way things are configured it can t connect so this is ok the problem is that it keeps trying i think the solution should be having a blacklist so when connected to my wifi it does not try to connect the vpn thanks a original issue reported on code google com by besang grumulla net on nov at | 1 |
7,795 | 2,610,636,996 | IssuesEvent | 2015-02-26 21:33:43 | alistairreilly/open-ig | https://api.github.com/repos/alistairreilly/open-ig | closed | UI hangok finomhangolása&elmaradt hangok pótlása | auto-migrated Milestone-0.93.500 Priority-Low Sound-Effects Type-Defect | ```
Belső menü:
Nem megfelelő hangok:
-Hangok-->Gombhangok
-Játékmenet-->Épületek automatikus javítása.
Ennél a kettőnél click_low_1.wav helyett click_medium_2.wav-ot kellene
bejátszani.
```
Original issue reported on code.google.com by `Jozsef.T...@gmail.com` on 25 Aug 2011 at 6:18 | 1.0 | UI hangok finomhangolása&elmaradt hangok pótlása - ```
Belső menü:
Nem megfelelő hangok:
-Hangok-->Gombhangok
-Játékmenet-->Épületek automatikus javítása.
Ennél a kettőnél click_low_1.wav helyett click_medium_2.wav-ot kellene
bejátszani.
```
Original issue reported on code.google.com by `Jozsef.T...@gmail.com` on 25 Aug 2011 at 6:18 | defect | ui hangok finomhangolása elmaradt hangok pótlása belső menü nem megfelelő hangok hangok gombhangok játékmenet épületek automatikus javítása ennél a kettőnél click low wav helyett click medium wav ot kellene bejátszani original issue reported on code google com by jozsef t gmail com on aug at | 1 |
24,034 | 3,900,663,559 | IssuesEvent | 2016-04-18 07:29:16 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | simple rms.rmsd() ignores weights option | Component-Analysis defect Difficulty-easy | The `analysis.rms.rmsd` function doesn't weight a calculated rmsd when the weights are given
### Currently version of MDAnalysis:
(run `python -c "import MDAnalysis as mda; print(mda.__version__)"`)
0.15dev
| 1.0 | simple rms.rmsd() ignores weights option - The `analysis.rms.rmsd` function doesn't weight a calculated rmsd when the weights are given
### Currently version of MDAnalysis:
(run `python -c "import MDAnalysis as mda; print(mda.__version__)"`)
0.15dev
| defect | simple rms rmsd ignores weights option the analysis rms rmsd function doesn t weight a calculated rmsd when the weights are given currently version of mdanalysis run python c import mdanalysis as mda print mda version | 1 |
78,699 | 3,513,191,475 | IssuesEvent | 2016-01-11 09:05:52 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | New Contribute Flow: # Resources | bug Priority-Medium | Currently it is displayed: # Resources (drag and drop to sort)
"#" should be updated and display the total number of resources. When adding a new resource, the number will be updated | 1.0 | New Contribute Flow: # Resources - Currently it is displayed: # Resources (drag and drop to sort)
"#" should be updated and display the total number of resources. When adding a new resource, the number will be updated | non_defect | new contribute flow resources currently it is displayed resources drag and drop to sort should be updated and display the total number of resources when adding a new resource the number will be updated | 0 |
57,001 | 15,591,279,603 | IssuesEvent | 2021-03-18 10:16:24 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | DataTable: unselect not clearing selection in PFX | 10.0.1 defect | **Describe the defect**
Testing out PFX. While following the migration guide, I noticed a problem with Datatable when performing a _unselect_.
In my datatables, I noticed that _selection_ was not updated even though the rowUnselect event listener was called properly. Also, when I clear the selection from the model in my backing bean and update/re-render the datatable, the row would remain selected.
**Environment:**
- PF Version: _10.0.0_
- JSF + version: Mojarra 2.2.17
- Affected browsers: _ALL_
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the Showcase
2. Navigate to Datatable - Selection
3. Scroll down to 'Single with Row Click'
4. Try to click select a row, and press 'View'. Afterwards, CTRL click the row, thereby unselecting it. When pressing 'View' again, it will still display the previously selected item.
**Expected behavior**
I expected the selection property of the datatable to be set to null, like in previous versions of PF.
**Sample code**
I'm not attaching any code since it can be reproduced in the showcase itself. I hope that's ok. | 1.0 | DataTable: unselect not clearing selection in PFX - **Describe the defect**
Testing out PFX. While following the migration guide, I noticed a problem with Datatable when performing a _unselect_.
In my datatables, I noticed that _selection_ was not updated even though the rowUnselect event listener was called properly. Also, when I clear the selection from the model in my backing bean and update/re-render the datatable, the row would remain selected.
**Environment:**
- PF Version: _10.0.0_
- JSF + version: Mojarra 2.2.17
- Affected browsers: _ALL_
**To Reproduce**
Steps to reproduce the behavior:
1. Go to the Showcase
2. Navigate to Datatable - Selection
3. Scroll down to 'Single with Row Click'
4. Try to click select a row, and press 'View'. Afterwards, CTRL click the row, thereby unselecting it. When pressing 'View' again, it will still display the previously selected item.
**Expected behavior**
I expected the selection property of the datatable to be set to null, like in previous versions of PF.
**Sample code**
I'm not attaching any code since it can be reproduced in the showcase itself. I hope that's ok. | defect | datatable unselect not clearing selection in pfx describe the defect testing out pfx while following the migration guide i noticed a problem with datatable when performing a unselect in my datatables i noticed that selection was not updated even though the rowunselect event listener was called properly also when i clear the selection from the model in my backing bean and update re render the datatable the row would remain selected environment pf version jsf version mojarra affected browsers all to reproduce steps to reproduce the behavior go to the showcase navigate to datatable selection scroll down to single with row click try to click select a row and press view afterwards ctrl click the row thereby unselecting it when pressing view again it will still display the previously selected item expected behavior i expected the selection property of the datatable to be set to null like in previous versions of pf sample code i m not attaching any code since it can be reproduced in the showcase itself i hope that s ok | 1 |
124,824 | 12,239,674,860 | IssuesEvent | 2020-05-04 22:11:26 | MikeHeiber/Excimontec | https://api.github.com/repos/MikeHeiber/Excimontec | closed | Add recommended reading section to README | documentation/tutorial | List high-quality review papers and book chapters to help new users learn how KMC simulations have been used in the organic electronics field in the past. | 1.0 | Add recommended reading section to README - List high-quality review papers and book chapters to help new users learn how KMC simulations have been used in the organic electronics field in the past. | non_defect | add recommended reading section to readme list high quality review papers and book chapters to help new users learn how kmc simulations have been used in the organic electronics field in the past | 0 |
25,844 | 4,472,844,374 | IssuesEvent | 2016-08-26 00:08:17 | zaproxy/zaproxy | https://api.github.com/repos/zaproxy/zaproxy | closed | Performing only a scan via the API can cause an infinite loop | Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
Start a scan via the API without first proxying any traffic or running a spider.
What is the expected output? What do you see instead?
I would expect the scan either to fail or to immediately reach 100%. Instead the scan
is triggered and the status always returns 0.
What version of the product are you using? On what operating system?
D-2013-04-08 on Mac OS X
```
Original issue reported on code.google.com by `dave.hunt` on 2013-04-16 10:05:06 | 1.0 | Performing only a scan via the API can cause an infinite loop - ```
What steps will reproduce the problem?
Start a scan via the API without first proxying any traffic or running a spider.
What is the expected output? What do you see instead?
I would expect the scan either to fail or to immediately reach 100%. Instead the scan
is triggered and the status always returns 0.
What version of the product are you using? On what operating system?
D-2013-04-08 on Mac OS X
```
Original issue reported on code.google.com by `dave.hunt` on 2013-04-16 10:05:06 | defect | performing only a scan via the api can cause an infinite loop what steps will reproduce the problem start a scan via the api without first proxying any traffic or running a spider what is the expected output what do you see instead i would expect the scan either to fail or to immediately reach instead the scan is triggered and the status always returns what version of the product are you using on what operating system d on mac os x original issue reported on code google com by dave hunt on | 1 |
428,714 | 12,415,614,769 | IssuesEvent | 2020-05-22 16:35:18 | TykTechnologies/tyk-pump | https://api.github.com/repos/TykTechnologies/tyk-pump | closed | Prometheus metric "tyk_http_status_per_key" not getting incremented | Priority: Low sla wontfix | Version:
tyk-pump: 0.8.1
Prometheus metric "tyk_http_status_per_key" not getting incremented
Even after making multiple API calls
tyk_http_status_per_key{code="201",key="00000000"} 0
tyk_http_status_per_key{code="400",key="00000000"} 0
count remains 0 for all status codes , whereas is should have been incremented accordingly
e.g.
tyk_http_status_per_key{code="201",key="00000000"} 5
tyk_http_status_per_key{code="400",key="00000000"} 22
| 1.0 | Prometheus metric "tyk_http_status_per_key" not getting incremented - Version:
tyk-pump: 0.8.1
Prometheus metric "tyk_http_status_per_key" not getting incremented
Even after making multiple API calls
tyk_http_status_per_key{code="201",key="00000000"} 0
tyk_http_status_per_key{code="400",key="00000000"} 0
count remains 0 for all status codes , whereas is should have been incremented accordingly
e.g.
tyk_http_status_per_key{code="201",key="00000000"} 5
tyk_http_status_per_key{code="400",key="00000000"} 22
| non_defect | prometheus metric tyk http status per key not getting incremented version tyk pump prometheus metric tyk http status per key not getting incremented even after making multiple api calls tyk http status per key code key tyk http status per key code key count remains for all status codes whereas is should have been incremented accordingly e g tyk http status per key code key tyk http status per key code key | 0 |
7,388 | 2,601,760,591 | IssuesEvent | 2015-02-24 00:34:55 | chrsmith/bwapi | https://api.github.com/repos/chrsmith/bwapi | closed | Const correctness | auto-migrated Component-Persistence Priority-None Type-Task | ```
Many functions should be made constant, however it's more of a tedius chore.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 25 Feb 2011 at 10:35 | 1.0 | Const correctness - ```
Many functions should be made constant, however it's more of a tedius chore.
```
-----
Original issue reported on code.google.com by `AHeinerm` on 25 Feb 2011 at 10:35 | non_defect | const correctness many functions should be made constant however it s more of a tedius chore original issue reported on code google com by aheinerm on feb at | 0 |
67,130 | 8,074,999,014 | IssuesEvent | 2018-08-07 02:48:14 | GitHawkApp/GitHawk | https://api.github.com/repos/GitHawkApp/GitHawk | closed | Resign keyboard when tapping manage button | design starter-task | Kind of annoying to have the kb flying around
<details>
<summary>Bug Report Dump (Auto-generated)</summary>
<pre>
Version 1.22.0 (1532006598)
Device: iPhone X (iOS 11.4.1)
TestFlight: true
</pre>
</details> | 1.0 | Resign keyboard when tapping manage button - Kind of annoying to have the kb flying around
<details>
<summary>Bug Report Dump (Auto-generated)</summary>
<pre>
Version 1.22.0 (1532006598)
Device: iPhone X (iOS 11.4.1)
TestFlight: true
</pre>
</details> | non_defect | resign keyboard when tapping manage button kind of annoying to have the kb flying around bug report dump auto generated version device iphone x ios testflight true | 0 |
143,284 | 21,993,500,761 | IssuesEvent | 2022-05-26 02:11:45 | harryodubhghaill/CI-Portforlio-4-blogsocial | https://api.github.com/repos/harryodubhghaill/CI-Portforlio-4-blogsocial | closed | DESIGN: Group Page | Design | **Full group page**
**Page to hold a list of posts specific to that group**
**Criteria**
- [ ] Group Name
- [ ] Group Description
- [ ] Follow group button
- [ ] Create post button
- [ ] Paginated List of posts (similar to home page list #26)
- [ ] Number of members
| 1.0 | DESIGN: Group Page - **Full group page**
**Page to hold a list of posts specific to that group**
**Criteria**
- [ ] Group Name
- [ ] Group Description
- [ ] Follow group button
- [ ] Create post button
- [ ] Paginated List of posts (similar to home page list #26)
- [ ] Number of members
| non_defect | design group page full group page page to hold a list of posts specific to that group criteria group name group description follow group button create post button paginated list of posts similar to home page list number of members | 0 |
95,581 | 27,556,602,026 | IssuesEvent | 2023-03-07 18:25:35 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Build failure: pkgsCross.raspberryPi.hello (from aarch64-linux) | 0.kind: build failure | ### Steps To Reproduce
See also https://discourse.nixos.org/t/cross-compiling-hello-to-raspberrypi/26093
Steps to reproduce the behavior:
1. `nix build -f '<nixpkgs>' pkgsCross.raspberryPi.hello`
### Build log
https://gist.github.com/eliasnaur/b57b5c0337ba4fdf17271708db252b45
### Additional context
This is such as basic cross-compilation that I wonder whether something is wrong in my setup. However, the `exec format` errors are surprising, because I thought most packages were cross-compilation aware enough to not run host binaries on the build platform.
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
$ nix-shell -p nix-info --run "nix-info -m"
- system: `"aarch64-linux"`
- host os: `Linux 5.15.97, NixOS, 22.11 (Raccoon), 22.11.20230306.47c0034`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.11.1`
- channels(e): `""`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
```
| 1.0 | Build failure: pkgsCross.raspberryPi.hello (from aarch64-linux) - ### Steps To Reproduce
See also https://discourse.nixos.org/t/cross-compiling-hello-to-raspberrypi/26093
Steps to reproduce the behavior:
1. `nix build -f '<nixpkgs>' pkgsCross.raspberryPi.hello`
### Build log
https://gist.github.com/eliasnaur/b57b5c0337ba4fdf17271708db252b45
### Additional context
This is such as basic cross-compilation that I wonder whether something is wrong in my setup. However, the `exec format` errors are surprising, because I thought most packages were cross-compilation aware enough to not run host binaries on the build platform.
### Metadata
Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result.
```console
$ nix-shell -p nix-info --run "nix-info -m"
- system: `"aarch64-linux"`
- host os: `Linux 5.15.97, NixOS, 22.11 (Raccoon), 22.11.20230306.47c0034`
- multi-user?: `yes`
- sandbox: `yes`
- version: `nix-env (Nix) 2.11.1`
- channels(e): `""`
- nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos`
```
| non_defect | build failure pkgscross raspberrypi hello from linux steps to reproduce see also steps to reproduce the behavior nix build f pkgscross raspberrypi hello build log additional context this is such as basic cross compilation that i wonder whether something is wrong in my setup however the exec format errors are surprising because i thought most packages were cross compilation aware enough to not run host binaries on the build platform metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels e nixpkgs nix var nix profiles per user root channels nixos | 0 |
78,940 | 27,827,475,883 | IssuesEvent | 2023-03-19 22:43:07 | scoutplan/scoutplan | https://api.github.com/repos/scoutplan/scoutplan | opened | [Scoutplan Production/production] KeyError: key not found: "OPENWEATHER_API_KEY" | defect | ## Backtrace
line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: fetch
line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: api
line 21 of [PROJECT_ROOT]/app/services/weather_service.rb: current
[View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/94565340) | 1.0 | [Scoutplan Production/production] KeyError: key not found: "OPENWEATHER_API_KEY" - ## Backtrace
line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: fetch
line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: api
line 21 of [PROJECT_ROOT]/app/services/weather_service.rb: current
[View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/94565340) | defect | keyerror key not found openweather api key backtrace line of app services weather service rb fetch line of app services weather service rb api line of app services weather service rb current | 1 |
403,745 | 11,845,778,247 | IssuesEvent | 2020-03-24 09:01:02 | AY1920S2-CS2103T-F09-4/main | https://api.github.com/repos/AY1920S2-CS2103T-F09-4/main | closed | Create Json adapted objects for data storage | priority.High type.Epic | - [ ] JsonAdaptedActivity
- [ ] JsonAdaptedItinerary
- [ ] JsonAdaptedEntry
- [ ] JsonAdaptedJournal
| 1.0 | Create Json adapted objects for data storage - - [ ] JsonAdaptedActivity
- [ ] JsonAdaptedItinerary
- [ ] JsonAdaptedEntry
- [ ] JsonAdaptedJournal
| non_defect | create json adapted objects for data storage jsonadaptedactivity jsonadapteditinerary jsonadaptedentry jsonadaptedjournal | 0 |
684,948 | 23,439,366,576 | IssuesEvent | 2022-08-15 13:29:30 | Plutonomicon/cardano-transaction-lib | https://api.github.com/repos/Plutonomicon/cardano-transaction-lib | closed | Improve AlwaysSucceeds example | enhancement good first issue lower-priority needs investigation | `countToZero` is a boring way to ensure a tx passes. We can skip some of the waiting by checking utxos available at script address.
This code was suggested by Brian:
```
waitForTx :: Int -> ValidatorHash -> TransactionHash -> Contract () (Maybe TransactionInput)
waitForTx n vhash txid = do
let scriptAddress = scriptHashAddress vhash
UtxoM utxos <- fromMaybe (UtxoM Map.empty) <$> utxosAt scriptAddress
case fst <$> find hasTransactionId (Map.toUnfoldable utxos :: Array (TransactionInput /\ TransactionOutput)) of
Nothing ->
if (n <= 0)
then do
pure Nothing
else do
logInfo' $ "No tx yet, waiting for: " <> show n <> " more seconds"
(liftAff <<< delay <<< wrap) 1000.0
waitForTx (n - 1) vhash txid
Just txin -> do
logInfo' $ "found tx:" <> show txid
pure $ Just txin
where
hasTransactionId :: forall a. TransactionInput /\ a -> Boolean
hasTransactionId (TransactionInput tx /\ _) =
tx.transactionId == txid
``` | 1.0 | Improve AlwaysSucceeds example - `countToZero` is a boring way to ensure a tx passes. We can skip some of the waiting by checking utxos available at script address.
This code was suggested by Brian:
```
waitForTx :: Int -> ValidatorHash -> TransactionHash -> Contract () (Maybe TransactionInput)
waitForTx n vhash txid = do
let scriptAddress = scriptHashAddress vhash
UtxoM utxos <- fromMaybe (UtxoM Map.empty) <$> utxosAt scriptAddress
case fst <$> find hasTransactionId (Map.toUnfoldable utxos :: Array (TransactionInput /\ TransactionOutput)) of
Nothing ->
if (n <= 0)
then do
pure Nothing
else do
logInfo' $ "No tx yet, waiting for: " <> show n <> " more seconds"
(liftAff <<< delay <<< wrap) 1000.0
waitForTx (n - 1) vhash txid
Just txin -> do
logInfo' $ "found tx:" <> show txid
pure $ Just txin
where
hasTransactionId :: forall a. TransactionInput /\ a -> Boolean
hasTransactionId (TransactionInput tx /\ _) =
tx.transactionId == txid
``` | non_defect | improve alwayssucceeds example counttozero is a boring way to ensure a tx passes we can skip some of the waiting by checking utxos available at script address this code was suggested by brian waitfortx int validatorhash transactionhash contract maybe transactioninput waitfortx n vhash txid do let scriptaddress scripthashaddress vhash utxom utxos utxosat scriptaddress case fst find hastransactionid map tounfoldable utxos array transactioninput transactionoutput of nothing if n then do pure nothing else do loginfo no tx yet waiting for show n more seconds liftaff delay wrap waitfortx n vhash txid just txin do loginfo found tx show txid pure just txin where hastransactionid forall a transactioninput a boolean hastransactionid transactioninput tx tx transactionid txid | 0 |
46,689 | 13,055,959,860 | IssuesEvent | 2020-07-30 03:14:26 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | trigger-sim needs specially keyed volume trigger for each instance (Trac #1717) | Incomplete Migration Migrated from Trac combo simulation defect | Migrated from https://code.icecube.wisc.edu/ticket/1717
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:58",
"description": "When running the trigger-sim segment on hitspool data younger than May 9th 2016, I get the following error: \n`RuntimeError: Tray already contains module named \"InIceRawdata_triggersim_cyl\" of type CylinderTrigger (in I3Tray::param_setter I3Tray::AddModule(boost::python::api::object, std::string))`\n\nWhich is caused (I guess) by multiple volume trigger instances in the frame. Would be nice to have the same behaviour as we have for multiple SMTs already as well for multiple clyinder triggers.",
"reporter": "dheereman",
"cc": "",
"resolution": "fixed",
"_ts": "1550067178841456",
"component": "combo simulation",
"summary": "trigger-sim needs specially keyed volume trigger for each instance",
"priority": "critical",
"keywords": "trigger-sim, multiple volume rigger instances",
"time": "2016-05-30T20:05:37",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| 1.0 | trigger-sim needs specially keyed volume trigger for each instance (Trac #1717) - Migrated from https://code.icecube.wisc.edu/ticket/1717
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:58",
"description": "When running the trigger-sim segment on hitspool data younger than May 9th 2016, I get the following error: \n`RuntimeError: Tray already contains module named \"InIceRawdata_triggersim_cyl\" of type CylinderTrigger (in I3Tray::param_setter I3Tray::AddModule(boost::python::api::object, std::string))`\n\nWhich is caused (I guess) by multiple volume trigger instances in the frame. Would be nice to have the same behaviour as we have for multiple SMTs already as well for multiple clyinder triggers.",
"reporter": "dheereman",
"cc": "",
"resolution": "fixed",
"_ts": "1550067178841456",
"component": "combo simulation",
"summary": "trigger-sim needs specially keyed volume trigger for each instance",
"priority": "critical",
"keywords": "trigger-sim, multiple volume rigger instances",
"time": "2016-05-30T20:05:37",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
| defect | trigger sim needs specially keyed volume trigger for each instance trac migrated from json status closed changetime description when running the trigger sim segment on hitspool data younger than may i get the following error n runtimeerror tray already contains module named inicerawdata triggersim cyl of type cylindertrigger in param setter addmodule boost python api object std string n nwhich is caused i guess by multiple volume trigger instances in the frame would be nice to have the same behaviour as we have for multiple smts already as well for multiple clyinder triggers reporter dheereman cc resolution fixed ts component combo simulation summary trigger sim needs specially keyed volume trigger for each instance priority critical keywords trigger sim multiple volume rigger instances time milestone owner olivas type defect | 1 |
51,547 | 13,207,523,854 | IssuesEvent | 2020-08-14 23:26:36 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | examples project can not be built without phys-services (Trac #604) | IceTray Incomplete Migration Migrated from Trac defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/604">https://code.icecube.wisc.edu/projects/icecube/ticket/604</a>, reported by blaufussand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T20:52:58",
"_ts": "1351716778000000",
"description": "The examples project (now in icetray svn)\ncan not be build or cmake'd with out the icecube phys-services\nproject.\n\nNeed to cleave these better to examples can serve all by itself.\n\nSeveral of the modules in the examples project use the phys-services code (I3Calculator) and the scripts load phys-services libraries.",
"reporter": "blaufuss",
"cc": "",
"resolution": "wont or cant fix",
"time": "2010-03-11T15:28:56",
"component": "IceTray",
"summary": "examples project can not be built without phys-services",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | examples project can not be built without phys-services (Trac #604) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/604">https://code.icecube.wisc.edu/projects/icecube/ticket/604</a>, reported by blaufussand owned by troy</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T20:52:58",
"_ts": "1351716778000000",
"description": "The examples project (now in icetray svn)\ncan not be build or cmake'd with out the icecube phys-services\nproject.\n\nNeed to cleave these better to examples can serve all by itself.\n\nSeveral of the modules in the examples project use the phys-services code (I3Calculator) and the scripts load phys-services libraries.",
"reporter": "blaufuss",
"cc": "",
"resolution": "wont or cant fix",
"time": "2010-03-11T15:28:56",
"component": "IceTray",
"summary": "examples project can not be built without phys-services",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| defect | examples project can not be built without phys services trac migrated from json status closed changetime ts description the examples project now in icetray svn ncan not be build or cmake d with out the icecube phys services nproject n nneed to cleave these better to examples can serve all by itself n nseveral of the modules in the examples project use the phys services code and the scripts load phys services libraries reporter blaufuss cc resolution wont or cant fix time component icetray summary examples project can not be built without phys services priority normal keywords milestone owner troy type defect | 1 |
687,192 | 23,516,387,243 | IssuesEvent | 2022-08-18 21:55:42 | internetarchive/openlibrary | https://api.github.com/repos/internetarchive/openlibrary | closed | Don't allow merging a work with more than 50 editions into another work | Type: Bug Priority: 2 Affects: Librarians Needs: Investigation Lead: @cdrini | Sub-task of #6556
The Merge UI currently allows this, but only moves over the first 50 editions! Leaving the rest in a sort of limbo.
Prevent this from being allowed, or investigate if we can increase it to a larger limit, but we'll probably still need to block at whatever limit that is.
### Stakeholders
@mheiman @dcapillae | 1.0 | Don't allow merging a work with more than 50 editions into another work - Sub-task of #6556
The Merge UI currently allows this, but only moves over the first 50 editions! Leaving the rest in a sort of limbo.
Prevent this from being allowed, or investigate if we can increase it to a larger limit, but we'll probably still need to block at whatever limit that is.
### Stakeholders
@mheiman @dcapillae | non_defect | don t allow merging a work with more than editions into another work sub task of the merge ui currently allows this but only moves over the first editions leaving the rest in a sort of limbo prevent this from being allowed or investigate if we can increase it to a larger limit but we ll probably still need to block at whatever limit that is stakeholders mheiman dcapillae | 0 |
583,615 | 17,393,960,522 | IssuesEvent | 2021-08-02 11:03:54 | JonasMuehlmann/BIB_VPR | https://api.github.com/repos/JonasMuehlmann/BIB_VPR | opened | Role Permissions | bug frontend priority: medium | You can not save the changes for a Role when you add more then 2 Permissions. | 1.0 | Role Permissions - You can not save the changes for a Role when you add more then 2 Permissions. | non_defect | role permissions you can not save the changes for a role when you add more then permissions | 0 |
10,858 | 2,622,193,654 | IssuesEvent | 2015-03-04 00:24:12 | byzhang/cudpp | https://api.github.com/repos/byzhang/cudpp | opened | cudppPlan execution time | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. run cudppPlan in any function
What is the expected output? What do you see instead?
The output is correct (all tests in cudpp_testrig passed), but the first call
of the function cudppPlan is slow (about a second).
What version of the product are you using? On what operating system?
Motherboard Rampage III GENE
Intel(R) Core(TM) i7 CPU X 980 @ 3.33GHz
DIMM 1333 MHz 6GiB
GeForce GTX680
Ubuntu 12.04
Cudpp 2.0
Cuda 5.0
Nvidia Driver version 310.19
Please provide any additional information below.
I mentioned an "infinite loop" issue in cudppPlan here:
https://groups.google.com/forum/?fromgroups=#!topic/cudpp/J4AEijEFzW4
The two problems seem to be related somehow: if I make an exit(2) just before
the cudppPlan call (in simpleCUDPP.cu for example), the execution time is
normal. If I make it on the first line of cudppPlan, it takes about 0.8 second
to make the exit on my GTX680 card.
I also have a GT240 card. The same problem happens but it takes about 0.5
second.
I just made a new compilation of the cudpp library and the execution time of
simpleCUDPP was about 15 seconds. A second run and this is about one second
again.
```
Original issue reported on code.google.com by `nicolas....@gmail.com` on 10 Dec 2012 at 6:46 | 1.0 | cudppPlan execution time - ```
What steps will reproduce the problem?
1. run cudppPlan in any function
What is the expected output? What do you see instead?
The output is correct (all tests in cudpp_testrig passed), but the first call
of the function cudppPlan is slow (about a second).
What version of the product are you using? On what operating system?
Motherboard Rampage III GENE
Intel(R) Core(TM) i7 CPU X 980 @ 3.33GHz
DIMM 1333 MHz 6GiB
GeForce GTX680
Ubuntu 12.04
Cudpp 2.0
Cuda 5.0
Nvidia Driver version 310.19
Please provide any additional information below.
I mentioned an "infinite loop" issue in cudppPlan here:
https://groups.google.com/forum/?fromgroups=#!topic/cudpp/J4AEijEFzW4
The two problems seem to be related somehow: if I make an exit(2) just before
the cudppPlan call (in simpleCUDPP.cu for example), the execution time is
normal. If I make it on the first line of cudppPlan, it takes about 0.8 second
to make the exit on my GTX680 card.
I also have a GT240 card. The same problem happens but it takes about 0.5
second.
I just made a new compilation of the cudpp library and the execution time of
simpleCUDPP was about 15 seconds. A second run and this is about one second
again.
```
Original issue reported on code.google.com by `nicolas....@gmail.com` on 10 Dec 2012 at 6:46 | defect | cudppplan execution time what steps will reproduce the problem run cudppplan in any function what is the expected output what do you see instead the output is correct all tests in cudpp testrig passed but the first call of the function cudppplan is slow about a second what version of the product are you using on what operating system motherboard rampage iii gene intel r core tm cpu x dimm mhz geforce ubuntu cudpp cuda nvidia driver version please provide any additional information below i mentioned an infinite loop issue in cudppplan here the two problems seem to be related somehow if i make an exit just before the cudppplan call in simplecudpp cu for example the execution time is normal if i make it on the first line of cudppplan it takes about second to make the exit on my card i also have a card the same problem happens but it takes about second i just made a new compilation of the cudpp library and the execution time of simplecudpp was about seconds a second run and this is about one second again original issue reported on code google com by nicolas gmail com on dec at | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.