Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
55,881
3,075,100,780
IssuesEvent
2015-08-20 11:39:54
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
closed
Opera не перехватывает ссылки
bug imported Priority-Medium Usability
_From [stupidcu...@gmail.com](https://code.google.com/u/109602399200609281346/) on February 16, 2011 13:31:56_ What steps will reproduce the problem? В браузере Opera при нажатии ссылки пишет отсутствует связанное приложение для данного расширение, хотя в флайлинке стоит галочка перехватывать ссылки What version of the product are you using? On what operating system? FlylinkDC-x64- r500 -beta74-build-6047 Windows 7 x64 **Attachment:** [bug.jpg](http://code.google.com/p/flylinkdc/issues/detail?id=362) _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=362_
1.0
Opera не перехватывает ссылки - _From [stupidcu...@gmail.com](https://code.google.com/u/109602399200609281346/) on February 16, 2011 13:31:56_ What steps will reproduce the problem? В браузере Opera при нажатии ссылки пишет отсутствует связанное приложение для данного расширение, хотя в флайлинке стоит галочка перехватывать ссылки What version of the product are you using? On what operating system? FlylinkDC-x64- r500 -beta74-build-6047 Windows 7 x64 **Attachment:** [bug.jpg](http://code.google.com/p/flylinkdc/issues/detail?id=362) _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=362_
non_defect
opera не перехватывает ссылки from on february what steps will reproduce the problem в браузере opera при нажатии ссылки пишет отсутствует связанное приложение для данного расширение хотя в флайлинке стоит галочка перехватывать ссылки what version of the product are you using on what operating system flylinkdc build windows attachment original issue
0
22,483
4,807,153,139
IssuesEvent
2016-11-02 20:36:53
ipfs/go-ipfs
https://api.github.com/repos/ipfs/go-ipfs
closed
Inconsistent usage of IPFS and ipfs
documentation ready
We generally agreed on fact that capitalised IPFS should be used in reference to project and ipfs in reference to CLI or implementation. Unfortunately now we are inconsistent across codebase. One of great examples is help text of `ipfs tar`: ```  ipfs tar -h USAGE ipfs tar - Utility functions for tar files in IPFS. SUBCOMMANDS ipfs tar add <file> - Import a tar file into ipfs. ipfs tar cat <path> - Export a tar file from IPFS. Use 'ipfs tar --help' for more information about this command. ```
1.0
Inconsistent usage of IPFS and ipfs - We generally agreed on fact that capitalised IPFS should be used in reference to project and ipfs in reference to CLI or implementation. Unfortunately now we are inconsistent across codebase. One of great examples is help text of `ipfs tar`: ```  ipfs tar -h USAGE ipfs tar - Utility functions for tar files in IPFS. SUBCOMMANDS ipfs tar add <file> - Import a tar file into ipfs. ipfs tar cat <path> - Export a tar file from IPFS. Use 'ipfs tar --help' for more information about this command. ```
non_defect
inconsistent usage of ipfs and ipfs we generally agreed on fact that capitalised ipfs should be used in reference to project and ipfs in reference to cli or implementation unfortunately now we are inconsistent across codebase one of great examples is help text of ipfs tar  ipfs tar h usage ipfs tar utility functions for tar files in ipfs subcommands ipfs tar add import a tar file into ipfs ipfs tar cat export a tar file from ipfs use ipfs tar help for more information about this command
0
29,389
5,664,009,102
IssuesEvent
2017-04-11 00:26:12
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
MDAnalysisTests raises exception when imported
defect testing
### Expected behaviour ```python import MDAnalysisTests ``` and ```python import MDAnalysis.tests ``` should import (and make eg files available in `datafiles`. ### Actual behaviour Both imports fail with ``` In [3]: import MDAnalysisTests --------------------------------------------------------------------------- Exception Traceback (most recent call last) <ipython-input-3-52912a4f0416> in <module>() ----> 1 import MDAnalysisTests /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/__init__.py in <module>() 141 pass 142 --> 143 from MDAnalysisTests.util import ( 144 block_import, 145 executable_not_found, /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/util.py in <module>() 37 from functools import wraps 38 import importlib ---> 39 import mock 40 import os 41 /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/__init__.py in <module>() 1 from __future__ import absolute_import ----> 2 import mock.mock as _mock 3 from mock.mock import * 4 __all__ = _mock.__all__ 5 #import mock.mock as _mock /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/mock.py in <module>() 69 from pbr.version import VersionInfo 70 ---> 71 _v = VersionInfo('mock').semantic_version() 72 __version__ = _v.release_string() 73 version_info = _v.version_tuple() /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in semantic_version(self) 458 """Return the SemanticVersion object for this version.""" 459 if self._semantic is None: --> 460 self._semantic = self._get_version_from_pkg_resources() 461 return self._semantic 462 /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in _get_version_from_pkg_resources(self) 445 # installed into anything. Revert to setup-time logic. 446 from pbr import packaging --> 447 result_string = packaging.get_version(self.package) 448 return SemanticVersion.from_pip_string(result_string) 449 /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/packaging.pyc in get_version(package_name, pre_version) 748 " to pbr.version.VersionInfo. Project name {name} was" 749 " given, but was not able to be found.".format( --> 750 name=package_name)) 751 752 Exception: Versioning for this project requires either an sdist tarball, or access to an upstream git repository. It's also possible that there is a mismatch between the package name in setup.cfg and the argument given to pbr.version.VersionInfo. Project name mock was given, but was not able to be found. ``` ### Code to reproduce the behaviour ```python import MDAnalysisTests ``` ### Currently version of MDAnalysis: (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.16.0 (pip upgraded in a virtualenv)
1.0
MDAnalysisTests raises exception when imported - ### Expected behaviour ```python import MDAnalysisTests ``` and ```python import MDAnalysis.tests ``` should import (and make eg files available in `datafiles`. ### Actual behaviour Both imports fail with ``` In [3]: import MDAnalysisTests --------------------------------------------------------------------------- Exception Traceback (most recent call last) <ipython-input-3-52912a4f0416> in <module>() ----> 1 import MDAnalysisTests /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/__init__.py in <module>() 141 pass 142 --> 143 from MDAnalysisTests.util import ( 144 block_import, 145 executable_not_found, /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/MDAnalysisTests/util.py in <module>() 37 from functools import wraps 38 import importlib ---> 39 import mock 40 import os 41 /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/__init__.py in <module>() 1 from __future__ import absolute_import ----> 2 import mock.mock as _mock 3 from mock.mock import * 4 __all__ = _mock.__all__ 5 #import mock.mock as _mock /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/mock/mock.py in <module>() 69 from pbr.version import VersionInfo 70 ---> 71 _v = VersionInfo('mock').semantic_version() 72 __version__ = _v.release_string() 73 version_info = _v.version_tuple() /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in semantic_version(self) 458 """Return the SemanticVersion object for this version.""" 459 if self._semantic is None: --> 460 self._semantic = self._get_version_from_pkg_resources() 461 return self._semantic 462 /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/version.pyc in _get_version_from_pkg_resources(self) 445 # installed into anything. Revert to setup-time logic. 446 from pbr import packaging --> 447 result_string = packaging.get_version(self.package) 448 return SemanticVersion.from_pip_string(result_string) 449 /Users/oliver/.virtualenvs/mda_clean/lib/python2.7/site-packages/pbr/packaging.pyc in get_version(package_name, pre_version) 748 " to pbr.version.VersionInfo. Project name {name} was" 749 " given, but was not able to be found.".format( --> 750 name=package_name)) 751 752 Exception: Versioning for this project requires either an sdist tarball, or access to an upstream git repository. It's also possible that there is a mismatch between the package name in setup.cfg and the argument given to pbr.version.VersionInfo. Project name mock was given, but was not able to be found. ``` ### Code to reproduce the behaviour ```python import MDAnalysisTests ``` ### Currently version of MDAnalysis: (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.16.0 (pip upgraded in a virtualenv)
defect
mdanalysistests raises exception when imported expected behaviour python import mdanalysistests and python import mdanalysis tests should import and make eg files available in datafiles actual behaviour both imports fail with in import mdanalysistests exception traceback most recent call last in import mdanalysistests users oliver virtualenvs mda clean lib site packages mdanalysistests init py in pass from mdanalysistests util import block import executable not found users oliver virtualenvs mda clean lib site packages mdanalysistests util py in from functools import wraps import importlib import mock import os users oliver virtualenvs mda clean lib site packages mock init py in from future import absolute import import mock mock as mock from mock mock import all mock all import mock mock as mock users oliver virtualenvs mda clean lib site packages mock mock py in from pbr version import versioninfo v versioninfo mock semantic version version v release string version info v version tuple users oliver virtualenvs mda clean lib site packages pbr version pyc in semantic version self return the semanticversion object for this version if self semantic is none self semantic self get version from pkg resources return self semantic users oliver virtualenvs mda clean lib site packages pbr version pyc in get version from pkg resources self installed into anything revert to setup time logic from pbr import packaging result string packaging get version self package return semanticversion from pip string result string users oliver virtualenvs mda clean lib site packages pbr packaging pyc in get version package name pre version to pbr version versioninfo project name name was given but was not able to be found format name package name exception versioning for this project requires either an sdist tarball or access to an upstream git repository it s also possible that there is a mismatch between the package name in setup cfg and the argument given to pbr version versioninfo project name mock was given but was not able to be found code to reproduce the behaviour python import mdanalysistests currently version of mdanalysis run python c import mdanalysis as mda print mda version pip upgraded in a virtualenv
1
59,999
17,023,307,846
IssuesEvent
2021-07-03 01:20:56
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
highway=motorway_junction nodes render with an incorrect font - t@h
Component: osmarender Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 8.35am, Thursday, 9th October 2008]** Some stray t@h clients are rendering these nodes with an incorrect font. Some examples of this can be found at http://wiki.openstreetmap.org/index.php?title=User:Chriscf/bad_renders In most of the cases, it can be seen that only highway=motorway_junction nodes are affected. It would seem that this one style, and no others, specifies "sans-serif" as a fallback font. It is unclear *why* some clients are falling back in this case, but it is causing undesirable results, especially on those captions that cross tile boundaries. In the meantime, proposing that sans-serif be removed from this one feature, since there doesn't appear to be a reason why this feature and this one alone is marked out in this way.
1.0
highway=motorway_junction nodes render with an incorrect font - t@h - **[Submitted to the original trac issue database at 8.35am, Thursday, 9th October 2008]** Some stray t@h clients are rendering these nodes with an incorrect font. Some examples of this can be found at http://wiki.openstreetmap.org/index.php?title=User:Chriscf/bad_renders In most of the cases, it can be seen that only highway=motorway_junction nodes are affected. It would seem that this one style, and no others, specifies "sans-serif" as a fallback font. It is unclear *why* some clients are falling back in this case, but it is causing undesirable results, especially on those captions that cross tile boundaries. In the meantime, proposing that sans-serif be removed from this one feature, since there doesn't appear to be a reason why this feature and this one alone is marked out in this way.
defect
highway motorway junction nodes render with an incorrect font t h some stray t h clients are rendering these nodes with an incorrect font some examples of this can be found at in most of the cases it can be seen that only highway motorway junction nodes are affected it would seem that this one style and no others specifies sans serif as a fallback font it is unclear why some clients are falling back in this case but it is causing undesirable results especially on those captions that cross tile boundaries in the meantime proposing that sans serif be removed from this one feature since there doesn t appear to be a reason why this feature and this one alone is marked out in this way
1
594,594
18,049,158,666
IssuesEvent
2021-09-19 12:40:45
codeforcroatia/imamopravoznati-theme
https://api.github.com/repos/codeforcroatia/imamopravoznati-theme
closed
Update donate URL
tasks ⚡ priority
**Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is; e.g. I'm always frustrated when [...] --> **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> New URL (new WP web): https://gong.hr/tko-smo/podrzi-gong/ **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
1.0
Update donate URL - **Is your feature request related to a problem? Please describe.** <!-- A clear and concise description of what the problem is; e.g. I'm always frustrated when [...] --> **Describe the solution you'd like** <!-- A clear and concise description of what you want to happen. --> New URL (new WP web): https://gong.hr/tko-smo/podrzi-gong/ **Describe alternatives you've considered** <!-- A clear and concise description of any alternative solutions or features you've considered. --> **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
non_defect
update donate url is your feature request related to a problem please describe a clear and concise description of what the problem is e g i m always frustrated when describe the solution you d like a clear and concise description of what you want to happen new url new wp web describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
0
203,022
15,340,155,633
IssuesEvent
2021-02-27 05:32:59
momentum-mod/game
https://api.github.com/repos/momentum-mod/game
closed
Rendering issues when alt-tabbing in fullscreen
Blocked: Needs testing & verification Type: Bug
**Describe the bug** ALt-tabbing in momentum on chaos causes the game to flicker and/or freeze for roughly half a second, Happens anywhere in the game. Under some circumstances it can show the setting panel for half a second even if it was not open prior to alt-tabbing. **To Reproduce** 1. set game to fullscreen 2. alt-tab out and in, notice flickering/freezes (freezes seem to happen more often when in map compared to main menu) **Expected behavior** No visual corruption. **Desktop/Branch (please complete the following information):** - OS: Windows - Branch: 0.8.8-rc **Additional context** Does not happened in windowed mode. See if `-d3d9ex` changes anything since I know that affects alt-tabbing in other source games to some degree.
1.0
Rendering issues when alt-tabbing in fullscreen - **Describe the bug** ALt-tabbing in momentum on chaos causes the game to flicker and/or freeze for roughly half a second, Happens anywhere in the game. Under some circumstances it can show the setting panel for half a second even if it was not open prior to alt-tabbing. **To Reproduce** 1. set game to fullscreen 2. alt-tab out and in, notice flickering/freezes (freezes seem to happen more often when in map compared to main menu) **Expected behavior** No visual corruption. **Desktop/Branch (please complete the following information):** - OS: Windows - Branch: 0.8.8-rc **Additional context** Does not happened in windowed mode. See if `-d3d9ex` changes anything since I know that affects alt-tabbing in other source games to some degree.
non_defect
rendering issues when alt tabbing in fullscreen describe the bug alt tabbing in momentum on chaos causes the game to flicker and or freeze for roughly half a second happens anywhere in the game under some circumstances it can show the setting panel for half a second even if it was not open prior to alt tabbing to reproduce set game to fullscreen alt tab out and in notice flickering freezes freezes seem to happen more often when in map compared to main menu expected behavior no visual corruption desktop branch please complete the following information os windows branch rc additional context does not happened in windowed mode see if changes anything since i know that affects alt tabbing in other source games to some degree
0
353,370
10,552,050,208
IssuesEvent
2019-10-03 14:28:40
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
closed
pass on ?to url to edit account profile page
component: firefox accounts priority: p3 state: pull request ready
Needed for mozilla/addons-frontend#8666 When a new user is redirected to the edit account profile page on frontend, we should pass along the intended destination url as a GET parameter in the url. Note: currently not all login/registration requests from frontend result in a new user being redirected to the edit page - #12417 will change that but we can't do that until frontend supports the `?to=xxx` param.
1.0
pass on ?to url to edit account profile page - Needed for mozilla/addons-frontend#8666 When a new user is redirected to the edit account profile page on frontend, we should pass along the intended destination url as a GET parameter in the url. Note: currently not all login/registration requests from frontend result in a new user being redirected to the edit page - #12417 will change that but we can't do that until frontend supports the `?to=xxx` param.
non_defect
pass on to url to edit account profile page needed for mozilla addons frontend when a new user is redirected to the edit account profile page on frontend we should pass along the intended destination url as a get parameter in the url note currently not all login registration requests from frontend result in a new user being redirected to the edit page will change that but we can t do that until frontend supports the to xxx param
0
44,091
11,962,106,747
IssuesEvent
2020-04-05 11:06:50
gitblit/gitblit
https://api.github.com/repos/gitblit/gitblit
closed
Manifest Change in 1.9.0 Requires Explicit Classpath
Catg-Defect Status-Fixed
First of all thank you very much for the development of Gitblit, it's a really nice software :) I was just helping a colleage to update Gitblit GO on a FreeBSD server from 1.8.0 to 1.9.0. We ran into problems regarding a change in `MANIFEST.MF` in `gitblit.jar`. Since the issues are Java-related, they most likely apply for all operating systems. The manifest from 1.8.0 looks like this: ``` Manifest-Version: 1.0 Ant-Version: Apache Ant 1.9.3 Created-By: Moxie v0.9.4 Build-Jdk: 1.8.0_91 Build-Date: 2016-06-22 Implementation-Title: Gitblit Implementation-Vendor-Id: com.gitblit Implementation-Vendor-URL: http://gitblit.com Implementation-Version: 1.8.0 Bundle-Name: Gitblit Bundle-SymbolicName: gitblit Bundle-Version: 1.8.0 Maven-Url: http://gitblit.github.io/gitblit-maven Main-Class: org.moxie.MxLauncher mxMain-Class: com.gitblit.GitBlitServer mxMain-Paths: ext ``` and the manifest from 1.9.0 looks like this: ``` Manifest-Version: 1.0 Ant-Version: Apache Ant 1.9.2 Created-By: Moxie v0.9.4 Build-Jdk: 1.7.0_252 Build-Date: 2020-02-01 Implementation-Title: Gitblit Implementation-Vendor-Id: com.gitblit Implementation-Vendor-URL: http://gitblit.com Implementation-Version: 1.9.0 Bundle-Name: Gitblit Bundle-SymbolicName: gitblit Bundle-Version: 1.9.0 Maven-Url: http://gitblit.github.io/gitblit-maven Main-Class: com.gitblit.GitBlitServer ``` The most notable changes are: - Main class changed from `org.moxie.MxLauncher` to `com.gitblit.GitblitServer` - `mxMain-Class: com.gitblit.GitBlitServer` and `mxMain-Paths: ext` was removed When using 1.8.0, it was possible to start gitblit using a command line like this: ``` java -jar gitblit.jar --baseFolder data ``` This was possible because the Moxie launcher automatically added the jars in the `ext` directory to the classpath. Now that this was removed, we end up with `NoClassDefFoundErrors`. To resolve this, we have to specify the classpath explicitly **and** replace `-jar gitblit.jar` with the main class: ``` java -cp "gitblit.jar:ext/*" com.gitblit.GitBlitServer --baseFolder data ``` Note: you **have to add the quotes around the classpath**, otherwise it does not work. Is this behavior intended? If yes, then several parts of the documentation should be updated. Also, the service scripts do not work, because they use the `-jar` version of the command. If the behavior is not intended, `ext` should be added to the classpath of gitblit.jar in the manifest.
1.0
Manifest Change in 1.9.0 Requires Explicit Classpath - First of all thank you very much for the development of Gitblit, it's a really nice software :) I was just helping a colleage to update Gitblit GO on a FreeBSD server from 1.8.0 to 1.9.0. We ran into problems regarding a change in `MANIFEST.MF` in `gitblit.jar`. Since the issues are Java-related, they most likely apply for all operating systems. The manifest from 1.8.0 looks like this: ``` Manifest-Version: 1.0 Ant-Version: Apache Ant 1.9.3 Created-By: Moxie v0.9.4 Build-Jdk: 1.8.0_91 Build-Date: 2016-06-22 Implementation-Title: Gitblit Implementation-Vendor-Id: com.gitblit Implementation-Vendor-URL: http://gitblit.com Implementation-Version: 1.8.0 Bundle-Name: Gitblit Bundle-SymbolicName: gitblit Bundle-Version: 1.8.0 Maven-Url: http://gitblit.github.io/gitblit-maven Main-Class: org.moxie.MxLauncher mxMain-Class: com.gitblit.GitBlitServer mxMain-Paths: ext ``` and the manifest from 1.9.0 looks like this: ``` Manifest-Version: 1.0 Ant-Version: Apache Ant 1.9.2 Created-By: Moxie v0.9.4 Build-Jdk: 1.7.0_252 Build-Date: 2020-02-01 Implementation-Title: Gitblit Implementation-Vendor-Id: com.gitblit Implementation-Vendor-URL: http://gitblit.com Implementation-Version: 1.9.0 Bundle-Name: Gitblit Bundle-SymbolicName: gitblit Bundle-Version: 1.9.0 Maven-Url: http://gitblit.github.io/gitblit-maven Main-Class: com.gitblit.GitBlitServer ``` The most notable changes are: - Main class changed from `org.moxie.MxLauncher` to `com.gitblit.GitblitServer` - `mxMain-Class: com.gitblit.GitBlitServer` and `mxMain-Paths: ext` was removed When using 1.8.0, it was possible to start gitblit using a command line like this: ``` java -jar gitblit.jar --baseFolder data ``` This was possible because the Moxie launcher automatically added the jars in the `ext` directory to the classpath. Now that this was removed, we end up with `NoClassDefFoundErrors`. To resolve this, we have to specify the classpath explicitly **and** replace `-jar gitblit.jar` with the main class: ``` java -cp "gitblit.jar:ext/*" com.gitblit.GitBlitServer --baseFolder data ``` Note: you **have to add the quotes around the classpath**, otherwise it does not work. Is this behavior intended? If yes, then several parts of the documentation should be updated. Also, the service scripts do not work, because they use the `-jar` version of the command. If the behavior is not intended, `ext` should be added to the classpath of gitblit.jar in the manifest.
defect
manifest change in requires explicit classpath first of all thank you very much for the development of gitblit it s a really nice software i was just helping a colleage to update gitblit go on a freebsd server from to we ran into problems regarding a change in manifest mf in gitblit jar since the issues are java related they most likely apply for all operating systems the manifest from looks like this manifest version ant version apache ant created by moxie build jdk build date implementation title gitblit implementation vendor id com gitblit implementation vendor url implementation version bundle name gitblit bundle symbolicname gitblit bundle version maven url main class org moxie mxlauncher mxmain class com gitblit gitblitserver mxmain paths ext and the manifest from looks like this manifest version ant version apache ant created by moxie build jdk build date implementation title gitblit implementation vendor id com gitblit implementation vendor url implementation version bundle name gitblit bundle symbolicname gitblit bundle version maven url main class com gitblit gitblitserver the most notable changes are main class changed from org moxie mxlauncher to com gitblit gitblitserver mxmain class com gitblit gitblitserver and mxmain paths ext was removed when using it was possible to start gitblit using a command line like this java jar gitblit jar basefolder data this was possible because the moxie launcher automatically added the jars in the ext directory to the classpath now that this was removed we end up with noclassdeffounderrors to resolve this we have to specify the classpath explicitly and replace jar gitblit jar with the main class java cp gitblit jar ext com gitblit gitblitserver basefolder data note you have to add the quotes around the classpath otherwise it does not work is this behavior intended if yes then several parts of the documentation should be updated also the service scripts do not work because they use the jar version of the command if the behavior is not intended ext should be added to the classpath of gitblit jar in the manifest
1
62,704
17,155,327,467
IssuesEvent
2021-07-14 05:50:38
martinrotter/rssguard
https://api.github.com/repos/martinrotter/rssguard
closed
[BUG]: inoreader login failed when used with VPN v2ray
Type-Defect
**Brief description of the issue.** When I try to log in inoreader(and feedly), I always fail to log in. **How to reproduce the bug?** Try to use Inoreader and test the login. Network: I am in China, but I have used v2rayN to hook up the proxy, and I can log in to the inoreader webpage. Login process: I have applied for app id and app key and re(picture1), and when I click login in rssguard, the webpage jumps to the inoreader authorization page (picture 2), after clicking the authorization, it will jump to the prompt to close the window page (picture 3). But the software crystal surface prompts an unknown error(pic4). **What was the expected result?** For the login to report success. **What actually happened?** The login fails with the following message: There is error. unknown error Hope to reply,thanks a lot! * OS: win10 20H2 and macos big sur * RSS Guard version:3.92 and 4.0 <img width="122" alt="pic1" src="https://user-images.githubusercontent.com/54013763/123372407-12234e80-d5b6-11eb-8ad8-7684bba09a3e.png"> <img width="699" alt="pic2" src="https://user-images.githubusercontent.com/54013763/123372766-c1602580-d5b6-11eb-81a6-c831c8807d91.png"> <img width="518" alt="pic3" src="https://user-images.githubusercontent.com/54013763/123372430-1d767a00-d5b6-11eb-8470-8069e50549cd.png"> <img width="122" alt="pic4" src="https://user-images.githubusercontent.com/54013763/123372435-1f403d80-d5b6-11eb-8dfb-75785fe7e5d4.png">
1.0
[BUG]: inoreader login failed when used with VPN v2ray - **Brief description of the issue.** When I try to log in inoreader(and feedly), I always fail to log in. **How to reproduce the bug?** Try to use Inoreader and test the login. Network: I am in China, but I have used v2rayN to hook up the proxy, and I can log in to the inoreader webpage. Login process: I have applied for app id and app key and re(picture1), and when I click login in rssguard, the webpage jumps to the inoreader authorization page (picture 2), after clicking the authorization, it will jump to the prompt to close the window page (picture 3). But the software crystal surface prompts an unknown error(pic4). **What was the expected result?** For the login to report success. **What actually happened?** The login fails with the following message: There is error. unknown error Hope to reply,thanks a lot! * OS: win10 20H2 and macos big sur * RSS Guard version:3.92 and 4.0 <img width="122" alt="pic1" src="https://user-images.githubusercontent.com/54013763/123372407-12234e80-d5b6-11eb-8ad8-7684bba09a3e.png"> <img width="699" alt="pic2" src="https://user-images.githubusercontent.com/54013763/123372766-c1602580-d5b6-11eb-81a6-c831c8807d91.png"> <img width="518" alt="pic3" src="https://user-images.githubusercontent.com/54013763/123372430-1d767a00-d5b6-11eb-8470-8069e50549cd.png"> <img width="122" alt="pic4" src="https://user-images.githubusercontent.com/54013763/123372435-1f403d80-d5b6-11eb-8dfb-75785fe7e5d4.png">
defect
inoreader login failed when used with vpn brief description of the issue when i try to log in inoreader(and feedly) i always fail to log in how to reproduce the bug try to use inoreader and test the login network i am in china but i have used to hook up the proxy and i can log in to the inoreader webpage login process i have applied for app id and app key and re( ) and when i click login in rssguard the webpage jumps to the inoreader authorization page picture after clicking the authorization it will jump to the prompt to close the window page picture but the software crystal surface prompts an unknown error what was the expected result for the login to report success what actually happened the login fails with the following message there is error unknown error hope to reply,thanks a lot os and macos big sur rss guard version and img width alt src img width alt src img width alt src img width alt src
1
362,037
10,722,221,516
IssuesEvent
2019-10-27 10:17:29
Jinnaix/nemesis-bug-reporting
https://api.github.com/repos/Jinnaix/nemesis-bug-reporting
closed
[Heroic and Normal Steamvault] Mekgineer Steamrigger [Boss]
C++ Script Creature Dungeon Fixed Medium Priority
You can pull the boss with a ranged weapon and the gnomes around him doesnt trigger combat ------------------- Normal mode missing ability; Gnome Repairers: A few times during the fight at 75%, 50% and 25% health, he will call upon 3-4 gnomes to heal him. These gnomes have 5k hp and deal 200-500 melee damage on hit. These are non-elite mobs and should receive all focus from DPS, to prevent boss from healing. The healing is a channeled spell which break on damage, and after it has been broken the gnome will start it again after a few seconds, so keep the DPS up. ---------- Heroic mode missing 2 abilities; 1- Steamrigger in heroic mode is almost identical to normal mode, except that an extra gnome appears every 7 seconds or so 2- Enrage: Enrages after 5 minutes (heroic mode only) -------- https://wowwiki.fandom.com/wiki/Mekgineer_Steamrigger
1.0
[Heroic and Normal Steamvault] Mekgineer Steamrigger [Boss] - You can pull the boss with a ranged weapon and the gnomes around him doesnt trigger combat ------------------- Normal mode missing ability; Gnome Repairers: A few times during the fight at 75%, 50% and 25% health, he will call upon 3-4 gnomes to heal him. These gnomes have 5k hp and deal 200-500 melee damage on hit. These are non-elite mobs and should receive all focus from DPS, to prevent boss from healing. The healing is a channeled spell which break on damage, and after it has been broken the gnome will start it again after a few seconds, so keep the DPS up. ---------- Heroic mode missing 2 abilities; 1- Steamrigger in heroic mode is almost identical to normal mode, except that an extra gnome appears every 7 seconds or so 2- Enrage: Enrages after 5 minutes (heroic mode only) -------- https://wowwiki.fandom.com/wiki/Mekgineer_Steamrigger
non_defect
mekgineer steamrigger you can pull the boss with a ranged weapon and the gnomes around him doesnt trigger combat normal mode missing ability gnome repairers a few times during the fight at and health he will call upon gnomes to heal him these gnomes have hp and deal melee damage on hit these are non elite mobs and should receive all focus from dps to prevent boss from healing the healing is a channeled spell which break on damage and after it has been broken the gnome will start it again after a few seconds so keep the dps up heroic mode missing abilities steamrigger in heroic mode is almost identical to normal mode except that an extra gnome appears every seconds or so enrage enrages after minutes heroic mode only
0
77,848
27,192,301,272
IssuesEvent
2023-02-19 23:09:09
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Element Call video room PIP (always on top) avatars are not round
T-Defect
### Steps to reproduce 1. join an element call video room 2. another call member has an avatar that is rectangular (height > width), which usually isn't visible as they're cropped to circles 3. click on another room, watch the call become a PIP floaty that's always on top 4. notice the kinder surprise shaped avatar in the PIP (screenshot) ### Outcome #### What did you expect? avatars are always perfect circles #### What happened instead? ![image](https://user-images.githubusercontent.com/2803622/219980557-3ca978ee-8e49-4a1b-8ab1-951d5ad7e2e2.png) ```html <div class="_avatar_w2fgr_17 _avatar_7m54n_80" style="background-color: rgb(45, 194, 197); width: 110px; height: 110px; border-radius: 110px; font-size: 55px;"> <img src="https://matrix.example.org/_matrix/media/r0/thumbnail/example.org/abcdef1234567890?width=110&amp;height=110&amp;method=scale" > </div> ``` ### Operating system arch ### Application version Element Nightly version: 2023021901 Olm version: 3.2.12 ### How did you install the app? aur ### Homeserver synapse 1.77.0" ### Will you send logs? No
1.0
Element Call video room PIP (always on top) avatars are not round - ### Steps to reproduce 1. join an element call video room 2. another call member has an avatar that is rectangular (height > width), which usually isn't visible as they're cropped to circles 3. click on another room, watch the call become a PIP floaty that's always on top 4. notice the kinder surprise shaped avatar in the PIP (screenshot) ### Outcome #### What did you expect? avatars are always perfect circles #### What happened instead? ![image](https://user-images.githubusercontent.com/2803622/219980557-3ca978ee-8e49-4a1b-8ab1-951d5ad7e2e2.png) ```html <div class="_avatar_w2fgr_17 _avatar_7m54n_80" style="background-color: rgb(45, 194, 197); width: 110px; height: 110px; border-radius: 110px; font-size: 55px;"> <img src="https://matrix.example.org/_matrix/media/r0/thumbnail/example.org/abcdef1234567890?width=110&amp;height=110&amp;method=scale" > </div> ``` ### Operating system arch ### Application version Element Nightly version: 2023021901 Olm version: 3.2.12 ### How did you install the app? aur ### Homeserver synapse 1.77.0" ### Will you send logs? No
defect
element call video room pip always on top avatars are not round steps to reproduce join an element call video room another call member has an avatar that is rectangular height width which usually isn t visible as they re cropped to circles click on another room watch the call become a pip floaty that s always on top notice the kinder surprise shaped avatar in the pip screenshot outcome what did you expect avatars are always perfect circles what happened instead html operating system arch application version element nightly version olm version how did you install the app aur homeserver synapse will you send logs no
1
3,668
2,610,066,731
IssuesEvent
2015-02-26 18:19:36
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
路桥治前列腺炎一般需要多少钱
auto-migrated Priority-Medium Type-Defect
``` 路桥治前列腺炎一般需要多少钱【台州五洲生殖医院】24小时 健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地� ��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1 04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、 105、109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:39
1.0
路桥治前列腺炎一般需要多少钱 - ``` 路桥治前列腺炎一般需要多少钱【台州五洲生殖医院】24小时 健康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地� ��:台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐1 04、108、118、198及椒江一金清公交车直达枫南小区,乘坐107、 105、109、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:39
defect
路桥治前列腺炎一般需要多少钱 路桥治前列腺炎一般需要多少钱【台州五洲生殖医院】 健康咨询热线 微信号tzwzszyy 医院地� �� (枫南大转盘旁)乘车线路 、 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
1
75,608
25,943,666,033
IssuesEvent
2022-12-16 21:19:43
bigbluebutton/bigbluebutton
https://api.github.com/repos/bigbluebutton/bigbluebutton
closed
Breakout rooms don't save room assigning
type: defect status: accepted module: client component: breakout rooms
<!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. This issue tracker is only for bbb development related issues.--> **Describe the bug** When creating breakout rooms in a single session, only room names and amount of rooms are saved. Participants have to be assigned to rooms manually every time. **To Reproduce** Steps to reproduce the behavior: 1. Create breakout rooms as normal and assign participants 2. End breakout room session 3. Create new round of breakout rooms 4. Every student is not assigned to room **Expected behavior** Update notes BBB 2.5: "Breakout rooms now remember your previous rooms assignments within the current session. This means if you assign students to breakout rooms and then, later in the class want to reuse the breakout rooms, you don’t need to re-assign students." **Actual behavior** Assignment is not saved. Only amount and names of rooms. **BBB version:** BigBlueButton continually evolves. Providing the version/build helps us to pinpoint when an issue was introduced. BigBlueButton Server 2.5.5 Issue also happens on 2.6 test servers **Desktop (please complete the following information):** - OS: [multiple (MacOS Big Sur, Windows 10)] - Browser [multiple: Chrome, Safari, Edge, Firefox]
1.0
Breakout rooms don't save room assigning - <!--PLEASE DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. This issue tracker is only for bbb development related issues.--> **Describe the bug** When creating breakout rooms in a single session, only room names and amount of rooms are saved. Participants have to be assigned to rooms manually every time. **To Reproduce** Steps to reproduce the behavior: 1. Create breakout rooms as normal and assign participants 2. End breakout room session 3. Create new round of breakout rooms 4. Every student is not assigned to room **Expected behavior** Update notes BBB 2.5: "Breakout rooms now remember your previous rooms assignments within the current session. This means if you assign students to breakout rooms and then, later in the class want to reuse the breakout rooms, you don’t need to re-assign students." **Actual behavior** Assignment is not saved. Only amount and names of rooms. **BBB version:** BigBlueButton continually evolves. Providing the version/build helps us to pinpoint when an issue was introduced. BigBlueButton Server 2.5.5 Issue also happens on 2.6 test servers **Desktop (please complete the following information):** - OS: [multiple (MacOS Big Sur, Windows 10)] - Browser [multiple: Chrome, Safari, Edge, Firefox]
defect
breakout rooms don t save room assigning please do not file issues for general support questions this issue tracker is only for bbb development related issues describe the bug when creating breakout rooms in a single session only room names and amount of rooms are saved participants have to be assigned to rooms manually every time to reproduce steps to reproduce the behavior create breakout rooms as normal and assign participants end breakout room session create new round of breakout rooms every student is not assigned to room expected behavior update notes bbb breakout rooms now remember your previous rooms assignments within the current session this means if you assign students to breakout rooms and then later in the class want to reuse the breakout rooms you don’t need to re assign students actual behavior assignment is not saved only amount and names of rooms bbb version bigbluebutton continually evolves providing the version build helps us to pinpoint when an issue was introduced bigbluebutton server issue also happens on test servers desktop please complete the following information os browser
1
58,215
16,439,562,163
IssuesEvent
2021-05-20 13:01:37
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Test Failing: Duplicate Benefit Hub references should not be allowed
Automated testing Core Application Team Critical defect Drupal engineering Unplanned work
``` Running: entity_reference_validation.feature (2 of 6) Entity Reference Validation 1) Duplicate Benefit Hub references should not be allowed 0 passing (46s) 1) Entity Reference Validation Duplicate Benefit Hub references should not be allowed: AssertionError: Timed out retrying after 10000ms: Expected to find content: '1 error has been found' but never did. at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:174838:28) at Context.resolveAndRunStepDefinition (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23780:9) at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23121:35) ``` _Originally posted by @ElijahLynn in https://github.com/department-of-veterans-affairs/va.gov-cms/issues/5363#issuecomment-842578031_
1.0
Test Failing: Duplicate Benefit Hub references should not be allowed - ``` Running: entity_reference_validation.feature (2 of 6) Entity Reference Validation 1) Duplicate Benefit Hub references should not be allowed 0 passing (46s) 1) Entity Reference Validation Duplicate Benefit Hub references should not be allowed: AssertionError: Timed out retrying after 10000ms: Expected to find content: '1 error has been found' but never did. at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:174838:28) at Context.resolveAndRunStepDefinition (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23780:9) at Context.eval (https://pr5363-wzzeapb2fhjgbyb7dsiyua0xtvokozhg.ci.cms.va.gov/__cypress/tests?p=cypress/integration/entity_reference_validation.feature:23121:35) ``` _Originally posted by @ElijahLynn in https://github.com/department-of-veterans-affairs/va.gov-cms/issues/5363#issuecomment-842578031_
defect
test failing duplicate benefit hub references should not be allowed running entity reference validation feature of entity reference validation duplicate benefit hub references should not be allowed passing entity reference validation duplicate benefit hub references should not be allowed assertionerror timed out retrying after expected to find content error has been found but never did at context eval at context resolveandrunstepdefinition at context eval originally posted by elijahlynn in
1
9,444
2,615,150,382
IssuesEvent
2015-03-01 06:27:40
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Hung on [+] Waiting for beacon
auto-migrated Priority-Triage Type-Defect
``` 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) v1.3 1. What operating system are you using (Linux is the only supported OS)? Windows 7 32bit with VMware 8 running Back Track 5 R1 2. Is your wireless card in monitor mode (yes/no)? Yes under mon0 3. What is the signal strength of the Access Point you are trying to crack? signal strength is -53 4. What is the manufacturer and model # of the device you are trying to crack? Manufacturer SMC Model # D3GN-RRR http://www.smc.com/files/AJ/SMCD3GN-UserManual_Rogers.pdf 5. What is the entire command line string you are supplying to reaver? # reaver -i mon0 -b 78:CD:8E:CD:18:99 -vv 6. Please describe what you think the issue is. I have no idea of what the problem is :( 7. Paste the output from Reaver below. [+] Waiting for beacon from 78:CD:8E:CD:18:99 Additional info Modem/router has a WPS button so i assume is supports it. CTRL + ALT produces the following [+] Nothing done, nothing to save. [+] Session saved. Using Linksys WUSB100 Ralink RT2870/3070 rt2800usb ``` Original issue reported on code.google.com by `gordonJe...@gmail.com` on 18 Jan 2012 at 11:23
1.0
Hung on [+] Waiting for beacon - ``` 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) v1.3 1. What operating system are you using (Linux is the only supported OS)? Windows 7 32bit with VMware 8 running Back Track 5 R1 2. Is your wireless card in monitor mode (yes/no)? Yes under mon0 3. What is the signal strength of the Access Point you are trying to crack? signal strength is -53 4. What is the manufacturer and model # of the device you are trying to crack? Manufacturer SMC Model # D3GN-RRR http://www.smc.com/files/AJ/SMCD3GN-UserManual_Rogers.pdf 5. What is the entire command line string you are supplying to reaver? # reaver -i mon0 -b 78:CD:8E:CD:18:99 -vv 6. Please describe what you think the issue is. I have no idea of what the problem is :( 7. Paste the output from Reaver below. [+] Waiting for beacon from 78:CD:8E:CD:18:99 Additional info Modem/router has a WPS button so i assume is supports it. CTRL + ALT produces the following [+] Nothing done, nothing to save. [+] Session saved. Using Linksys WUSB100 Ralink RT2870/3070 rt2800usb ``` Original issue reported on code.google.com by `gordonJe...@gmail.com` on 18 Jan 2012 at 11:23
defect
hung on waiting for beacon what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os windows with vmware running back track is your wireless card in monitor mode yes no yes under what is the signal strength of the access point you are trying to crack signal strength is what is the manufacturer and model of the device you are trying to crack manufacturer smc model rrr what is the entire command line string you are supplying to reaver reaver i b cd cd vv please describe what you think the issue is i have no idea of what the problem is paste the output from reaver below waiting for beacon from cd cd additional info modem router has a wps button so i assume is supports it ctrl alt produces the following nothing done nothing to save session saved using linksys ralink original issue reported on code google com by gordonje gmail com on jan at
1
110,331
11,698,355,243
IssuesEvent
2020-03-06 13:44:32
MeteoSwiss-APN/dawn
https://api.github.com/repos/MeteoSwiss-APN/dawn
closed
Clean up Atlas Utils Repo
Testing documentation enhancement infrastructure
The utilities in https://github.com/mroethlin/AtlasUtilities are poorly documented, there is no build instructions and testing is done manually. Address these issues. Properly document the utilities, provide a basic CMake, make a test script.
1.0
Clean up Atlas Utils Repo - The utilities in https://github.com/mroethlin/AtlasUtilities are poorly documented, there is no build instructions and testing is done manually. Address these issues. Properly document the utilities, provide a basic CMake, make a test script.
non_defect
clean up atlas utils repo the utilities in are poorly documented there is no build instructions and testing is done manually address these issues properly document the utilities provide a basic cmake make a test script
0
583,506
17,390,573,199
IssuesEvent
2021-08-02 06:43:18
codidact/qpixel
https://api.github.com/repos/codidact/qpixel
closed
I can see a removed thread in Linux Systems
area: html/css/js area: ruby complexity: unassessed priority: medium type: bug
I can see a removed thread in Linux Systems. https://linux.codidact.com/comments/thread/3720 ![Screenshot from 2021-07-30 01-57-21](https://user-images.githubusercontent.com/58106197/127607671-eaaf82e2-678d-4d64-998b-249b7df5ab2b.png) The thread were created by Elgonzo. And, I guess, it was deleted by a moderator.
1.0
I can see a removed thread in Linux Systems - I can see a removed thread in Linux Systems. https://linux.codidact.com/comments/thread/3720 ![Screenshot from 2021-07-30 01-57-21](https://user-images.githubusercontent.com/58106197/127607671-eaaf82e2-678d-4d64-998b-249b7df5ab2b.png) The thread were created by Elgonzo. And, I guess, it was deleted by a moderator.
non_defect
i can see a removed thread in linux systems i can see a removed thread in linux systems the thread were created by elgonzo and i guess it was deleted by a moderator
0
563,968
16,706,707,153
IssuesEvent
2021-06-09 10:51:20
googleapis/google-api-ruby-client
https://api.github.com/repos/googleapis/google-api-ruby-client
closed
Synthesis failed for language-v1beta1
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate language-v1beta1. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the language-v1beta1 API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:17:53,723 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:17:54,515 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:17:54,518 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:17:54,520 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:17:54,522 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:17:54,524 autosynth [DEBUG] > Running: git branch -f autosynth-language-v1beta1 2021-06-08 03:17:54,527 autosynth [DEBUG] > Running: git checkout autosynth-language-v1beta1 Switched to branch 'autosynth-language-v1beta1' 2021-06-08 03:17:54,726 autosynth [INFO] > Running synthtool 2021-06-08 03:17:54,726 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--'] 2021-06-08 03:17:54,726 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/language/v1beta1/sponge_log.log 2021-06-08 03:17:54,728 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-language_v1beta1/synth.metadata synth.py -- language v1beta1 2021-06-08 03:17:54,919 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-language-v1beta1 nothing to commit, working tree clean 2021-06-08 03:17:54,977 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1 DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1 git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching source index from https://rubygems.org/ Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> chown -R 1000:1000 /workspace/generated 2021-06-08 03:17:56,329 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'language', 'v1beta1']' returned non-zero exit status 1. 2021-06-08 03:17:56,355 autosynth [ERROR] > Synthesis failed 2021-06-08 03:17:56,355 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--', 'language', 'v1beta1']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
1.0
Synthesis failed for language-v1beta1 - Hello! Autosynth couldn't regenerate language-v1beta1. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the language-v1beta1 API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` 2021-06-08 03:17:53,723 autosynth [INFO] > logs will be written to: /tmpfs/src/logs/google-api-ruby-client 2021-06-08 03:17:54,515 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2021-06-08 03:17:54,518 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2021-06-08 03:17:54,520 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2021-06-08 03:17:54,522 autosynth [DEBUG] > Running: git config push.default simple 2021-06-08 03:17:54,524 autosynth [DEBUG] > Running: git branch -f autosynth-language-v1beta1 2021-06-08 03:17:54,527 autosynth [DEBUG] > Running: git checkout autosynth-language-v1beta1 Switched to branch 'autosynth-language-v1beta1' 2021-06-08 03:17:54,726 autosynth [INFO] > Running synthtool 2021-06-08 03:17:54,726 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--'] 2021-06-08 03:17:54,726 autosynth [DEBUG] > log_file_path: /tmpfs/src/logs/google-api-ruby-client/language/v1beta1/sponge_log.log 2021-06-08 03:17:54,728 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata generated/google-apis-language_v1beta1/synth.metadata synth.py -- language v1beta1 2021-06-08 03:17:54,919 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py. On branch autosynth-language-v1beta1 nothing to commit, working tree clean 2021-06-08 03:17:54,977 synthtool [DEBUG] > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1 DEBUG:synthtool:Running: docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1 git clean -df bundle install Don't run Bundler as root. Bundler can ask for sudo if it is needed, and installing your bundle as root will break this application for all non-root users on this machine. Fetching source index from https://rubygems.org/ Retrying fetcher due to error (2/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (3/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Retrying fetcher due to error (4/4): Bundler::HTTPError Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> Could not fetch specs from https://rubygems.org/ due to underlying error <bad response Service Unavailable 503 (https://rubygems.org/specs.4.8.gz)> chown -R 1000:1000 /workspace/generated 2021-06-08 03:17:56,329 synthtool [ERROR] > Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1: None ERROR:synthtool:Failed executing docker run --rm -v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace -v/var/run/docker.sock:/var/run/docker.sock -w /workspace -e USER_GROUP=1000:1000 --entrypoint script/synth.rb gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth language v1beta1: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1137, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1062, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1404, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 763, in invoke return __callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/google-api-ruby-client/synth.py", line 41, in <module> shell.run(command, hide_output=False) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['docker', 'run', '--rm', '-v/home/kbuilder/.cache/synthtool/google-api-ruby-client:/workspace', '-v/var/run/docker.sock:/var/run/docker.sock', '-w', '/workspace', '-e', 'USER_GROUP=1000:1000', '--entrypoint', 'script/synth.rb', 'gcr.io/cloud-devrel-kokoro-resources/yoshi-ruby/autosynth', 'language', 'v1beta1']' returned non-zero exit status 1. 2021-06-08 03:17:56,355 autosynth [ERROR] > Synthesis failed 2021-06-08 03:17:56,355 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 293, in _inner_main ).synthesize(synth_log_path / "sponge_log.log") File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'generated/google-apis-language_v1beta1/synth.metadata', 'synth.py', '--', 'language', 'v1beta1']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/bdd2ce63-997e-4095-81b3-cd54e5fccbe4/targets/github%2Fsynthtool;config=default/tests;query=google-api-ruby-client;failed=false).
non_defect
synthesis failed for language hello autosynth couldn t regenerate language broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the language api and the library grows stale see for trouble shooting tips here s the output from running synth py autosynth logs will be written to tmpfs src logs google api ruby client autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth language autosynth running git checkout autosynth language switched to branch autosynth language autosynth running synthtool autosynth autosynth log file path tmpfs src logs google api ruby client language sponge log log autosynth running tmpfs src github synthtool env bin m synthtool metadata generated google apis language synth metadata synth py language synthtool executing home kbuilder cache synthtool google api ruby client synth py on branch autosynth language nothing to commit working tree clean synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language debug synthtool running docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language git clean df bundle install don t run bundler as root bundler can ask for sudo if it is needed and installing your bundle as root will break this application for all non root users on this machine fetching source index from retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable retrying fetcher due to error bundler httperror could not fetch specs from due to underlying error bad response service unavailable could not fetch specs from due to underlying error bad response service unavailable chown r workspace generated synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language none error synthtool failed executing docker run rm v home kbuilder cache synthtool google api ruby client workspace v var run docker sock var run docker sock w workspace e user group entrypoint script synth rb gcr io cloud devrel kokoro resources yoshi ruby autosynth language none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool google api ruby client synth py line in shell run command hide output false file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize synth log path sponge log log file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
35,651
7,795,087,015
IssuesEvent
2018-06-08 06:43:32
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
CLUBB can't read GrADS files with a 5 minute timestep (Trac #119)
Migrated from Trac clubb_src defect fasching@uwm.edu
Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce: Edit ../input/case_setups/fire_model.in to use stats_tout = 300. $ ./run_scm.bash fire Edit ../input/case_setup/fire_model.in to use fname_prefix= "fire_tmp" Edit ../input_misc/inputfields/fire_inputfields to use datafile = "../output/fire" $ ./run_inputfields.bash fire Output: At line 390 of file ../src/inputgrads.F90 (Unit 15 "../output/fire_zt.dat") Traceback: (Innermost first) Called from line 246 of file ../src/inputfields.F90 Called from line 707 of file ../src/clubb_driver.F90 Called from line 170 of file ../src/clubb_inputfields.F90 Fortran runtime error: Record number must be positive I suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/119 ```json { "status": "closed", "changetime": "2009-07-22T20:30:12", "description": "Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce:\n\nEdit ../input/case_setups/fire_model.in to use stats_tout = 300.\n$ ./run_scm.bash fire\nEdit ../input/case_setup/fire_model.in to use fname_prefix= \"fire_tmp\"\nEdit ../input_misc/inputfields/fire_inputfields to use datafile = \"../output/fire\"\n$ ./run_inputfields.bash fire\n\nOutput:\nAt line 390 of file ../src/inputgrads.F90 (Unit 15 \"../output/fire_zt.dat\")\nTraceback: (Innermost first)\n Called from line 246 of file ../src/inputfields.F90\n Called from line 707 of file ../src/clubb_driver.F90\n Called from line 170 of file ../src/clubb_inputfields.F90\nFortran runtime error: Record number must be positive\nI suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug.", "reporter": "dschanen@uwm.edu", "cc": "senkbeil@uwm.edu, fasching@uwm.edu", "resolution": "Verified by V. Larson", "_ts": "1248294612000000", "component": "clubb_src", "summary": "CLUBB can't read GrADS files with a 5 minute timestep", "priority": "major", "keywords": "GrADS, inputfields", "time": "2009-07-09T20:22:15", "milestone": "", "owner": "fasching@uwm.edu", "type": "defect" } ```
1.0
CLUBB can't read GrADS files with a 5 minute timestep (Trac #119) - Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce: Edit ../input/case_setups/fire_model.in to use stats_tout = 300. $ ./run_scm.bash fire Edit ../input/case_setup/fire_model.in to use fname_prefix= "fire_tmp" Edit ../input_misc/inputfields/fire_inputfields to use datafile = "../output/fire" $ ./run_inputfields.bash fire Output: At line 390 of file ../src/inputgrads.F90 (Unit 15 "../output/fire_zt.dat") Traceback: (Innermost first) Called from line 246 of file ../src/inputfields.F90 Called from line 707 of file ../src/clubb_driver.F90 Called from line 170 of file ../src/clubb_inputfields.F90 Fortran runtime error: Record number must be positive I suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/119 ```json { "status": "closed", "changetime": "2009-07-22T20:30:12", "description": "Ryan discovered when trying to do a run_inputfields simulation that our inputgrads code for reading a variable does not work for GrADS data output every 300 seconds. To reproduce:\n\nEdit ../input/case_setups/fire_model.in to use stats_tout = 300.\n$ ./run_scm.bash fire\nEdit ../input/case_setup/fire_model.in to use fname_prefix= \"fire_tmp\"\nEdit ../input_misc/inputfields/fire_inputfields to use datafile = \"../output/fire\"\n$ ./run_inputfields.bash fire\n\nOutput:\nAt line 390 of file ../src/inputgrads.F90 (Unit 15 \"../output/fire_zt.dat\")\nTraceback: (Innermost first)\n Called from line 246 of file ../src/inputfields.F90\n Called from line 707 of file ../src/clubb_driver.F90\n Called from line 170 of file ../src/clubb_inputfields.F90\nFortran runtime error: Record number must be positive\nI suspect the formula [http://carson.math.uwm.edu/trac/clubb/browser/trunk/src/inputgrads.F90#L380 here] might have a bug.", "reporter": "dschanen@uwm.edu", "cc": "senkbeil@uwm.edu, fasching@uwm.edu", "resolution": "Verified by V. Larson", "_ts": "1248294612000000", "component": "clubb_src", "summary": "CLUBB can't read GrADS files with a 5 minute timestep", "priority": "major", "keywords": "GrADS, inputfields", "time": "2009-07-09T20:22:15", "milestone": "", "owner": "fasching@uwm.edu", "type": "defect" } ```
defect
clubb can t read grads files with a minute timestep trac ryan discovered when trying to do a run inputfields simulation that our inputgrads code for reading a variable does not work for grads data output every seconds to reproduce edit input case setups fire model in to use stats tout run scm bash fire edit input case setup fire model in to use fname prefix fire tmp edit input misc inputfields fire inputfields to use datafile output fire run inputfields bash fire output at line of file src inputgrads unit output fire zt dat traceback innermost first called from line of file src inputfields called from line of file src clubb driver called from line of file src clubb inputfields fortran runtime error record number must be positive i suspect the formula might have a bug attachments migrated from json status closed changetime description ryan discovered when trying to do a run inputfields simulation that our inputgrads code for reading a variable does not work for grads data output every seconds to reproduce n nedit input case setups fire model in to use stats tout n run scm bash fire nedit input case setup fire model in to use fname prefix fire tmp nedit input misc inputfields fire inputfields to use datafile output fire n run inputfields bash fire n noutput nat line of file src inputgrads unit output fire zt dat ntraceback innermost first n called from line of file src inputfields n called from line of file src clubb driver n called from line of file src clubb inputfields nfortran runtime error record number must be positive ni suspect the formula might have a bug reporter dschanen uwm edu cc senkbeil uwm edu fasching uwm edu resolution verified by v larson ts component clubb src summary clubb can t read grads files with a minute timestep priority major keywords grads inputfields time milestone owner fasching uwm edu type defect
1
21,304
3,487,385,789
IssuesEvent
2016-01-01 21:11:11
jOOQ/jOOL
https://api.github.com/repos/jOOQ/jOOL
closed
NoSuchElementException is thrown on Seq.cycle() for empty Seq
P: Medium R: Fixed T: Defect
The following code: ```java Seq.empty().cycle().toList(); ``` throws: ``` java.util.NoSuchElementException at java.util.ArrayList$Itr.next(ArrayList.java:854) at org.jooq.lambda.Seq$1Cycle.next(Seq.java:1910) at java.util.Iterator.forEachRemaining(Iterator.java:116) at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) at org.jooq.lambda.SeqImpl.collect(SeqImpl.java:176) at org.jooq.lambda.Seq.toList(Seq.java:4634) at org.jooq.lambda.Seq.toList(Seq.java:1237) at org.jooq.lambda.SeqTest.testCycle(SeqTest.java:1133) ```
1.0
NoSuchElementException is thrown on Seq.cycle() for empty Seq - The following code: ```java Seq.empty().cycle().toList(); ``` throws: ``` java.util.NoSuchElementException at java.util.ArrayList$Itr.next(ArrayList.java:854) at org.jooq.lambda.Seq$1Cycle.next(Seq.java:1910) at java.util.Iterator.forEachRemaining(Iterator.java:116) at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) at org.jooq.lambda.SeqImpl.collect(SeqImpl.java:176) at org.jooq.lambda.Seq.toList(Seq.java:4634) at org.jooq.lambda.Seq.toList(Seq.java:1237) at org.jooq.lambda.SeqTest.testCycle(SeqTest.java:1133) ```
defect
nosuchelementexception is thrown on seq cycle for empty seq the following code java seq empty cycle tolist throws java util nosuchelementexception at java util arraylist itr next arraylist java at org jooq lambda seq next seq java at java util iterator foreachremaining iterator java at java util spliterators iteratorspliterator foreachremaining spliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at org jooq lambda seqimpl collect seqimpl java at org jooq lambda seq tolist seq java at org jooq lambda seq tolist seq java at org jooq lambda seqtest testcycle seqtest java
1
259,256
22,423,011,530
IssuesEvent
2022-06-20 06:26:02
opencurve/curve
https://api.github.com/repos/opencurve/curve
closed
after restart curvefs mds, the partition num become 0 in metaserver
bug need test
**Describe the bug (描述bug)** 1. curvefs create one fs, then metric shows metaserver has partition num. > topology_metric_metaserver_id_1_partition_num 3 > topology_metric_metaserver_id_2_partition_num 3 > topology_metric_metaserver_id_3_partition_num 3 > topology_metric_pool_default_pool_partition_num 3 2. curvefs restart mds, then metric shows metaserver has no partition num > topology_metric_metaserver_id_1_partition_num 0 > topology_metric_metaserver_id_2_partition_num 0 > topology_metric_metaserver_id_3_partition_num 0 > topology_metric_pool_default_pool_partition_num 3 **To Reproduce (复现方法)** Mount 1 fs on curvefs, restart mds. **Expected behavior (期望行为)** **Versions (各种版本)** OS: Compiler: branch: master commit id: 941e593a4e8818266cc695cd20fe2956e5807910 **Additional context/screenshots (更多上下文/截图)**
1.0
after restart curvefs mds, the partition num become 0 in metaserver - **Describe the bug (描述bug)** 1. curvefs create one fs, then metric shows metaserver has partition num. > topology_metric_metaserver_id_1_partition_num 3 > topology_metric_metaserver_id_2_partition_num 3 > topology_metric_metaserver_id_3_partition_num 3 > topology_metric_pool_default_pool_partition_num 3 2. curvefs restart mds, then metric shows metaserver has no partition num > topology_metric_metaserver_id_1_partition_num 0 > topology_metric_metaserver_id_2_partition_num 0 > topology_metric_metaserver_id_3_partition_num 0 > topology_metric_pool_default_pool_partition_num 3 **To Reproduce (复现方法)** Mount 1 fs on curvefs, restart mds. **Expected behavior (期望行为)** **Versions (各种版本)** OS: Compiler: branch: master commit id: 941e593a4e8818266cc695cd20fe2956e5807910 **Additional context/screenshots (更多上下文/截图)**
non_defect
after restart curvefs mds the partition num become in metaserver describe the bug 描述bug curvefs create one fs then metric shows metaserver has partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric pool default pool partition num curvefs restart mds then metric shows metaserver has no partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric metaserver id partition num topology metric pool default pool partition num to reproduce 复现方法 mount fs on curvefs restart mds expected behavior 期望行为 versions 各种版本 os compiler branch master commit id additional context screenshots 更多上下文 截图
0
55,332
6,470,512,135
IssuesEvent
2017-08-17 09:13:09
elastic/logstash
https://api.github.com/repos/elastic/logstash
opened
Module Tests are Broken in Part due to Removing Max Bucket Size Setting
test failure
Obvious issue :) the fix for #8010 broke two tests it seems: ```sh 1) LogStash::Modules::Scaffold kibana operation provides a list of importable files Failure/Error: expect(resource1.content.size).to eq(2) expected: 2 got: 1 (compared using ==) # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' 2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}) expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}} got: {"changes"=>{"defaultIndex"=>"tester-*"}} (compared using ==) Diff: @@ -1,2 +1,2 @@ -"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}, +"changes" => {"defaultIndex"=>"tester-*"}, # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' Finished in 6 minutes 54 seconds (files took 21.47 seconds to load) 2462 examples, 2 failures, 3 pending 1) LogStash::Modules::Scaffold kibana operation provides a list of importable files Failure/Error: expect(resource1.content.size).to eq(2) expected: 2 got: 1 (compared using ==) # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' 2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}) expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}} got: {"changes"=>{"defaultIndex"=>"tester-*"}} (compared using ==) Diff: @@ -1,2 +1,2 @@ -"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}, +"changes" => {"defaultIndex"=>"tester-*"}, # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' Finished in 6 minutes 54 seconds (files took 21.47 seconds to load) 2462 examples, 2 failures, 3 pending ```
1.0
Module Tests are Broken in Part due to Removing Max Bucket Size Setting - Obvious issue :) the fix for #8010 broke two tests it seems: ```sh 1) LogStash::Modules::Scaffold kibana operation provides a list of importable files Failure/Error: expect(resource1.content.size).to eq(2) expected: 2 got: 1 (compared using ==) # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' 2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}) expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}} got: {"changes"=>{"defaultIndex"=>"tester-*"}} (compared using ==) Diff: @@ -1,2 +1,2 @@ -"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}, +"changes" => {"defaultIndex"=>"tester-*"}, # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' Finished in 6 minutes 54 seconds (files took 21.47 seconds to load) 2462 examples, 2 failures, 3 pending 1) LogStash::Modules::Scaffold kibana operation provides a list of importable files Failure/Error: expect(resource1.content.size).to eq(2) expected: 2 got: 1 (compared using ==) # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:127:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' 2) LogStash::Modules::Scaffold importing to elasticsearch stubbed client calls the import method Failure/Error: expect(contents[0]).to eq({"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}}) expected: {"changes"=>{"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}} got: {"changes"=>{"defaultIndex"=>"tester-*"}} (compared using ==) Diff: @@ -1,2 +1,2 @@ -"changes" => {"defaultIndex"=>"tester-*", "metrics:max_buckets"=>"86400"}, +"changes" => {"defaultIndex"=>"tester-*"}, # ./logstash-core/spec/logstash/modules/scaffold_spec.rb:210:in `block in (root)' # ./vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb:46:in `block in /Users/brownbear/src/logstash/vendor/bundle/jruby/2.3.0/gems/rspec-wait-0.0.9/lib/rspec/wait.rb' # ./lib/bootstrap/rspec.rb:13:in `<main>' Finished in 6 minutes 54 seconds (files took 21.47 seconds to load) 2462 examples, 2 failures, 3 pending ```
non_defect
module tests are broken in part due to removing max bucket size setting obvious issue the fix for broke two tests it seems sh logstash modules scaffold kibana operation provides a list of importable files failure error expect content size to eq expected got compared using logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in logstash modules scaffold importing to elasticsearch stubbed client calls the import method failure error expect contents to eq changes defaultindex tester metrics max buckets expected changes defaultindex tester metrics max buckets got changes defaultindex tester compared using diff changes defaultindex tester metrics max buckets changes defaultindex tester logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in finished in minutes seconds files took seconds to load examples failures pending logstash modules scaffold kibana operation provides a list of importable files failure error expect content size to eq expected got compared using logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in logstash modules scaffold importing to elasticsearch stubbed client calls the import method failure error expect contents to eq changes defaultindex tester metrics max buckets expected changes defaultindex tester metrics max buckets got changes defaultindex tester compared using diff changes defaultindex tester metrics max buckets changes defaultindex tester logstash core spec logstash modules scaffold spec rb in block in root vendor bundle jruby gems rspec wait lib rspec wait rb in block in users brownbear src logstash vendor bundle jruby gems rspec wait lib rspec wait rb lib bootstrap rspec rb in finished in minutes seconds files took seconds to load examples failures pending
0
44,092
5,732,291,974
IssuesEvent
2017-04-21 14:32:02
gudell/bwardp
https://api.github.com/repos/gudell/bwardp
opened
Update HTML on Website for New Pricing
design open development
Implement changes to website to reflect new pricing and the features that would go under each subscription type
1.0
Update HTML on Website for New Pricing - Implement changes to website to reflect new pricing and the features that would go under each subscription type
non_defect
update html on website for new pricing implement changes to website to reflect new pricing and the features that would go under each subscription type
0
30,202
6,041,037,933
IssuesEvent
2017-06-10 20:05:32
Advanced-Post-List/advanced-post-list
https://api.github.com/repos/Advanced-Post-List/advanced-post-list
closed
[Show Preview] Not Correctly Working On Admin Side
P3 - Major T-Defect X-Fixed
The show preview checkbox next to the save button does not appear to do anything. ![show_preview_issue_and_details](https://f.cloud.github.com/assets/2794445/1498123/11a6372e-4824-11e3-9dd8-e955e76cb77d.PNG)
1.0
[Show Preview] Not Correctly Working On Admin Side - The show preview checkbox next to the save button does not appear to do anything. ![show_preview_issue_and_details](https://f.cloud.github.com/assets/2794445/1498123/11a6372e-4824-11e3-9dd8-e955e76cb77d.PNG)
defect
not correctly working on admin side the show preview checkbox next to the save button does not appear to do anything
1
29,617
11,763,768,999
IssuesEvent
2020-03-14 09:04:00
uniquelyparticular/shipengine-request
https://api.github.com/repos/uniquelyparticular/shipengine-request
opened
WS-2020-0042 (High) detected in acorn-5.7.3.tgz, acorn-6.1.1.tgz
security vulnerability
## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>acorn-5.7.3.tgz</b>, <b>acorn-6.1.1.tgz</b></p></summary> <p> <details><summary><b>acorn-5.7.3.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - jest-24.8.0.tgz (Root Library) - jest-cli-24.8.0.tgz - jest-config-24.8.0.tgz - jest-environment-jsdom-24.8.0.tgz - jsdom-11.12.0.tgz - :x: **acorn-5.7.3.tgz** (Vulnerable Library) </details> <details><summary><b>acorn-6.1.1.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz">https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn-globals/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - jest-24.8.0.tgz (Root Library) - jest-cli-24.8.0.tgz - jest-config-24.8.0.tgz - jest-environment-jsdom-24.8.0.tgz - jsdom-11.12.0.tgz - acorn-globals-4.3.2.tgz - :x: **acorn-6.1.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/shipengine-request/commit/f53abdebfdb119b33faab5e4131106b799eb636c">f53abdebfdb119b33faab5e4131106b799eb636c</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-08 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-08</p> <p>Fix Resolution: 7.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0042 (High) detected in acorn-5.7.3.tgz, acorn-6.1.1.tgz - ## WS-2020-0042 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>acorn-5.7.3.tgz</b>, <b>acorn-6.1.1.tgz</b></p></summary> <p> <details><summary><b>acorn-5.7.3.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz">https://registry.npmjs.org/acorn/-/acorn-5.7.3.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - jest-24.8.0.tgz (Root Library) - jest-cli-24.8.0.tgz - jest-config-24.8.0.tgz - jest-environment-jsdom-24.8.0.tgz - jsdom-11.12.0.tgz - :x: **acorn-5.7.3.tgz** (Vulnerable Library) </details> <details><summary><b>acorn-6.1.1.tgz</b></p></summary> <p>ECMAScript parser</p> <p>Library home page: <a href="https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz">https://registry.npmjs.org/acorn/-/acorn-6.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/shipengine-request/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/shipengine-request/node_modules/acorn-globals/node_modules/acorn/package.json</p> <p> Dependency Hierarchy: - jest-24.8.0.tgz (Root Library) - jest-cli-24.8.0.tgz - jest-config-24.8.0.tgz - jest-environment-jsdom-24.8.0.tgz - jsdom-11.12.0.tgz - acorn-globals-4.3.2.tgz - :x: **acorn-6.1.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/shipengine-request/commit/f53abdebfdb119b33faab5e4131106b799eb636c">f53abdebfdb119b33faab5e4131106b799eb636c</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> acorn is vulnerable to REGEX DoS. A regex of the form /[x-\ud800]/u causes the parser to enter an infinite loop. attackers may leverage the vulnerability leading to a Denial of Service since the string is not valid UTF16 and it results in it being sanitized before reaching the parser. <p>Publish Date: 2020-03-08 <p>URL: <a href=https://github.com/acornjs/acorn/commit/b5c17877ac0511e31579ea31e7650ba1a5871e51>WS-2020-0042</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1488">https://www.npmjs.com/advisories/1488</a></p> <p>Release Date: 2020-03-08</p> <p>Fix Resolution: 7.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
ws high detected in acorn tgz acorn tgz ws high severity vulnerability vulnerable libraries acorn tgz acorn tgz acorn tgz ecmascript parser library home page a href path to dependency file tmp ws scm shipengine request package json path to vulnerable library tmp ws scm shipengine request node modules acorn package json dependency hierarchy jest tgz root library jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz x acorn tgz vulnerable library acorn tgz ecmascript parser library home page a href path to dependency file tmp ws scm shipengine request package json path to vulnerable library tmp ws scm shipengine request node modules acorn globals node modules acorn package json dependency hierarchy jest tgz root library jest cli tgz jest config tgz jest environment jsdom tgz jsdom tgz acorn globals tgz x acorn tgz vulnerable library found in head commit a href vulnerability details acorn is vulnerable to regex dos a regex of the form u causes the parser to enter an infinite loop attackers may leverage the vulnerability leading to a denial of service since the string is not valid and it results in it being sanitized before reaching the parser publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
62,224
8,582,249,667
IssuesEvent
2018-11-13 16:31:52
godaddy/wp-primer-theme
https://api.github.com/repos/godaddy/wp-primer-theme
closed
Update documentation for latest release
bug documentation
The documentation hasn't been updated since v1.8.2. Investigate why this hasn't been updating in our build & release process and fix it. https://godaddy.github.io/wp-primer-theme/index.html
1.0
Update documentation for latest release - The documentation hasn't been updated since v1.8.2. Investigate why this hasn't been updating in our build & release process and fix it. https://godaddy.github.io/wp-primer-theme/index.html
non_defect
update documentation for latest release the documentation hasn t been updated since investigate why this hasn t been updating in our build release process and fix it
0
55,238
14,289,208,147
IssuesEvent
2020-11-23 18:53:22
idaholab/raven
https://api.github.com/repos/idaholab/raven
closed
[DEFECT] statsmodels error
defect priority_normal
-------- Defect Description -------- **Describe the defect** When using the internal parallel as True, and sampling the ARMA, it might cause the error of `statsmodels.tsa.statespace._initialization.dInitialization.__init__` ##### What did you expect to see happen? ##### What did you see instead? ``` ( 33.51 sec) STEP MULTIRUN : Message -> *** Beginning run *** (pid=57717) Unable to load "timingProfile" decorator; replacing with passthrough ... (pid=57717) Unable to load "memoryProfile" decorator; replacing with passthrough ... (pid=57718) Unable to load "timingProfile" decorator; replacing with passthrough ... (pid=57718) Unable to load "memoryProfile" decorator; replacing with passthrough ... (pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1 (pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1 (pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1 (pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1 (pid=57718) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working (pid=57718) from collections import Mapping (pid=57717) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working (pid=57717) from collections import Mapping Traceback (most recent call last): File "/home/zhoujia/projects/raven/framework/Driver.py", line 305, in <module> raven() File "/home/zhoujia/projects/raven/framework/Driver.py", line 258, in raven simulation.run() File "/home/zhoujia/projects/raven/framework/Simulation.py", line 762, in run stepInstance.takeAstep(stepInputDict) File "/home/zhoujia/projects/raven/framework/Steps.py", line 346, in takeAstep self._localTakeAstepRun(inDictionary) File "/home/zhoujia/projects/raven/framework/Steps.py", line 690, in _localTakeAstepRun myLambda([finishedJob,outputs[outIndex]]) File "/home/zhoujia/projects/raven/framework/Steps.py", line 637, in <lambda> self._outputCollectionLambda.append( (lambda x: inDictionary['Model'].collectOutput(x[0],x[1]), outIndex) ) File "/home/zhoujia/projects/raven/framework/Models/EnsembleModel.py", line 453, in collectOutput evaluation = finishedJob.getEvaluation() File "/home/zhoujia/projects/raven/framework/Runners/InternalRunner.py", line 124, in getEvaluation self._collectRunnerResponse() File "/home/zhoujia/projects/raven/framework/Runners/DistributedMemoryRunner.py", line 109, in _collectRunnerResponse self.runReturn = ray.get(self.thread) if im.isLibAvail("ray") else self.thread() File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/worker.py", line 1428, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(ValueError): ray::evaluateSample() (pid=57718, ip=141.221.84.10) File "python/ray/_raylet.pyx", line 448, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 465, in ray._raylet.execute_task File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 308, in deserialize_objects self._deserialize_object(data, metadata, object_ref)) File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 247, in _deserialize_object return self._deserialize_msgpack_data(data, metadata) File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 226, in _deserialize_msgpack_data python_objects = self._deserialize_pickle5_data(pickle5_data) File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 214, in _deserialize_pickle5_data obj = pickle.loads(in_band, buffers=buffers) File "statsmodels/tsa/statespace/_initialization.pyx", line 227, in statsmodels.tsa.statespace._initialization.dInitialization.__init__ File "stringsource", line 658, in View.MemoryView.memoryview_cwrapper File "stringsource", line 349, in View.MemoryView.memoryview.__cinit__ ValueError: buffer source array is read-only ``` ##### Do you have a suggested fix for the development team? **Describe how to Reproduce** Steps to reproduce the behavior: 1. 2. 3. 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or task? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
1.0
[DEFECT] statsmodels error - -------- Defect Description -------- **Describe the defect** When using the internal parallel as True, and sampling the ARMA, it might cause the error of `statsmodels.tsa.statespace._initialization.dInitialization.__init__` ##### What did you expect to see happen? ##### What did you see instead? ``` ( 33.51 sec) STEP MULTIRUN : Message -> *** Beginning run *** (pid=57717) Unable to load "timingProfile" decorator; replacing with passthrough ... (pid=57717) Unable to load "memoryProfile" decorator; replacing with passthrough ... (pid=57718) Unable to load "timingProfile" decorator; replacing with passthrough ... (pid=57718) Unable to load "memoryProfile" decorator; replacing with passthrough ... (pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1 (pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1 (pid=57717) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1 (pid=57717) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in UniformDiscrete len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in UniformDiscrete len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to lowerBound in LogUniform len 1 (pid=57718) INPUT SPEC ERROR adding checked and unchecked to upperBound in LogUniform len 1 (pid=57718) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working (pid=57718) from collections import Mapping (pid=57717) /home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/patsy/constraint.py:13: DeprecationWarning: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3,and in 3.9 it will stop working (pid=57717) from collections import Mapping Traceback (most recent call last): File "/home/zhoujia/projects/raven/framework/Driver.py", line 305, in <module> raven() File "/home/zhoujia/projects/raven/framework/Driver.py", line 258, in raven simulation.run() File "/home/zhoujia/projects/raven/framework/Simulation.py", line 762, in run stepInstance.takeAstep(stepInputDict) File "/home/zhoujia/projects/raven/framework/Steps.py", line 346, in takeAstep self._localTakeAstepRun(inDictionary) File "/home/zhoujia/projects/raven/framework/Steps.py", line 690, in _localTakeAstepRun myLambda([finishedJob,outputs[outIndex]]) File "/home/zhoujia/projects/raven/framework/Steps.py", line 637, in <lambda> self._outputCollectionLambda.append( (lambda x: inDictionary['Model'].collectOutput(x[0],x[1]), outIndex) ) File "/home/zhoujia/projects/raven/framework/Models/EnsembleModel.py", line 453, in collectOutput evaluation = finishedJob.getEvaluation() File "/home/zhoujia/projects/raven/framework/Runners/InternalRunner.py", line 124, in getEvaluation self._collectRunnerResponse() File "/home/zhoujia/projects/raven/framework/Runners/DistributedMemoryRunner.py", line 109, in _collectRunnerResponse self.runReturn = ray.get(self.thread) if im.isLibAvail("ray") else self.thread() File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/worker.py", line 1428, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(ValueError): ray::evaluateSample() (pid=57718, ip=141.221.84.10) File "python/ray/_raylet.pyx", line 448, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 465, in ray._raylet.execute_task File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 308, in deserialize_objects self._deserialize_object(data, metadata, object_ref)) File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 247, in _deserialize_object return self._deserialize_msgpack_data(data, metadata) File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 226, in _deserialize_msgpack_data python_objects = self._deserialize_pickle5_data(pickle5_data) File "/home/zhoujia/.conda/envs/raven_libraries/lib/python3.7/site-packages/ray/serialization.py", line 214, in _deserialize_pickle5_data obj = pickle.loads(in_band, buffers=buffers) File "statsmodels/tsa/statespace/_initialization.pyx", line 227, in statsmodels.tsa.statespace._initialization.dInitialization.__init__ File "stringsource", line 658, in View.MemoryView.memoryview_cwrapper File "stringsource", line 349, in View.MemoryView.memoryview.__cinit__ ValueError: buffer source array is read-only ``` ##### Do you have a suggested fix for the development team? **Describe how to Reproduce** Steps to reproduce the behavior: 1. 2. 3. 4. **Screenshots and Input Files** Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. **Platform (please complete the following information):** - OS: [e.g. iOS] - Version: [e.g. 22] - Dependencies Installation: [CONDA or PIP] ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or task? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest release branch? If yes, is there any issue tagged with release (create if needed)? - [ ] 5. If the issue is being closed without a pull request, has an explanation of why it is being closed been provided?
defect
statsmodels error defect description describe the defect when using the internal parallel as true and sampling the arma it might cause the error of statsmodels tsa statespace initialization dinitialization init what did you expect to see happen what did you see instead sec step multirun message beginning run pid unable to load timingprofile decorator replacing with passthrough pid unable to load memoryprofile decorator replacing with passthrough pid unable to load timingprofile decorator replacing with passthrough pid unable to load memoryprofile decorator replacing with passthrough pid input spec error adding checked and unchecked to lowerbound in uniformdiscrete len pid input spec error adding checked and unchecked to upperbound in uniformdiscrete len pid input spec error adding checked and unchecked to lowerbound in loguniform len pid input spec error adding checked and unchecked to upperbound in loguniform len pid input spec error adding checked and unchecked to lowerbound in uniformdiscrete len pid input spec error adding checked and unchecked to upperbound in uniformdiscrete len pid input spec error adding checked and unchecked to lowerbound in loguniform len pid input spec error adding checked and unchecked to upperbound in loguniform len pid home zhoujia conda envs raven libraries lib site packages patsy constraint py deprecationwarning using or importing the abcs from collections instead of from collections abc is deprecated since python and in it will stop working pid from collections import mapping pid home zhoujia conda envs raven libraries lib site packages patsy constraint py deprecationwarning using or importing the abcs from collections instead of from collections abc is deprecated since python and in it will stop working pid from collections import mapping traceback most recent call last file home zhoujia projects raven framework driver py line in raven file home zhoujia projects raven framework driver py line in raven simulation run file home zhoujia projects raven framework simulation py line in run stepinstance takeastep stepinputdict file home zhoujia projects raven framework steps py line in takeastep self localtakeasteprun indictionary file home zhoujia projects raven framework steps py line in localtakeasteprun mylambda file home zhoujia projects raven framework steps py line in self outputcollectionlambda append lambda x indictionary collectoutput x x outindex file home zhoujia projects raven framework models ensemblemodel py line in collectoutput evaluation finishedjob getevaluation file home zhoujia projects raven framework runners internalrunner py line in getevaluation self collectrunnerresponse file home zhoujia projects raven framework runners distributedmemoryrunner py line in collectrunnerresponse self runreturn ray get self thread if im islibavail ray else self thread file home zhoujia conda envs raven libraries lib site packages ray worker py line in get raise value as instanceof cause ray exceptions raytaskerror valueerror ray evaluatesample pid ip file python ray raylet pyx line in ray raylet execute task file python ray raylet pyx line in ray raylet execute task file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize objects self deserialize object data metadata object ref file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize object return self deserialize msgpack data data metadata file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize msgpack data python objects self deserialize data data file home zhoujia conda envs raven libraries lib site packages ray serialization py line in deserialize data obj pickle loads in band buffers buffers file statsmodels tsa statespace initialization pyx line in statsmodels tsa statespace initialization dinitialization init file stringsource line in view memoryview memoryview cwrapper file stringsource line in view memoryview memoryview cinit valueerror buffer source array is read only do you have a suggested fix for the development team describe how to reproduce steps to reproduce the behavior screenshots and input files please attach the input file s that generate this error the simpler the input the faster we can find the issue platform please complete the following information os version dependencies installation for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or task is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest release branch if yes is there any issue tagged with release create if needed if the issue is being closed without a pull request has an explanation of why it is being closed been provided
1
53,643
13,262,039,387
IssuesEvent
2020-08-20 20:59:43
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[mue] Memory Leak (Trac #1801)
Migrated from Trac combo reconstruction defect
found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1801">https://code.icecube.wisc.edu/projects/icecube/ticket/1801</a>, reported by kjmeagherand owned by dima</em></summary> <p> ```json { "status": "closed", "changetime": "2016-08-01T18:41:59", "_ts": "1470076919611236", "description": "found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2016-07-27T08:06:31", "component": "combo reconstruction", "summary": "[mue] Memory Leak", "priority": "normal", "keywords": "", "milestone": "Long-Term Future", "owner": "dima", "type": "defect" } ``` </p> </details>
1.0
[mue] Memory Leak (Trac #1801) - found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1801">https://code.icecube.wisc.edu/projects/icecube/ticket/1801</a>, reported by kjmeagherand owned by dima</em></summary> <p> ```json { "status": "closed", "changetime": "2016-08-01T18:41:59", "_ts": "1470076919611236", "description": "found by static analyser http://software.icecube.wisc.edu/static_analysis/2016-07-26-030212-26135-1/report-ca0310.html#EndPath", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "time": "2016-07-27T08:06:31", "component": "combo reconstruction", "summary": "[mue] Memory Leak", "priority": "normal", "keywords": "", "milestone": "Long-Term Future", "owner": "dima", "type": "defect" } ``` </p> </details>
defect
memory leak trac found by static analyser migrated from json status closed changetime ts description found by static analyser reporter kjmeagher cc resolution fixed time component combo reconstruction summary memory leak priority normal keywords milestone long term future owner dima type defect
1
1,153
2,598,004,852
IssuesEvent
2015-02-22 01:41:55
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
opened
Can't build/land/etc on far right and/or bottom edge of the map
auto-migrated Component-Logic Priority-High Type-Defect Usability
``` What steps will reproduce the problem? 1. Call canBuildHere(), placing a building at a clear location whose edge is against the far right and/or bottom edge of the map. What is the expected output? What do you see instead? It should return true, but it always returns false. What version of the product are you using? On what operating system? HEAD of branch bwapi4 of the src Please provide any additional information below. I think Shared\Templates.h:canBuildHere is returning false because the rb variable is initialised to an invalid location (i.e. off the edge of the map) during this statement: TilePosition rb = position + TilePosition(width, height); It would probably be simple to fix this function, but I expect that similar bugs exist in several other places in the BWAPI src, and there may be a lot of src in BWAPI (and other libraries & bots) that is only called if this function returns true and manipulates position-related objects without doing map array bounds safety checks. My concern is that if just this function is fixed then unless all such src is fixed too then there may be crashes or strange problems caused by memory corruption, and these bugs may be hard to reproduce/debug because they would probably only occur in rare scenarios, and may not even be consistently reproducable. So, risk may not be worth the effort. ``` ----- Original issue reported on code.google.com by `chris.c...@gmail.com` on 23 May 2013 at 2:01
1.0
Can't build/land/etc on far right and/or bottom edge of the map - ``` What steps will reproduce the problem? 1. Call canBuildHere(), placing a building at a clear location whose edge is against the far right and/or bottom edge of the map. What is the expected output? What do you see instead? It should return true, but it always returns false. What version of the product are you using? On what operating system? HEAD of branch bwapi4 of the src Please provide any additional information below. I think Shared\Templates.h:canBuildHere is returning false because the rb variable is initialised to an invalid location (i.e. off the edge of the map) during this statement: TilePosition rb = position + TilePosition(width, height); It would probably be simple to fix this function, but I expect that similar bugs exist in several other places in the BWAPI src, and there may be a lot of src in BWAPI (and other libraries & bots) that is only called if this function returns true and manipulates position-related objects without doing map array bounds safety checks. My concern is that if just this function is fixed then unless all such src is fixed too then there may be crashes or strange problems caused by memory corruption, and these bugs may be hard to reproduce/debug because they would probably only occur in rare scenarios, and may not even be consistently reproducable. So, risk may not be worth the effort. ``` ----- Original issue reported on code.google.com by `chris.c...@gmail.com` on 23 May 2013 at 2:01
defect
can t build land etc on far right and or bottom edge of the map what steps will reproduce the problem call canbuildhere placing a building at a clear location whose edge is against the far right and or bottom edge of the map what is the expected output what do you see instead it should return true but it always returns false what version of the product are you using on what operating system head of branch of the src please provide any additional information below i think shared templates h canbuildhere is returning false because the rb variable is initialised to an invalid location i e off the edge of the map during this statement tileposition rb position tileposition width height it would probably be simple to fix this function but i expect that similar bugs exist in several other places in the bwapi src and there may be a lot of src in bwapi and other libraries bots that is only called if this function returns true and manipulates position related objects without doing map array bounds safety checks my concern is that if just this function is fixed then unless all such src is fixed too then there may be crashes or strange problems caused by memory corruption and these bugs may be hard to reproduce debug because they would probably only occur in rare scenarios and may not even be consistently reproducable so risk may not be worth the effort original issue reported on code google com by chris c gmail com on may at
1
52,231
7,754,237,826
IssuesEvent
2018-05-31 05:39:08
dustinRDF/food-roulette
https://api.github.com/repos/dustinRDF/food-roulette
closed
Create Cards for Each Section
Documentation
- [x] Application Information #11 - [x] Application Design #12 - [x] Development and Testing #13 - [x] Deployment #14
1.0
Create Cards for Each Section - - [x] Application Information #11 - [x] Application Design #12 - [x] Development and Testing #13 - [x] Deployment #14
non_defect
create cards for each section application information application design development and testing deployment
0
802,968
29,086,615,502
IssuesEvent
2023-05-16 00:50:19
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
pubsublite/internal/wire: TestRoutingPublisherUnloadIdlePublisher failed
type: bug priority: p1 api: pubsublite flakybot: issue
Note: #7854 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 85d084cc9d22f03ec60356d9f875c143d92a4aba buildURL: [Build Status](https://source.cloud.google.com/results/invocations/5cb50424-759c-407f-867d-e2264dd350dc), [Sponge](http://sponge2/5cb50424-759c-407f-867d-e2264dd350dc) status: failed <details><summary>Test output</summary><br><pre> verifier.go:226: call(2): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"2"} first_sequence_number:1} verifier.go:226: call(3): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"3"} first_sequence_number:2} verifier.go:226: call(4): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"4"} first_sequence_number:3} verifier.go:226: call(5): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"5"} first_sequence_number:4} verifier.go:226: call(6): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"6"} first_sequence_number:5} verifier.go:257: unexpected stream index 1 for key publish:projects/123456/locations/us-central1-b/topics/my-topic:0 publish_batcher_test.go:71: Publish("2") error: (rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection), want: partition=0,offset=11 publish_batcher_test.go:71: Publish("3") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=12 publish_batcher_test.go:71: Publish("4") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=13 publish_batcher_test.go:71: Publish("5") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=14 publish_batcher_test.go:71: Publish("6") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=15 publish_batcher_test.go:71: Publish("b") error: (pubsublite: service has stopped or is stopping), want: partition=1,offset=22 publisher_test.go:619: Stop() got err: (publisher(projects/123456/locations/us-central1-b/topics/my-topic/partitions/0): rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection) verifier.go:226: call(0): did not receive expected request: [*pubsublitepb.PublishRequest] initial_request:{topic:"projects/123456/locations/us-central1-b/topics/my-topic" partition:1 client_id:"publisher"} verifier.go:226: call(1): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"bar" data:"b"} first_sequence_number:1}</pre></details>
1.0
pubsublite/internal/wire: TestRoutingPublisherUnloadIdlePublisher failed - Note: #7854 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 85d084cc9d22f03ec60356d9f875c143d92a4aba buildURL: [Build Status](https://source.cloud.google.com/results/invocations/5cb50424-759c-407f-867d-e2264dd350dc), [Sponge](http://sponge2/5cb50424-759c-407f-867d-e2264dd350dc) status: failed <details><summary>Test output</summary><br><pre> verifier.go:226: call(2): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"2"} first_sequence_number:1} verifier.go:226: call(3): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"3"} first_sequence_number:2} verifier.go:226: call(4): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"4"} first_sequence_number:3} verifier.go:226: call(5): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"5"} first_sequence_number:4} verifier.go:226: call(6): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"baz" data:"6"} first_sequence_number:5} verifier.go:257: unexpected stream index 1 for key publish:projects/123456/locations/us-central1-b/topics/my-topic:0 publish_batcher_test.go:71: Publish("2") error: (rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection), want: partition=0,offset=11 publish_batcher_test.go:71: Publish("3") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=12 publish_batcher_test.go:71: Publish("4") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=13 publish_batcher_test.go:71: Publish("5") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=14 publish_batcher_test.go:71: Publish("6") error: (pubsublite: service has stopped or is stopping), want: partition=0,offset=15 publish_batcher_test.go:71: Publish("b") error: (pubsublite: service has stopped or is stopping), want: partition=1,offset=22 publisher_test.go:619: Stop() got err: (publisher(projects/123456/locations/us-central1-b/topics/my-topic/partitions/0): rpc error: code = FailedPrecondition desc = mockserver: got unexpected stream connection) verifier.go:226: call(0): did not receive expected request: [*pubsublitepb.PublishRequest] initial_request:{topic:"projects/123456/locations/us-central1-b/topics/my-topic" partition:1 client_id:"publisher"} verifier.go:226: call(1): did not receive expected request: [*pubsublitepb.PublishRequest] message_publish_request:{messages:{key:"bar" data:"b"} first_sequence_number:1}</pre></details>
non_defect
pubsublite internal wire testroutingpublisherunloadidlepublisher failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go call did not receive expected request message publish request messages key baz data first sequence number verifier go unexpected stream index for key publish projects locations us b topics my topic publish batcher test go publish error rpc error code failedprecondition desc mockserver got unexpected stream connection want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish error pubsublite service has stopped or is stopping want partition offset publish batcher test go publish b error pubsublite service has stopped or is stopping want partition offset publisher test go stop got err publisher projects locations us b topics my topic partitions rpc error code failedprecondition desc mockserver got unexpected stream connection verifier go call did not receive expected request initial request topic projects locations us b topics my topic partition client id publisher verifier go call did not receive expected request message publish request messages key bar data b first sequence number
0
21,445
11,660,236,226
IssuesEvent
2020-03-03 02:39:09
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
opened
Signal Timing and Phasing Dashboard
Product: Data & Performance Hub Service: Dev Type: New Application Workgroup: AMD migrated
Description: See also: SR19-104475 . Brian has requested that we display signal timing and phasing information from D4 on a map/dashboard so that it can be viewed by TMC staff and field technicians. Definitely need to clarify use case. Request Date: 2019-03-15 17:01:00 Request ID: DTS19-104480 Status: Backlog Customer Priority: 4 Level of Effort: Small Project DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c8c20a2a49e7c3051e72d52 *Migrated from [atd-data-and-performance #268](https://github.com/cityofaustin/atd-data-and-performance/issues/268)*
1.0
Signal Timing and Phasing Dashboard - Description: See also: SR19-104475 . Brian has requested that we display signal timing and phasing information from D4 on a map/dashboard so that it can be viewed by TMC staff and field technicians. Definitely need to clarify use case. Request Date: 2019-03-15 17:01:00 Request ID: DTS19-104480 Status: Backlog Customer Priority: 4 Level of Effort: Small Project DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c8c20a2a49e7c3051e72d52 *Migrated from [atd-data-and-performance #268](https://github.com/cityofaustin/atd-data-and-performance/issues/268)*
non_defect
signal timing and phasing dashboard description see also brian has requested that we display signal timing and phasing information from on a map dashboard so that it can be viewed by tmc staff and field technicians definitely need to clarify use case request date request id status backlog customer priority level of effort small project dts url migrated from
0
19,348
3,193,184,581
IssuesEvent
2015-09-30 02:32:00
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
dart2js: assert_with_message_test fails in minified code
Area-Dart2JS Priority-High Triaged Type-Defect
The minified interceptor name for JSNull is being rejected, leading to the evaluation of null.constructor. This happens in the formatting of the TypeError for a failed boolean conversion for the case when the assertion expression evaluates to non-bool. The null deref is trapped an converted to a NullError. Seems related to: https://github.com/dart-lang/sdk/issues/19137 ---- @@@BUILD_STEP dart2js_extra-drt tests --minified --dart2js-batch --csp --checked failures@@@ FAILED: dart2js-drt-checked release_ia32 dart2js_extra/assert_with_message_test Expected: Pass Actual: RuntimeError CommandOutput[dart2js]: CommandOutput[content_shell]: stdout: #READY CONSOLE MESSAGE: line 78: { "type": "debug", "value": "test_controller.js started", "timestamp": 0 } CONSOLE MESSAGE: line 78: { "type": "print", "value": "dart-calling-main", "timestamp": 0.064 } CONSOLE MESSAGE: line 78: { "type": "sync_exception", "value": "Exception: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\nStack: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\n at dart.d (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:682:3)\n at dart.aB (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1510:24)\n at dart.aC (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1509:3)\n at ak.dart.ak.$2 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1295:3)\n at dart.c5 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1242:3)\n at dart.bw (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1266:3)\n at dartMainRunner (http://127.0.0.1:41436/root_dart/tools/testing/dart/test_controller.js:311:5)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1638:55\n at init.currentScript (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1634:55)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1637:84", "timestamp": 0.092 }
1.0
dart2js: assert_with_message_test fails in minified code - The minified interceptor name for JSNull is being rejected, leading to the evaluation of null.constructor. This happens in the formatting of the TypeError for a failed boolean conversion for the case when the assertion expression evaluates to non-bool. The null deref is trapped an converted to a NullError. Seems related to: https://github.com/dart-lang/sdk/issues/19137 ---- @@@BUILD_STEP dart2js_extra-drt tests --minified --dart2js-batch --csp --checked failures@@@ FAILED: dart2js-drt-checked release_ia32 dart2js_extra/assert_with_message_test Expected: Pass Actual: RuntimeError CommandOutput[dart2js]: CommandOutput[content_shell]: stdout: #READY CONSOLE MESSAGE: line 78: { "type": "debug", "value": "test_controller.js started", "timestamp": 0 } CONSOLE MESSAGE: line 78: { "type": "print", "value": "dart-calling-main", "timestamp": 0.064 } CONSOLE MESSAGE: line 78: { "type": "sync_exception", "value": "Exception: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\nStack: Expect.isTrue(false, 'name: <<NullError: method not found: 'constructor' on null>> (F) is TypeError') fails.\n at dart.d (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:682:3)\n at dart.aB (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1510:24)\n at dart.aC (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1509:3)\n at ak.dart.ak.$2 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1295:3)\n at dart.c5 (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1242:3)\n at dart.bw (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1266:3)\n at dartMainRunner (http://127.0.0.1:41436/root_dart/tools/testing/dart/test_controller.js:311:5)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1638:55\n at init.currentScript (http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1634:55)\n at http://127.0.0.1:41436/root_build/generated_compilations/dart2js-checked-minified-csp-sdk/tests_compiler_dart2js_extra_assert_with_message_test/test.js:1637:84", "timestamp": 0.092 }
defect
assert with message test fails in minified code the minified interceptor name for jsnull is being rejected leading to the evaluation of null constructor this happens in the formatting of the typeerror for a failed boolean conversion for the case when the assertion expression evaluates to non bool the null deref is trapped an converted to a nullerror seems related to build step extra drt tests minified batch csp checked failures failed drt checked release extra assert with message test expected pass actual runtimeerror commandoutput commandoutput stdout ready console message line type debug value test controller js started timestamp console message line type print value dart calling main timestamp console message line type sync exception value exception expect istrue false name f is typeerror fails nstack expect istrue false name f is typeerror fails n at dart d at dart ab at dart ac at ak dart ak at dart at dart bw at dartmainrunner at at init currentscript at timestamp
1
59,583
17,023,168,320
IssuesEvent
2021-07-03 00:40:50
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
cannot access account settings
Component: website Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 3.38pm, Saturday, 26th May 2007]** Whenever I try to access my account settings at http://www.openstreetmap.org/user/Marc+Sch%FCtz/account, I get the following error: Application error Change this error message for exceptions thrown outside of an action (like in Dispatcher setups or broken Ruby code) in public/500.html Maybe this is related to the %FC in the URL?
1.0
cannot access account settings - **[Submitted to the original trac issue database at 3.38pm, Saturday, 26th May 2007]** Whenever I try to access my account settings at http://www.openstreetmap.org/user/Marc+Sch%FCtz/account, I get the following error: Application error Change this error message for exceptions thrown outside of an action (like in Dispatcher setups or broken Ruby code) in public/500.html Maybe this is related to the %FC in the URL?
defect
cannot access account settings whenever i try to access my account settings at i get the following error application error change this error message for exceptions thrown outside of an action like in dispatcher setups or broken ruby code in public html maybe this is related to the fc in the url
1
335,118
24,454,563,436
IssuesEvent
2022-10-07 04:53:52
justicehub-in/justice-hub-docs
https://api.github.com/repos/justicehub-in/justice-hub-docs
closed
Data points to curate
documentation data-request
- [ ] Parliament session wise questions for Law and Justice - [ ] High Court data released by Daksh
1.0
Data points to curate - - [ ] Parliament session wise questions for Law and Justice - [ ] High Court data released by Daksh
non_defect
data points to curate parliament session wise questions for law and justice high court data released by daksh
0
69,135
22,199,980,864
IssuesEvent
2022-06-07 10:16:55
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Mismatch between reactions on thread root between main timeline and thread view
T-Defect S-Major A-Reactions O-Occasional A-Threads
### Steps to reproduce 1. Where are you starting? What can you see? 2. What do you click? 3. More steps… ### Outcome #### What did you expect? Reactions to be the same #### What happened instead? Fewer reactions on the main list ![Screenshot from 2022-05-20 17-35-14](https://user-images.githubusercontent.com/8428120/169573734-e61e7ffb-a05b-429c-b77d-a233a82658e7.png) ### Operating system Ubuntu ### Browser information Chrome ### URL for webapp develop.element.io ### Application version _No response_ ### Homeserver jki.re ### Will you send logs? No
1.0
Mismatch between reactions on thread root between main timeline and thread view - ### Steps to reproduce 1. Where are you starting? What can you see? 2. What do you click? 3. More steps… ### Outcome #### What did you expect? Reactions to be the same #### What happened instead? Fewer reactions on the main list ![Screenshot from 2022-05-20 17-35-14](https://user-images.githubusercontent.com/8428120/169573734-e61e7ffb-a05b-429c-b77d-a233a82658e7.png) ### Operating system Ubuntu ### Browser information Chrome ### URL for webapp develop.element.io ### Application version _No response_ ### Homeserver jki.re ### Will you send logs? No
defect
mismatch between reactions on thread root between main timeline and thread view steps to reproduce where are you starting what can you see what do you click more steps… outcome what did you expect reactions to be the same what happened instead fewer reactions on the main list operating system ubuntu browser information chrome url for webapp develop element io application version no response homeserver jki re will you send logs no
1
5,721
2,610,214,096
IssuesEvent
2015-02-26 19:08:21
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
yoksel dll
auto-migrated Priority-Medium Type-Defect
``` '''Владелин Лапин''' День добрый никак не могу найти .yoksel dll. где то видел уже '''Гарибальди Анисимов''' Качай тут http://bit.ly/16sJ2P3 '''Афиноген Рожков''' Спасибо вроде то но просит телефон вводить '''Гелеон Соколов''' Не это не влияет на баланс '''Аркадий Русаков''' Неа все ок у меня ничего не списало Информация о файле: yoksel dll Загружен: В этом месяце Скачан раз: 234 Рейтинг: 1475 Средняя скорость скачивания: 181 Похожих файлов: 15 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 12:25
1.0
yoksel dll - ``` '''Владелин Лапин''' День добрый никак не могу найти .yoksel dll. где то видел уже '''Гарибальди Анисимов''' Качай тут http://bit.ly/16sJ2P3 '''Афиноген Рожков''' Спасибо вроде то но просит телефон вводить '''Гелеон Соколов''' Не это не влияет на баланс '''Аркадий Русаков''' Неа все ок у меня ничего не списало Информация о файле: yoksel dll Загружен: В этом месяце Скачан раз: 234 Рейтинг: 1475 Средняя скорость скачивания: 181 Похожих файлов: 15 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 16 Dec 2013 at 12:25
defect
yoksel dll владелин лапин день добрый никак не могу найти yoksel dll где то видел уже гарибальди анисимов качай тут афиноген рожков спасибо вроде то но просит телефон вводить гелеон соколов не это не влияет на баланс аркадий русаков неа все ок у меня ничего не списало информация о файле yoksel dll загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
12,095
5,145,208,685
IssuesEvent
2017-01-12 20:54:32
dotnet/cli
https://api.github.com/repos/dotnet/cli
closed
Razor not copied to folder on change
msbuild-mitigated
When I change only a razor (cshtml) page and build/rebuild the solution I do not see the changes applied, files dont get copied to build folder. need to remove bin folder or change a controller to copy all new razor content. ## Steps to reproduce modify a razor page (.cshtml) ## Expected behaviour always copy content files even if they are not changed ## Actual behaviour Does not copy Views unless a .cs file has been modified. ## Environment data `dotnet --info` output: Product Information: Version: 1.0.0-preview1-002700 Commit Sha: 0336f6bb34 Runtime Environment: OS Name: Mac OS X OS Version: 10.11 OS Platform: Darwin RID: osx.10.11-x64
1.0
Razor not copied to folder on change - When I change only a razor (cshtml) page and build/rebuild the solution I do not see the changes applied, files dont get copied to build folder. need to remove bin folder or change a controller to copy all new razor content. ## Steps to reproduce modify a razor page (.cshtml) ## Expected behaviour always copy content files even if they are not changed ## Actual behaviour Does not copy Views unless a .cs file has been modified. ## Environment data `dotnet --info` output: Product Information: Version: 1.0.0-preview1-002700 Commit Sha: 0336f6bb34 Runtime Environment: OS Name: Mac OS X OS Version: 10.11 OS Platform: Darwin RID: osx.10.11-x64
non_defect
razor not copied to folder on change when i change only a razor cshtml page and build rebuild the solution i do not see the changes applied files dont get copied to build folder need to remove bin folder or change a controller to copy all new razor content steps to reproduce modify a razor page cshtml expected behaviour always copy content files even if they are not changed actual behaviour does not copy views unless a cs file has been modified environment data dotnet info output product information version commit sha runtime environment os name mac os x os version os platform darwin rid osx
0
59,436
24,767,496,805
IssuesEvent
2022-10-22 18:16:22
openkfw/TruBudget
https://api.github.com/repos/openkfw/TruBudget
closed
Improve notification link to item
bug wontfix blockchain api frontend storage-service
- [ ] I have searched the [issues](https://github.com/openkfw/TruBudget/issues) of this repository and believe that this is not a duplicate. ## Description 😯 Problem: when a user clicks on a notification linked to an WFI in trubudget, he is directed to the "project" level. He does not know therefore in some component is the WFI. Need: When users click on a notification, trubudget should take them to the relevant level in the user interface. Thus, a notification concerning an WFI must lead the user who clicks on this notification to the WFI concerned. ### How to reproduce 🕹 ## Your Environment 🌎 | Tech | Version | | --------------- | ------- | | Frontend | v1.?.? | | API | v1.?.? | | Blockchain | v1.?.? | | Export-Service | v1.?.? | | Email-Service | v1.?.? | | Storage-Service | v1.?.? |
1.0
Improve notification link to item - - [ ] I have searched the [issues](https://github.com/openkfw/TruBudget/issues) of this repository and believe that this is not a duplicate. ## Description 😯 Problem: when a user clicks on a notification linked to an WFI in trubudget, he is directed to the "project" level. He does not know therefore in some component is the WFI. Need: When users click on a notification, trubudget should take them to the relevant level in the user interface. Thus, a notification concerning an WFI must lead the user who clicks on this notification to the WFI concerned. ### How to reproduce 🕹 ## Your Environment 🌎 | Tech | Version | | --------------- | ------- | | Frontend | v1.?.? | | API | v1.?.? | | Blockchain | v1.?.? | | Export-Service | v1.?.? | | Email-Service | v1.?.? | | Storage-Service | v1.?.? |
non_defect
improve notification link to item i have searched the of this repository and believe that this is not a duplicate description 😯 problem when a user clicks on a notification linked to an wfi in trubudget he is directed to the project level he does not know therefore in some component is the wfi need when users click on a notification trubudget should take them to the relevant level in the user interface thus a notification concerning an wfi must lead the user who clicks on this notification to the wfi concerned how to reproduce 🕹 your environment 🌎 tech version frontend api blockchain export service email service storage service
0
814,543
30,511,454,531
IssuesEvent
2023-07-18 21:12:58
DiscoTrayStudios/hendrix-today
https://api.github.com/repos/DiscoTrayStudios/hendrix-today
closed
Hendrix Today Uploader - CI/CD with GitHub Actions
category: backend priority: medium size: medium type: incomplete
**Describe the incomplete feature** Building and deployment of the Hendrix Today Uploader is not automated. **Describe the solution you'd like** The uploader should be rebuilt and re-deployed on each push to `main`.
1.0
Hendrix Today Uploader - CI/CD with GitHub Actions - **Describe the incomplete feature** Building and deployment of the Hendrix Today Uploader is not automated. **Describe the solution you'd like** The uploader should be rebuilt and re-deployed on each push to `main`.
non_defect
hendrix today uploader ci cd with github actions describe the incomplete feature building and deployment of the hendrix today uploader is not automated describe the solution you d like the uploader should be rebuilt and re deployed on each push to main
0
76,669
26,545,312,014
IssuesEvent
2023-01-19 23:20:51
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
Preview button not working for some Lovell nodes
Defect ⭐️ Facilities Needs refining
## Describe the defect As a Lovell editor, I want to be able to preview content that is in the Lovell-VA or Lovell-TRICARE sections. This is currently broken for VAMC Detail Pages. The Preview button works if they are in the Lovell Federal section, but not if they are in the Lovell-VA or Lovell-TRICARE sections. I created a test page to investigate the cases in which this is true. I have not investigated other content types. ## To Reproduce Steps to reproduce the behavior: 1. Go to [Revision log for this test page](https://prod.cms.va.gov/node/52805/revisions) 2. Revert to Revision #743400 or Revision #743399 (Section = Lovell - VA or TRICARE) 3. Confirm that Preview button leads to Internal Server Error 4. Revert to Revision #743398 (Section = Lovell Federal health care) 4. Confirm that Preview button magically works again ## AC / Expected behavior The Preview button should work consistently for content types regardless of which Lovell section they are in. ## Screenshots ![screencapture-prod-cms-va-gov-chicago-health-care-programs-test-page-do-not-publish-2023-01-19-16_46_24.png](https://images.zenhubusercontent.com/62acf41e5be34fcdb6c44c0b/1d3bc79d-e759-400a-b620-ccb69cf3c665) ![screencapture-preview-prod-vfs-va-gov-preview-2023-01-19-16_46_30.png](https://images.zenhubusercontent.com/62acf41e5be34fcdb6c44c0b/bb2e24d2-860c-45ad-a74f-c24cbf31c06e) ![Screen Shot 2023-01-19 at 5.12.29 PM.png](https://images.zenhubusercontent.com/62acf41e5be34fcdb6c44c0b/87ef4525-3656-4334-a049-1396a908f710)
1.0
Preview button not working for some Lovell nodes - ## Describe the defect As a Lovell editor, I want to be able to preview content that is in the Lovell-VA or Lovell-TRICARE sections. This is currently broken for VAMC Detail Pages. The Preview button works if they are in the Lovell Federal section, but not if they are in the Lovell-VA or Lovell-TRICARE sections. I created a test page to investigate the cases in which this is true. I have not investigated other content types. ## To Reproduce Steps to reproduce the behavior: 1. Go to [Revision log for this test page](https://prod.cms.va.gov/node/52805/revisions) 2. Revert to Revision #743400 or Revision #743399 (Section = Lovell - VA or TRICARE) 3. Confirm that Preview button leads to Internal Server Error 4. Revert to Revision #743398 (Section = Lovell Federal health care) 4. Confirm that Preview button magically works again ## AC / Expected behavior The Preview button should work consistently for content types regardless of which Lovell section they are in. ## Screenshots ![screencapture-prod-cms-va-gov-chicago-health-care-programs-test-page-do-not-publish-2023-01-19-16_46_24.png](https://images.zenhubusercontent.com/62acf41e5be34fcdb6c44c0b/1d3bc79d-e759-400a-b620-ccb69cf3c665) ![screencapture-preview-prod-vfs-va-gov-preview-2023-01-19-16_46_30.png](https://images.zenhubusercontent.com/62acf41e5be34fcdb6c44c0b/bb2e24d2-860c-45ad-a74f-c24cbf31c06e) ![Screen Shot 2023-01-19 at 5.12.29 PM.png](https://images.zenhubusercontent.com/62acf41e5be34fcdb6c44c0b/87ef4525-3656-4334-a049-1396a908f710)
defect
preview button not working for some lovell nodes describe the defect as a lovell editor i want to be able to preview content that is in the lovell va or lovell tricare sections this is currently broken for vamc detail pages the preview button works if they are in the lovell federal section but not if they are in the lovell va or lovell tricare sections i created a test page to investigate the cases in which this is true i have not investigated other content types to reproduce steps to reproduce the behavior go to revert to revision or revision section lovell va or tricare confirm that preview button leads to internal server error revert to revision section lovell federal health care confirm that preview button magically works again ac expected behavior the preview button should work consistently for content types regardless of which lovell section they are in screenshots
1
34,234
7,431,748,273
IssuesEvent
2018-03-25 17:43:29
Yahkal/replicaisland
https://api.github.com/repos/Yahkal/replicaisland
closed
"Google Code will be turning read-only on August 25th. See this post for more information."
Priority-Medium Type-Defect auto-migrated
``` https://code.google.com/p/support/wiki/ReadOnlyTransition I hope you're moving this repository someplace active.… ``` Original issue reported on code.google.com by `TPSam...@gmail.com` on 2 Aug 2015 at 4:10
1.0
"Google Code will be turning read-only on August 25th. See this post for more information." - ``` https://code.google.com/p/support/wiki/ReadOnlyTransition I hope you're moving this repository someplace active.… ``` Original issue reported on code.google.com by `TPSam...@gmail.com` on 2 Aug 2015 at 4:10
defect
google code will be turning read only on august see this post for more information i hope you re moving this repository someplace active … original issue reported on code google com by tpsam gmail com on aug at
1
701,144
24,088,097,498
IssuesEvent
2022-09-19 12:44:53
AY2223S1-CS2103T-T09-4/tp
https://api.github.com/repos/AY2223S1-CS2103T-T09-4/tp
closed
Add student and Next of Kin phone number fields
priority.HIGH type.Task type.Task.Add
Phone number will be of 8 digits in length, starting with 8 or 9.
1.0
Add student and Next of Kin phone number fields - Phone number will be of 8 digits in length, starting with 8 or 9.
non_defect
add student and next of kin phone number fields phone number will be of digits in length starting with or
0
60,365
17,023,406,747
IssuesEvent
2021-07-03 01:52:05
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Potlatch not saving comments from offline save
Component: potlatch (flash editor) Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 6.15pm, Saturday, 23rd May 2009]** When doing a save in offline mode, Potlatch prompts for a comment, but this doesn't get used as the comment on the changeset (and the changeset doesn't get closed out, but perhaps that is expected behaviour).
1.0
Potlatch not saving comments from offline save - **[Submitted to the original trac issue database at 6.15pm, Saturday, 23rd May 2009]** When doing a save in offline mode, Potlatch prompts for a comment, but this doesn't get used as the comment on the changeset (and the changeset doesn't get closed out, but perhaps that is expected behaviour).
defect
potlatch not saving comments from offline save when doing a save in offline mode potlatch prompts for a comment but this doesn t get used as the comment on the changeset and the changeset doesn t get closed out but perhaps that is expected behaviour
1
596,958
18,151,489,119
IssuesEvent
2021-09-26 10:44:22
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
opened
people-and-planet-ai.image-classification.e2e_test: test_predict failed
priority: p1 type: bug flakybot: issue
Note: #6463 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 35873ab69eaf0f2a6dd4a7007a12917265e2b848 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8), [Sponge](http://sponge2/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 67, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1632652201.879730366","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Deadline Exceeded","grpc_status":4}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/people-and-planet-ai/image-classification/e2e_test.py", line 112, in model_endpoint_id PROJECT, REGION, MODEL_PATH, MODEL_ENDPOINT, endpoint_id File "/workspace/people-and-planet-ai/image-classification/deploy_model.py", line 80, in deploy_model deployed_model = response.result() File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 130, in result self._blocking_poll(timeout=timeout, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 108, in _blocking_poll retry_(self._done_or_raise)(**kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func on_error=on_error, File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target return target() File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 86, in _done_or_raise if not self.done(**kwargs): File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 170, in done self._refresh_and_update(retry) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 158, in _refresh_and_update self._operation = self._refresh(retry=retry) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operations_v1/operations_client.py", line 143, in get_operation request, retry=retry, timeout=timeout, metadata=metadata File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func on_error=on_error, File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target return target() File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout return func(*args, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 69, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded</pre></details>
1.0
people-and-planet-ai.image-classification.e2e_test: test_predict failed - Note: #6463 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 35873ab69eaf0f2a6dd4a7007a12917265e2b848 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8), [Sponge](http://sponge2/7932aaf3-ea85-4f33-b0c9-2ec91b2682f8) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 67, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "{"created":"@1632652201.879730366","description":"Error received from peer ipv4:74.125.142.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Deadline Exceeded","grpc_status":4}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/people-and-planet-ai/image-classification/e2e_test.py", line 112, in model_endpoint_id PROJECT, REGION, MODEL_PATH, MODEL_ENDPOINT, endpoint_id File "/workspace/people-and-planet-ai/image-classification/deploy_model.py", line 80, in deploy_model deployed_model = response.result() File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 130, in result self._blocking_poll(timeout=timeout, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 108, in _blocking_poll retry_(self._done_or_raise)(**kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func on_error=on_error, File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target return target() File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/future/polling.py", line 86, in _done_or_raise if not self.done(**kwargs): File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 170, in done self._refresh_and_update(retry) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operation.py", line 158, in _refresh_and_update self._operation = self._refresh(retry=retry) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/operations_v1/operations_client.py", line 143, in get_operation request, retry=retry, timeout=timeout, metadata=metadata File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 291, in retry_wrapped_func on_error=on_error, File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 189, in retry_target return target() File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 214, in func_with_timeout return func(*args, **kwargs) File "/workspace/people-and-planet-ai/image-classification/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 69, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded</pre></details>
non_defect
people and planet ai image classification test test predict failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace people and planet ai image classification nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace people and planet ai image classification nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace people and planet ai image classification nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string created description error received from peer file src core lib surface call cc file line grpc message deadline exceeded grpc status the above exception was the direct cause of the following exception traceback most recent call last file workspace people and planet ai image classification test py line in model endpoint id project region model path model endpoint endpoint id file workspace people and planet ai image classification deploy model py line in deploy model deployed model response result file workspace people and planet ai image classification nox py lib site packages google api core future polling py line in result self blocking poll timeout timeout kwargs file workspace people and planet ai image classification nox py lib site packages google api core future polling py line in blocking poll retry self done or raise kwargs file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry target return target file workspace people and planet ai image classification nox py lib site packages google api core future polling py line in done or raise if not self done kwargs file workspace people and planet ai image classification nox py lib site packages google api core operation py line in done self refresh and update retry file workspace people and planet ai image classification nox py lib site packages google api core operation py line in refresh and update self operation self refresh retry retry file workspace people and planet ai image classification nox py lib site packages google api core operations operations client py line in get operation request retry retry timeout timeout metadata metadata file workspace people and planet ai image classification nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace people and planet ai image classification nox py lib site packages google api core retry py line in retry target return target file workspace people and planet ai image classification nox py lib site packages google api core timeout py line in func with timeout return func args kwargs file workspace people and planet ai image classification nox py lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions deadlineexceeded deadline exceeded
0
23,063
3,756,128,559
IssuesEvent
2016-03-13 04:20:35
StarsOCV/eve-tf2hud
https://api.github.com/repos/StarsOCV/eve-tf2hud
closed
I broke it didnt I. TF2 Update.
auto-migrated OpSys-All OpSys-Windows Priority-Medium Type-Defect update Usability
``` Okay so here it is. Tf2 update. It updates i try to start game, i get disk write error and it doesnt start. I try opening up eve hud updater to uninstall the hud to see if that's the problem and i get this. See attached picture. ``` Original issue reported on code.google.com by `byron.ar...@gmail.com` on 20 Dec 2013 at 6:38 Attachments: * [i broke it.png](https://storage.googleapis.com/google-code-attachments/eve-tf2hud/issue-101/comment-0/i broke it.png)
1.0
I broke it didnt I. TF2 Update. - ``` Okay so here it is. Tf2 update. It updates i try to start game, i get disk write error and it doesnt start. I try opening up eve hud updater to uninstall the hud to see if that's the problem and i get this. See attached picture. ``` Original issue reported on code.google.com by `byron.ar...@gmail.com` on 20 Dec 2013 at 6:38 Attachments: * [i broke it.png](https://storage.googleapis.com/google-code-attachments/eve-tf2hud/issue-101/comment-0/i broke it.png)
defect
i broke it didnt i update okay so here it is update it updates i try to start game i get disk write error and it doesnt start i try opening up eve hud updater to uninstall the hud to see if that s the problem and i get this see attached picture original issue reported on code google com by byron ar gmail com on dec at attachments broke it png
1
46,286
13,055,885,660
IssuesEvent
2020-07-30 03:01:25
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting (Trac #926)
Incomplete Migration Migrated from Trac combo reconstruction defect
Migrated from https://code.icecube.wisc.edu/ticket/926 ```json { "status": "closed", "changetime": "2015-05-12T14:06:42", "description": "there is no check of the particle-shape in steamshovel/trunk/private/shovelart/artists/MCTreeArtist.cpp\n\nneeds be fixed, so that neutrinos are actually plotted as incoming particles!", "reporter": "mzoll", "cc": "", "resolution": "fixed", "_ts": "1431439602511916", "component": "combo reconstruction", "summary": "Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting", "priority": "normal", "keywords": "", "time": "2015-04-10T16:15:03", "milestone": "", "owner": "hdembinski", "type": "defect" } ```
1.0
Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting (Trac #926) - Migrated from https://code.icecube.wisc.edu/ticket/926 ```json { "status": "closed", "changetime": "2015-05-12T14:06:42", "description": "there is no check of the particle-shape in steamshovel/trunk/private/shovelart/artists/MCTreeArtist.cpp\n\nneeds be fixed, so that neutrinos are actually plotted as incoming particles!", "reporter": "mzoll", "cc": "", "resolution": "fixed", "_ts": "1431439602511916", "component": "combo reconstruction", "summary": "Steamshovel::I3MCTreeArtist wronly plots Stopping-Tracks as starting", "priority": "normal", "keywords": "", "time": "2015-04-10T16:15:03", "milestone": "", "owner": "hdembinski", "type": "defect" } ```
defect
steamshovel wronly plots stopping tracks as starting trac migrated from json status closed changetime description there is no check of the particle shape in steamshovel trunk private shovelart artists mctreeartist cpp n nneeds be fixed so that neutrinos are actually plotted as incoming particles reporter mzoll cc resolution fixed ts component combo reconstruction summary steamshovel wronly plots stopping tracks as starting priority normal keywords time milestone owner hdembinski type defect
1
61,380
17,023,680,032
IssuesEvent
2021-07-03 03:15:59
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
addr:street is ignored
Component: nominatim Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 9.25am, Monday, 7th February 2011]** I perform search: For "18, . , " (sorry for Russian text). The result is http://goo.gl/LB3yG It returns wrong building: http://www.openstreetmap.org/browse/way/38535341 addr:housenumber: 18 addr:street: . building: yes Expected result is: http://www.openstreetmap.org/browse/way/38323699 addr:housenumber: 18 addr:street: . building: yes It seems the nearest calculation is used instead of addr:street.
1.0
addr:street is ignored - **[Submitted to the original trac issue database at 9.25am, Monday, 7th February 2011]** I perform search: For "18, . , " (sorry for Russian text). The result is http://goo.gl/LB3yG It returns wrong building: http://www.openstreetmap.org/browse/way/38535341 addr:housenumber: 18 addr:street: . building: yes Expected result is: http://www.openstreetmap.org/browse/way/38323699 addr:housenumber: 18 addr:street: . building: yes It seems the nearest calculation is used instead of addr:street.
defect
addr street is ignored i perform search for sorry for russian text the result is it returns wrong building addr housenumber addr street building yes expected result is addr housenumber addr street building yes it seems the nearest calculation is used instead of addr street
1
4,939
2,891,810,781
IssuesEvent
2015-06-15 08:49:23
LibreCat/Catmandu
https://api.github.com/repos/LibreCat/Catmandu
closed
Extend help command to importers, exporters and fixes
documentation enhancement
The `help` command should also show help about importers, exporter, fixes etc.: catmandu help export JSON catmandu help RDF importer catmandu help add_field Some general help would also be nice, for instance a short introduction into fix language: catmandu help fix Help messages can be generated from module documentation. See https://github.com/LibreCat/Catmandu/tree/dev_help fo a start.
1.0
Extend help command to importers, exporters and fixes - The `help` command should also show help about importers, exporter, fixes etc.: catmandu help export JSON catmandu help RDF importer catmandu help add_field Some general help would also be nice, for instance a short introduction into fix language: catmandu help fix Help messages can be generated from module documentation. See https://github.com/LibreCat/Catmandu/tree/dev_help fo a start.
non_defect
extend help command to importers exporters and fixes the help command should also show help about importers exporter fixes etc catmandu help export json catmandu help rdf importer catmandu help add field some general help would also be nice for instance a short introduction into fix language catmandu help fix help messages can be generated from module documentation see fo a start
0
276,436
20,983,776,922
IssuesEvent
2022-03-28 23:18:10
bounswe/bounswe2022group2
https://api.github.com/repos/bounswe/bounswe2022group2
closed
Identifying Profile Page Requirements under User Requirements
priority-high type-documentation status-completed
## Description - Profile page requirements for both types of users (students and teachers) will be identified under the title of User Requirements as a part of the Functional Requirements for this project. - It will include the details for but not limited to * Profile page creation for students * Profile page creation for teachers * Profile page edit for students * Profile page edit for teachers * Experience tracking * Achievements and badges - This issue is a sub-issue of #16 . It will be linked there to easily access but specific details will be under this issue. - After the initial phase for the profile page requirements is finished, this issue will be closed and reviewing process will continue under the main issue mentioned above. **Deadline:** 13.03.2022 - Sunday - 20:00 **Reviewer:** Will be reviewed under the main issue after all user requirements are specified.
1.0
Identifying Profile Page Requirements under User Requirements - ## Description - Profile page requirements for both types of users (students and teachers) will be identified under the title of User Requirements as a part of the Functional Requirements for this project. - It will include the details for but not limited to * Profile page creation for students * Profile page creation for teachers * Profile page edit for students * Profile page edit for teachers * Experience tracking * Achievements and badges - This issue is a sub-issue of #16 . It will be linked there to easily access but specific details will be under this issue. - After the initial phase for the profile page requirements is finished, this issue will be closed and reviewing process will continue under the main issue mentioned above. **Deadline:** 13.03.2022 - Sunday - 20:00 **Reviewer:** Will be reviewed under the main issue after all user requirements are specified.
non_defect
identifying profile page requirements under user requirements description profile page requirements for both types of users students and teachers will be identified under the title of user requirements as a part of the functional requirements for this project it will include the details for but not limited to profile page creation for students profile page creation for teachers profile page edit for students profile page edit for teachers experience tracking achievements and badges this issue is a sub issue of it will be linked there to easily access but specific details will be under this issue after the initial phase for the profile page requirements is finished this issue will be closed and reviewing process will continue under the main issue mentioned above deadline sunday reviewer will be reviewed under the main issue after all user requirements are specified
0
26,647
4,775,714,828
IssuesEvent
2016-10-27 11:25:00
wheeler-microfluidics/microdrop
https://api.github.com/repos/wheeler-microfluidics/microdrop
opened
Add mechanism to add scheduling constraints when running emit_signal (Trac #68)
defect Incomplete Migration microdrop Migrated from Trac
Migrated from http://microfluidics.utoronto.ca/ticket/68 ```json { "status": "closed", "changetime": "2014-04-17T19:39:01", "description": "This ticket takes over the second issue in ticket #21.\n\nThe idea is to provide a way to ensure plugins are processed in the correct order upon an emit_signal() call. ", "reporter": "cfobel", "cc": "", "resolution": "fixed", "_ts": "1397763541728826", "component": "microdrop", "summary": "Add mechanism to add scheduling constraints when running emit_signal", "priority": "major", "keywords": "", "version": "0.1", "time": "2012-03-14T15:41:14", "milestone": "Microdrop 1.0", "owner": "cfobel", "type": "defect" } ```
1.0
Add mechanism to add scheduling constraints when running emit_signal (Trac #68) - Migrated from http://microfluidics.utoronto.ca/ticket/68 ```json { "status": "closed", "changetime": "2014-04-17T19:39:01", "description": "This ticket takes over the second issue in ticket #21.\n\nThe idea is to provide a way to ensure plugins are processed in the correct order upon an emit_signal() call. ", "reporter": "cfobel", "cc": "", "resolution": "fixed", "_ts": "1397763541728826", "component": "microdrop", "summary": "Add mechanism to add scheduling constraints when running emit_signal", "priority": "major", "keywords": "", "version": "0.1", "time": "2012-03-14T15:41:14", "milestone": "Microdrop 1.0", "owner": "cfobel", "type": "defect" } ```
defect
add mechanism to add scheduling constraints when running emit signal trac migrated from json status closed changetime description this ticket takes over the second issue in ticket n nthe idea is to provide a way to ensure plugins are processed in the correct order upon an emit signal call reporter cfobel cc resolution fixed ts component microdrop summary add mechanism to add scheduling constraints when running emit signal priority major keywords version time milestone microdrop owner cfobel type defect
1
411,273
12,016,238,631
IssuesEvent
2020-04-10 15:40:57
yedhink/covid19-kerala-api
https://api.github.com/repos/yedhink/covid19-kerala-api
closed
Major bug - the pdf has another inconsistency where "on today" in annex1 has prefix
bug high-priority python
Probably changing the regex to just an "in" check should fix this.
1.0
Major bug - the pdf has another inconsistency where "on today" in annex1 has prefix - Probably changing the regex to just an "in" check should fix this.
non_defect
major bug the pdf has another inconsistency where on today in has prefix probably changing the regex to just an in check should fix this
0
60,539
17,023,451,285
IssuesEvent
2021-07-03 02:05:56
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Changeset Not Being Rendered
Component: mapnik Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 4.35pm, Friday, 31st July 2009]** Take a look at this boundary where a forest and national park meet: http://osm.org/go/TwUljNo-- Notice that the boundaries don't line up. This is because the national park is in slightly the wrong place. The national park is this changeset uploaded yesterday: http://www.openstreetmap.org/browse/changeset/1980439 Today I moved the national park into the correct position. The changeset was closed at 31 Jul 00:09: http://www.openstreetmap.org/browse/changeset/1989864 I then marked the tile you are looking at as dirty. It was apparently rendered by Mapnik on 31 Jul 03:21: http://a.tile.openstreetmap.org/12/772/1608.png/status As you can see the data from my new changeset has not been used. On 31 Jul 01:33 I added a new changeset with some trails: http://www.openstreetmap.org/browse/changeset/1990063 This was rendered with trails at 31 Jul 03:13: http://a.tile.openstreetmap.org/13/1567/3318.png/status If data I uploaded at 01:33 was rendered at 3:13, how come data I uploaded at 00:09 has not been rendered at the time of writing this? (03:21)? One clue might be that the trails are new data but the movement of nodes was not. Also JOSM gave me an error of "unexpected end of file" when the changeset was closing, but the changeset is listed in my edits as being closed anyway. It also has all 23573 nodes. I have cleared my browser cache and tried two browsers.
1.0
Changeset Not Being Rendered - **[Submitted to the original trac issue database at 4.35pm, Friday, 31st July 2009]** Take a look at this boundary where a forest and national park meet: http://osm.org/go/TwUljNo-- Notice that the boundaries don't line up. This is because the national park is in slightly the wrong place. The national park is this changeset uploaded yesterday: http://www.openstreetmap.org/browse/changeset/1980439 Today I moved the national park into the correct position. The changeset was closed at 31 Jul 00:09: http://www.openstreetmap.org/browse/changeset/1989864 I then marked the tile you are looking at as dirty. It was apparently rendered by Mapnik on 31 Jul 03:21: http://a.tile.openstreetmap.org/12/772/1608.png/status As you can see the data from my new changeset has not been used. On 31 Jul 01:33 I added a new changeset with some trails: http://www.openstreetmap.org/browse/changeset/1990063 This was rendered with trails at 31 Jul 03:13: http://a.tile.openstreetmap.org/13/1567/3318.png/status If data I uploaded at 01:33 was rendered at 3:13, how come data I uploaded at 00:09 has not been rendered at the time of writing this? (03:21)? One clue might be that the trails are new data but the movement of nodes was not. Also JOSM gave me an error of "unexpected end of file" when the changeset was closing, but the changeset is listed in my edits as being closed anyway. It also has all 23573 nodes. I have cleared my browser cache and tried two browsers.
defect
changeset not being rendered take a look at this boundary where a forest and national park meet notice that the boundaries don t line up this is because the national park is in slightly the wrong place the national park is this changeset uploaded yesterday today i moved the national park into the correct position the changeset was closed at jul i then marked the tile you are looking at as dirty it was apparently rendered by mapnik on jul as you can see the data from my new changeset has not been used on jul i added a new changeset with some trails this was rendered with trails at jul if data i uploaded at was rendered at how come data i uploaded at has not been rendered at the time of writing this one clue might be that the trails are new data but the movement of nodes was not also josm gave me an error of unexpected end of file when the changeset was closing but the changeset is listed in my edits as being closed anyway it also has all nodes i have cleared my browser cache and tried two browsers
1
278,064
21,058,031,895
IssuesEvent
2022-04-01 06:38:51
medajet/ped
https://api.github.com/repos/medajet/ped
opened
"upt" command has incorrect format in UG
severity.Medium type.DocumentationBug
For the updt command, both fields are listed as optional even though at least 1 is mandatory. ![Picture14.png](https://raw.githubusercontent.com/medajet/ped/main/files/bce0f5c7-9c4c-499c-a1ff-c2d946008a39.png) <!--session: 1648792880809-a0d404c0-a5ce-4319-b3e0-6ef579fd4865--> <!--Version: Web v3.4.2-->
1.0
"upt" command has incorrect format in UG - For the updt command, both fields are listed as optional even though at least 1 is mandatory. ![Picture14.png](https://raw.githubusercontent.com/medajet/ped/main/files/bce0f5c7-9c4c-499c-a1ff-c2d946008a39.png) <!--session: 1648792880809-a0d404c0-a5ce-4319-b3e0-6ef579fd4865--> <!--Version: Web v3.4.2-->
non_defect
upt command has incorrect format in ug for the updt command both fields are listed as optional even though at least is mandatory
0
52,837
13,225,122,571
IssuesEvent
2020-08-17 20:32:11
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
Healpix port does not build shared libraries (Trac #438)
Migrated from Trac defect other
I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/438">https://code.icecube.wisc.edu/projects/icecube/ticket/438</a>, reported by gladstoneand owned by dladieu</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T17:59:20", "_ts": "1351706360000000", "description": "I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray. ", "reporter": "gladstone", "cc": "gladstone@icecube.wisc.edu", "resolution": "fixed", "time": "2012-08-15T23:13:51", "component": "other", "summary": "Healpix port does not build shared libraries", "priority": "normal", "keywords": "ports", "milestone": "", "owner": "dladieu", "type": "defect" } ``` </p> </details>
1.0
Healpix port does not build shared libraries (Trac #438) - I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/438">https://code.icecube.wisc.edu/projects/icecube/ticket/438</a>, reported by gladstoneand owned by dladieu</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T17:59:20", "_ts": "1351706360000000", "description": "I'm trying to use the healpix port instead of the healpix-cxx project within icerec, and the port hasn't built properly. I'm working on /net/local in Madison. It's only built static libraries, so I can't use it to link to shared libraries in Icetray. ", "reporter": "gladstone", "cc": "gladstone@icecube.wisc.edu", "resolution": "fixed", "time": "2012-08-15T23:13:51", "component": "other", "summary": "Healpix port does not build shared libraries", "priority": "normal", "keywords": "ports", "milestone": "", "owner": "dladieu", "type": "defect" } ``` </p> </details>
defect
healpix port does not build shared libraries trac i m trying to use the healpix port instead of the healpix cxx project within icerec and the port hasn t built properly i m working on net local in madison it s only built static libraries so i can t use it to link to shared libraries in icetray migrated from json status closed changetime ts description i m trying to use the healpix port instead of the healpix cxx project within icerec and the port hasn t built properly i m working on net local in madison it s only built static libraries so i can t use it to link to shared libraries in icetray reporter gladstone cc gladstone icecube wisc edu resolution fixed time component other summary healpix port does not build shared libraries priority normal keywords ports milestone owner dladieu type defect
1
72,628
24,207,431,154
IssuesEvent
2022-09-25 12:38:23
bcpierce00/unison
https://api.github.com/repos/bcpierce00/unison
opened
Backups locations are mixed for backupcurr and backup, and shared across replicas (central)
defect wontfix impact-low effort-medium
As discussed on theunison-users list in 2022-09, `backupcurr` (which is not really backup, but merge pre-images) and `backup` store data in the same place. However, they have different rules. Related, central storage blurs copies across replicas, but each of these is a per-replica concept. To resolve: 1) Central storage should somehow be per-replica 2) backupcurr and backup data should not collide 3) The manual should advise maintaining this separation.
1.0
Backups locations are mixed for backupcurr and backup, and shared across replicas (central) - As discussed on theunison-users list in 2022-09, `backupcurr` (which is not really backup, but merge pre-images) and `backup` store data in the same place. However, they have different rules. Related, central storage blurs copies across replicas, but each of these is a per-replica concept. To resolve: 1) Central storage should somehow be per-replica 2) backupcurr and backup data should not collide 3) The manual should advise maintaining this separation.
defect
backups locations are mixed for backupcurr and backup and shared across replicas central as discussed on theunison users list in backupcurr which is not really backup but merge pre images and backup store data in the same place however they have different rules related central storage blurs copies across replicas but each of these is a per replica concept to resolve central storage should somehow be per replica backupcurr and backup data should not collide the manual should advise maintaining this separation
1
54,996
14,117,132,710
IssuesEvent
2020-11-08 07:32:49
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
Oops, Unable to handle kernel paging request in scan_prefetch_queue_compare
Status: Triage Needed Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information Type | Version/Name --- | --- Distribution Name | Ubuntu (armbian) Distribution Version | Focal Linux Kernel | 5.8.17 Architecture | amd64 ZFS Version | 2.0.0-rc5 SPL Version | 2.0.0-rc5 ### Describe the problem you're observing zpool + zfs hang, seemingly uninterruptible, high load, Oops in dmesg ### Describe how to reproduce the problem `zpool import tank0` ### Include any warning/errors/backtraces from the system logs ``` [ 506.332804] Unable to handle kernel paging request at virtual address fffe0000c4f5f316 [ 506.333567] Mem abort info: [ 506.333815] ESR = 0x96000004 [ 506.334107] EC = 0x25: DABT (current EL), IL = 32 bits [ 506.334581] SET = 0, FnV = 0 [ 506.334921] EA = 0, S1PTW = 0 [ 506.335198] Data abort info: [ 506.335476] ISV = 0, ISS = 0x00000004 [ 506.335813] CM = 0, WnR = 0 [ 506.336076] [fffe0000c4f5f316] address between user and kernel address ranges [ 506.336702] Internal error: Oops: 96000004 [#1] PREEMPT SMP [ 506.337190] Modules linked in: xt_conntrack xt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo xt_addrtype iptable_filter iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 bpfilter br_netfilter bridge rfkill governor_performance zfs(POE) zunicode(POE) zzstd(OE) zlua(OE) zcommon(POE) znvpair(POE) zavl(POE) icp(POE) spl(OE) r8152 snd_soc_hdmi_codec leds_pwm pwm_fan gpio_charger panfrost gpu_sched snd_soc_rockchip_i2s snd_soc_core snd_pcm_dmaengine snd_pcm snd_timer hantro_vpu(C) snd videobuf2_vmalloc rockchip_vdec(C) fusb30x(C) rockchip_rga v4l2_h264 videobuf2_dma_contig soundcore v4l2_mem2mem videobuf2_dma_sg videobuf2_memops videobuf2_v4l2 videobuf2_common videodev mc zstd sg gpio_beeper cpufreq_dt zram sch_fq_codel lm75 ip_tables x_tables autofs4 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx raid1 raid0 multipath linear md_mod realtek rockchipdrm analogix_dp dw_hdmi dw_mipi_dsi drm_kms_helper cec rc_core dwmac_rk stmmac_platform drm [ 506.337260] stmmac mdio_xpcs drm_panel_orientation_quirks adc_keys [ 506.345397] CPU: 5 PID: 3192 Comm: z_rd_int Tainted: P C OE 5.8.17-rockchip64 #20.08.21 [ 506.346186] Hardware name: Helios64 (DT) [ 506.346533] pstate: 80000005 (Nzcv daif -PAN -UAO BTYPE=--) [ 506.347191] pc : scan_prefetch_queue_compare+0x20/0x40 [zfs] [ 506.347694] lr : avl_find+0x68/0xc8 [zavl] [ 506.348054] sp : ffff80001546baa0 [ 506.348347] x29: ffff80001546baa0 x28: ffff0000c207afa8 [ 506.348814] x27: ffff0000c316c9a8 x26: ffff0000c316c9c8 [ 506.349280] x25: ffff0000c30e5090 x24: 0000000000000000 [ 506.349746] x23: ffff80001546bb60 x22: ffff8000092b10d8 [ 506.350212] x21: ffff0000be677000 x20: ffff0000be57f200 [ 506.350677] x19: ffff0000be57f200 x18: 001c050ea761f9c0 [ 506.351143] x17: 00000445581d3e8e x16: 00001510366b591d [ 506.351609] x15: 0000000000000000 x14: 0000000000000000 [ 506.352074] x13: 0000000000000008 x12: ffff0000f6fa9328 [ 506.352541] x11: 0000000000000020 x10: 0000000000000000 [ 506.353006] x9 : 000000000001e067 x8 : 00000000003c0ce0 [ 506.353472] x7 : 00000000003e2880 x6 : 0000000080000000 [ 506.353938] x5 : ffff0000be57f2a0 x4 : ffff0000be6770a0 [ 506.354404] x3 : 0000000000000011 x2 : ffff0000c4f5f300 [ 506.354870] x1 : ffff0000be57f200 x0 : ffff0000c4f5f300 [ 506.355337] Call trace: [ 506.355661] scan_prefetch_queue_compare+0x20/0x40 [zfs] [ 506.356132] avl_find+0x68/0xc8 [zavl] [ 506.356553] dsl_scan_prefetch+0x1a8/0x228 [zfs] [ 506.357042] dsl_scan_prefetch_cb+0x108/0x268 [zfs] [ 506.357553] arc_read_done+0x20c/0x3f8 [zfs] [ 506.358011] zio_done+0x254/0xd40 [zfs] [ 506.358433] zio_execute+0xac/0x110 [zfs] [ 506.358798] taskq_thread+0x298/0x440 [spl] [ 506.359175] kthread+0x118/0x150 [ 506.359462] ret_from_fork+0x10/0x34 [ 506.359782] Code: f9400c22 91028025 f9400c00 39405043 (79402c42) [ 506.360318] ---[ end trace 5477bbfb085279e5 ]--- ```
1.0
Oops, Unable to handle kernel paging request in scan_prefetch_queue_compare - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information Type | Version/Name --- | --- Distribution Name | Ubuntu (armbian) Distribution Version | Focal Linux Kernel | 5.8.17 Architecture | amd64 ZFS Version | 2.0.0-rc5 SPL Version | 2.0.0-rc5 ### Describe the problem you're observing zpool + zfs hang, seemingly uninterruptible, high load, Oops in dmesg ### Describe how to reproduce the problem `zpool import tank0` ### Include any warning/errors/backtraces from the system logs ``` [ 506.332804] Unable to handle kernel paging request at virtual address fffe0000c4f5f316 [ 506.333567] Mem abort info: [ 506.333815] ESR = 0x96000004 [ 506.334107] EC = 0x25: DABT (current EL), IL = 32 bits [ 506.334581] SET = 0, FnV = 0 [ 506.334921] EA = 0, S1PTW = 0 [ 506.335198] Data abort info: [ 506.335476] ISV = 0, ISS = 0x00000004 [ 506.335813] CM = 0, WnR = 0 [ 506.336076] [fffe0000c4f5f316] address between user and kernel address ranges [ 506.336702] Internal error: Oops: 96000004 [#1] PREEMPT SMP [ 506.337190] Modules linked in: xt_conntrack xt_MASQUERADE nf_conntrack_netlink nfnetlink xfrm_user xfrm_algo xt_addrtype iptable_filter iptable_nat nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 bpfilter br_netfilter bridge rfkill governor_performance zfs(POE) zunicode(POE) zzstd(OE) zlua(OE) zcommon(POE) znvpair(POE) zavl(POE) icp(POE) spl(OE) r8152 snd_soc_hdmi_codec leds_pwm pwm_fan gpio_charger panfrost gpu_sched snd_soc_rockchip_i2s snd_soc_core snd_pcm_dmaengine snd_pcm snd_timer hantro_vpu(C) snd videobuf2_vmalloc rockchip_vdec(C) fusb30x(C) rockchip_rga v4l2_h264 videobuf2_dma_contig soundcore v4l2_mem2mem videobuf2_dma_sg videobuf2_memops videobuf2_v4l2 videobuf2_common videodev mc zstd sg gpio_beeper cpufreq_dt zram sch_fq_codel lm75 ip_tables x_tables autofs4 raid10 raid456 async_raid6_recov async_memcpy async_pq async_xor async_tx raid1 raid0 multipath linear md_mod realtek rockchipdrm analogix_dp dw_hdmi dw_mipi_dsi drm_kms_helper cec rc_core dwmac_rk stmmac_platform drm [ 506.337260] stmmac mdio_xpcs drm_panel_orientation_quirks adc_keys [ 506.345397] CPU: 5 PID: 3192 Comm: z_rd_int Tainted: P C OE 5.8.17-rockchip64 #20.08.21 [ 506.346186] Hardware name: Helios64 (DT) [ 506.346533] pstate: 80000005 (Nzcv daif -PAN -UAO BTYPE=--) [ 506.347191] pc : scan_prefetch_queue_compare+0x20/0x40 [zfs] [ 506.347694] lr : avl_find+0x68/0xc8 [zavl] [ 506.348054] sp : ffff80001546baa0 [ 506.348347] x29: ffff80001546baa0 x28: ffff0000c207afa8 [ 506.348814] x27: ffff0000c316c9a8 x26: ffff0000c316c9c8 [ 506.349280] x25: ffff0000c30e5090 x24: 0000000000000000 [ 506.349746] x23: ffff80001546bb60 x22: ffff8000092b10d8 [ 506.350212] x21: ffff0000be677000 x20: ffff0000be57f200 [ 506.350677] x19: ffff0000be57f200 x18: 001c050ea761f9c0 [ 506.351143] x17: 00000445581d3e8e x16: 00001510366b591d [ 506.351609] x15: 0000000000000000 x14: 0000000000000000 [ 506.352074] x13: 0000000000000008 x12: ffff0000f6fa9328 [ 506.352541] x11: 0000000000000020 x10: 0000000000000000 [ 506.353006] x9 : 000000000001e067 x8 : 00000000003c0ce0 [ 506.353472] x7 : 00000000003e2880 x6 : 0000000080000000 [ 506.353938] x5 : ffff0000be57f2a0 x4 : ffff0000be6770a0 [ 506.354404] x3 : 0000000000000011 x2 : ffff0000c4f5f300 [ 506.354870] x1 : ffff0000be57f200 x0 : ffff0000c4f5f300 [ 506.355337] Call trace: [ 506.355661] scan_prefetch_queue_compare+0x20/0x40 [zfs] [ 506.356132] avl_find+0x68/0xc8 [zavl] [ 506.356553] dsl_scan_prefetch+0x1a8/0x228 [zfs] [ 506.357042] dsl_scan_prefetch_cb+0x108/0x268 [zfs] [ 506.357553] arc_read_done+0x20c/0x3f8 [zfs] [ 506.358011] zio_done+0x254/0xd40 [zfs] [ 506.358433] zio_execute+0xac/0x110 [zfs] [ 506.358798] taskq_thread+0x298/0x440 [spl] [ 506.359175] kthread+0x118/0x150 [ 506.359462] ret_from_fork+0x10/0x34 [ 506.359782] Code: f9400c22 91028025 f9400c00 39405043 (79402c42) [ 506.360318] ---[ end trace 5477bbfb085279e5 ]--- ```
defect
oops unable to handle kernel paging request in scan prefetch queue compare thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name ubuntu armbian distribution version focal linux kernel architecture zfs version spl version describe the problem you re observing zpool zfs hang seemingly uninterruptible high load oops in dmesg describe how to reproduce the problem zpool import include any warning errors backtraces from the system logs unable to handle kernel paging request at virtual address mem abort info esr ec dabt current el il bits set fnv ea data abort info isv iss cm wnr address between user and kernel address ranges internal error oops preempt smp modules linked in xt conntrack xt masquerade nf conntrack netlink nfnetlink xfrm user xfrm algo xt addrtype iptable filter iptable nat nf nat nf conntrack nf defrag nf defrag bpfilter br netfilter bridge rfkill governor performance zfs poe zunicode poe zzstd oe zlua oe zcommon poe znvpair poe zavl poe icp poe spl oe snd soc hdmi codec leds pwm pwm fan gpio charger panfrost gpu sched snd soc rockchip snd soc core snd pcm dmaengine snd pcm snd timer hantro vpu c snd vmalloc rockchip vdec c c rockchip rga dma contig soundcore dma sg memops common videodev mc zstd sg gpio beeper cpufreq dt zram sch fq codel ip tables x tables async recov async memcpy async pq async xor async tx multipath linear md mod realtek rockchipdrm analogix dp dw hdmi dw mipi dsi drm kms helper cec rc core dwmac rk stmmac platform drm stmmac mdio xpcs drm panel orientation quirks adc keys cpu pid comm z rd int tainted p c oe hardware name dt pstate nzcv daif pan uao btype pc scan prefetch queue compare lr avl find sp call trace scan prefetch queue compare avl find dsl scan prefetch dsl scan prefetch cb arc read done zio done zio execute taskq thread kthread ret from fork code
1
17,087
2,974,593,191
IssuesEvent
2015-07-15 02:10:24
Reimashi/jotai
https://api.github.com/repos/Reimashi/jotai
closed
Inclusion of Network Download and Upload Speeds
auto-migrated Priority-Medium Type-Defect wontfix
``` What is the expected output? Download and Upload speeds for each network adapter. What version of the product are you using? On what operating system? 0.5.1 Beta on Windows 7 x64 Please provide any additional information below. This would be a great feature and definitely one that is most likely in demand. Thanks ``` Original issue reported on code.google.com by `sandeep....@gmail.com` on 7 Aug 2012 at 11:55
1.0
Inclusion of Network Download and Upload Speeds - ``` What is the expected output? Download and Upload speeds for each network adapter. What version of the product are you using? On what operating system? 0.5.1 Beta on Windows 7 x64 Please provide any additional information below. This would be a great feature and definitely one that is most likely in demand. Thanks ``` Original issue reported on code.google.com by `sandeep....@gmail.com` on 7 Aug 2012 at 11:55
defect
inclusion of network download and upload speeds what is the expected output download and upload speeds for each network adapter what version of the product are you using on what operating system beta on windows please provide any additional information below this would be a great feature and definitely one that is most likely in demand thanks original issue reported on code google com by sandeep gmail com on aug at
1
22,484
15,217,301,666
IssuesEvent
2021-02-17 16:27:30
airyhq/airy
https://api.github.com/repos/airyhq/airy
opened
Wait for resources to be created on airy create
cli infrastructure
Instead of waiting for services to start with the bash `wait-for-service` function, we need to implement this logic inside the CLI, for the full workflow of the `airy create` command.
1.0
Wait for resources to be created on airy create - Instead of waiting for services to start with the bash `wait-for-service` function, we need to implement this logic inside the CLI, for the full workflow of the `airy create` command.
non_defect
wait for resources to be created on airy create instead of waiting for services to start with the bash wait for service function we need to implement this logic inside the cli for the full workflow of the airy create command
0
104,521
13,095,925,180
IssuesEvent
2020-08-03 14:52:53
dusk-network/plonk
https://api.github.com/repos/dusk-network/plonk
closed
Make composer.public_inputs fn public again with `trace` feature
API-design constraint_system type:question
For some reason, this was removed with #255. It is useful when you test gadgets to have access to this feature in order to avoid making by yourself the `public_inputs` vector as we agreed in the past. Is there any reason why this was removed @kevaundray ? If not, we should probably implement `public_inputs()` for `Prover` which would get the public inputs from the `StandardComposer` and give them to the end user.
1.0
Make composer.public_inputs fn public again with `trace` feature - For some reason, this was removed with #255. It is useful when you test gadgets to have access to this feature in order to avoid making by yourself the `public_inputs` vector as we agreed in the past. Is there any reason why this was removed @kevaundray ? If not, we should probably implement `public_inputs()` for `Prover` which would get the public inputs from the `StandardComposer` and give them to the end user.
non_defect
make composer public inputs fn public again with trace feature for some reason this was removed with it is useful when you test gadgets to have access to this feature in order to avoid making by yourself the public inputs vector as we agreed in the past is there any reason why this was removed kevaundray if not we should probably implement public inputs for prover which would get the public inputs from the standardcomposer and give them to the end user
0
559,958
16,581,872,711
IssuesEvent
2021-05-31 12:58:10
gardener/landscaper
https://api.github.com/repos/gardener/landscaper
closed
Use new EnvTest Binaries Setup Tool
area/dev-productivity kind/enhancement priority/4
**How to categorize this issue?** <!-- Please select area, kind, and priority for this issue. This helps the community categorizing it. Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion. If multiple identifiers make sense you can also state the commands multiple times, e.g. /area control-plane /area auto-scaling ... "/area" identifiers: backup|certification|cost|delivery|deployers|manifest-deployer|helm-deployer|container-deployer|dev-productivity|documentation|high-availability|logging|monitoring|oci|open-source|operations|ops-productivity|performance|quality|robustness|scalability|security|storage|testing|usability|user-management "/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test "/priority" identifiers (numerical value): 1 (blocker)|2 (critical)|3 (normal)|4 (low priority)|5 (nice to have) --> /area dev-productivity /kind enhancement /priority 4 **Description**: Currently we install the test tools (kubebuilder apiserver + etcd) directly from the kubebuilder bucket using curl. The controller-runtime has introduced a new tool manage these binaries also to have multiple versions available. So lets switch our install requirements script to that new tool. https://github.com/kubernetes-sigs/controller-runtime/pull/1488
1.0
Use new EnvTest Binaries Setup Tool - **How to categorize this issue?** <!-- Please select area, kind, and priority for this issue. This helps the community categorizing it. Replace below TODOs or exchange the existing identifiers with those that fit best in your opinion. If multiple identifiers make sense you can also state the commands multiple times, e.g. /area control-plane /area auto-scaling ... "/area" identifiers: backup|certification|cost|delivery|deployers|manifest-deployer|helm-deployer|container-deployer|dev-productivity|documentation|high-availability|logging|monitoring|oci|open-source|operations|ops-productivity|performance|quality|robustness|scalability|security|storage|testing|usability|user-management "/kind" identifiers: api-change|bug|cleanup|discussion|enhancement|epic|impediment|poc|post-mortem|question|regression|task|technical-debt|test "/priority" identifiers (numerical value): 1 (blocker)|2 (critical)|3 (normal)|4 (low priority)|5 (nice to have) --> /area dev-productivity /kind enhancement /priority 4 **Description**: Currently we install the test tools (kubebuilder apiserver + etcd) directly from the kubebuilder bucket using curl. The controller-runtime has introduced a new tool manage these binaries also to have multiple versions available. So lets switch our install requirements script to that new tool. https://github.com/kubernetes-sigs/controller-runtime/pull/1488
non_defect
use new envtest binaries setup tool how to categorize this issue please select area kind and priority for this issue this helps the community categorizing it replace below todos or exchange the existing identifiers with those that fit best in your opinion if multiple identifiers make sense you can also state the commands multiple times e g area control plane area auto scaling area identifiers backup certification cost delivery deployers manifest deployer helm deployer container deployer dev productivity documentation high availability logging monitoring oci open source operations ops productivity performance quality robustness scalability security storage testing usability user management kind identifiers api change bug cleanup discussion enhancement epic impediment poc post mortem question regression task technical debt test priority identifiers numerical value blocker critical normal low priority nice to have area dev productivity kind enhancement priority description currently we install the test tools kubebuilder apiserver etcd directly from the kubebuilder bucket using curl the controller runtime has introduced a new tool manage these binaries also to have multiple versions available so lets switch our install requirements script to that new tool
0
710,477
24,419,968,151
IssuesEvent
2022-10-05 19:22:44
bireme/fi-admin
https://api.github.com/repos/bireme/fi-admin
reopened
Pesquisar por novos índices
enhancement priority 2 tesauro
**Novos índices** em "Campos": 1- "TR - Descritores relacionados" (mostra todas as ocorrências em que o string pesquisado ocorre no campo "descritores relacionados". Essa funcionalidade existe no sistema antigo http://decsweb.bireme.br) ![image](https://user-images.githubusercontent.com/25855735/113926659-d51f9b00-97c2-11eb-8de5-58cc817956f2.png) 2- "Termos em notas de escopo e de indexação" (no primeiro resultado, em lista, deve mostrar "Nota de indexação | [o descritor contendo a nota] | [a nota]" ou "Nota de escopo | [o descritor contendo a nota] | [a nota]") O segundo resultado é aquele onde o usuário seleciona a linha desejada e clica para abrir o registro. Os mais importantes são esses dois primeiros novos índices. 3- "Termos em notas históricas" (no primeiro resultado, em lista, deve mostrar "Nota histórica | [o descritor contendo essa nota] | [a nota]") (obs.: importante para podermos conhecer a história do termo da base. Hoje só conseguimos encontrar esses termos usando o webservice trigramas e pesquisando com o índice **DeCS em português com escopo**, pois esse usa o campo histórico do DeCS de consulta, que inclui esses termos. Por exemplo, hoje percebi que o termo Aparelhos Sanitários (no plural) aparecia na última versão do DeCS no [decsweb](http://decsweb.bireme.br/) e no DeCS 2020 de consulta usando o webservice [trigramas](http://trigramas.bireme.br/cgi-bin/mx/cgi=@1?collection=DeCSp&lang=p&minsim=0.30&maxrel=10&text=Aparelho+Sanit%E1rio&resubmit=DeCSxp&x=41&y=5) mas não foi para a nota histórica de nenhum termo no descritor Aparelho Sanitário. Se houvesse esse índice seria mais fácil, além do que não sabemos se o serviço de trigramas vai continuar a existir muito mais tempo. Acabo de recriar o termo no plural. Hoje, 7/4/2021, editei esse issue e ao tentar pesquisar qq termo desse descritor no sistema de consulta atual, o sistema não conseguiu encontrar o registro. Parece que ele estava ou está com algum problema. Favor verificar. 4- "Termos ativos, históricos e em notas" (obs.: históricos aqui inclui status histórico e em notas históricas) (agendar reunião para confirmar de forma consensual se este deveria ser o novo default). **Outros índices novos em termos ativos**: 5- "Descritores MeSH (ID do descritor iniciado com D0)", (obs.: pesquisa com string em qq idioma) 6- "Descritores exclusivos (ID do descritor iniciado com DD)", (obs.: pesquisa com string em qq idioma) 7- "Termos MeSH" (ID do termo iniciado com T) (obs.: pesquisa com string só em inglês, é claro) 8- "Termos exclusivos" (ID do termo iniciado com eng, spa, por, fre)
1.0
Pesquisar por novos índices - **Novos índices** em "Campos": 1- "TR - Descritores relacionados" (mostra todas as ocorrências em que o string pesquisado ocorre no campo "descritores relacionados". Essa funcionalidade existe no sistema antigo http://decsweb.bireme.br) ![image](https://user-images.githubusercontent.com/25855735/113926659-d51f9b00-97c2-11eb-8de5-58cc817956f2.png) 2- "Termos em notas de escopo e de indexação" (no primeiro resultado, em lista, deve mostrar "Nota de indexação | [o descritor contendo a nota] | [a nota]" ou "Nota de escopo | [o descritor contendo a nota] | [a nota]") O segundo resultado é aquele onde o usuário seleciona a linha desejada e clica para abrir o registro. Os mais importantes são esses dois primeiros novos índices. 3- "Termos em notas históricas" (no primeiro resultado, em lista, deve mostrar "Nota histórica | [o descritor contendo essa nota] | [a nota]") (obs.: importante para podermos conhecer a história do termo da base. Hoje só conseguimos encontrar esses termos usando o webservice trigramas e pesquisando com o índice **DeCS em português com escopo**, pois esse usa o campo histórico do DeCS de consulta, que inclui esses termos. Por exemplo, hoje percebi que o termo Aparelhos Sanitários (no plural) aparecia na última versão do DeCS no [decsweb](http://decsweb.bireme.br/) e no DeCS 2020 de consulta usando o webservice [trigramas](http://trigramas.bireme.br/cgi-bin/mx/cgi=@1?collection=DeCSp&lang=p&minsim=0.30&maxrel=10&text=Aparelho+Sanit%E1rio&resubmit=DeCSxp&x=41&y=5) mas não foi para a nota histórica de nenhum termo no descritor Aparelho Sanitário. Se houvesse esse índice seria mais fácil, além do que não sabemos se o serviço de trigramas vai continuar a existir muito mais tempo. Acabo de recriar o termo no plural. Hoje, 7/4/2021, editei esse issue e ao tentar pesquisar qq termo desse descritor no sistema de consulta atual, o sistema não conseguiu encontrar o registro. Parece que ele estava ou está com algum problema. Favor verificar. 4- "Termos ativos, históricos e em notas" (obs.: históricos aqui inclui status histórico e em notas históricas) (agendar reunião para confirmar de forma consensual se este deveria ser o novo default). **Outros índices novos em termos ativos**: 5- "Descritores MeSH (ID do descritor iniciado com D0)", (obs.: pesquisa com string em qq idioma) 6- "Descritores exclusivos (ID do descritor iniciado com DD)", (obs.: pesquisa com string em qq idioma) 7- "Termos MeSH" (ID do termo iniciado com T) (obs.: pesquisa com string só em inglês, é claro) 8- "Termos exclusivos" (ID do termo iniciado com eng, spa, por, fre)
non_defect
pesquisar por novos índices novos índices em campos tr descritores relacionados mostra todas as ocorrências em que o string pesquisado ocorre no campo descritores relacionados essa funcionalidade existe no sistema antigo termos em notas de escopo e de indexação no primeiro resultado em lista deve mostrar nota de indexação ou nota de escopo o segundo resultado é aquele onde o usuário seleciona a linha desejada e clica para abrir o registro os mais importantes são esses dois primeiros novos índices termos em notas históricas no primeiro resultado em lista deve mostrar nota histórica obs importante para podermos conhecer a história do termo da base hoje só conseguimos encontrar esses termos usando o webservice trigramas e pesquisando com o índice decs em português com escopo pois esse usa o campo histórico do decs de consulta que inclui esses termos por exemplo hoje percebi que o termo aparelhos sanitários no plural aparecia na última versão do decs no e no decs de consulta usando o webservice mas não foi para a nota histórica de nenhum termo no descritor aparelho sanitário se houvesse esse índice seria mais fácil além do que não sabemos se o serviço de trigramas vai continuar a existir muito mais tempo acabo de recriar o termo no plural hoje editei esse issue e ao tentar pesquisar qq termo desse descritor no sistema de consulta atual o sistema não conseguiu encontrar o registro parece que ele estava ou está com algum problema favor verificar termos ativos históricos e em notas obs históricos aqui inclui status histórico e em notas históricas agendar reunião para confirmar de forma consensual se este deveria ser o novo default outros índices novos em termos ativos descritores mesh id do descritor iniciado com obs pesquisa com string em qq idioma descritores exclusivos id do descritor iniciado com dd obs pesquisa com string em qq idioma termos mesh id do termo iniciado com t obs pesquisa com string só em inglês é claro termos exclusivos id do termo iniciado com eng spa por fre
0
58,259
16,453,935,780
IssuesEvent
2021-05-21 09:50:11
Questie/Questie
https://api.github.com/repos/Questie/Questie
closed
"[DND] Tar Pedestal" at lots of places
Type - Defect
<!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description Lots of blacksmith icons on the map, named "dnd tar pedestal", only reinstall fix this for me, but they reappear after a while. I don't have any profession trainers selected with a tick in questie, or in the default tracker ## Screenshots [WoWScrnShot_052121_103040](https://user-images.githubusercontent.com/84509345/119108259-4ee0b080-ba20-11eb-821d-c212c2aea89d.jpg) ## Questie version v6.3.11
1.0
"[DND] Tar Pedestal" at lots of places - <!-- READ THIS FIRST Hello, thanks for taking the time to report a bug! Before you proceed, please verify that you're running the latest version of Questie. The easiest way to do this is via the Twitch client, but you can also download the latest version here: https://www.curseforge.com/wow/addons/questie Questie is one of the most popular Classic WoW addons, with over 22M downloads. However, like almost all WoW addons, it's built and maintained by a team of volunteers. The current Questie team is: * @AeroScripts / Aero#1357 (Discord) * @BreakBB / TheCrux#1702 (Discord) * @drejjmit / Drejjmit#8241 (Discord) * @Dyaxler / Dyaxler#0086 (Discord) * @gogo1951 / Gogo#0298 (Discord) If you'd like to help, please consider making a donation. You can do so here: https://www.paypal.com/cgi-bin/webscr?cmd=_donations&business=aero1861%40gmail%2ecom&lc=CA&item_name=Questie%20Devs&currency_code=USD&bn=PP%2dDonationsBF%3abtn_donate_LG%2egif%3aNonHosted You can also help as a tester, developer or translator, please join the Questie Discord here https://discord.gg/fYcQfv7 --> ## Bug description Lots of blacksmith icons on the map, named "dnd tar pedestal", only reinstall fix this for me, but they reappear after a while. I don't have any profession trainers selected with a tick in questie, or in the default tracker ## Screenshots [WoWScrnShot_052121_103040](https://user-images.githubusercontent.com/84509345/119108259-4ee0b080-ba20-11eb-821d-c212c2aea89d.jpg) ## Questie version v6.3.11
defect
tar pedestal at lots of places read this first hello thanks for taking the time to report a bug before you proceed please verify that you re running the latest version of questie the easiest way to do this is via the twitch client but you can also download the latest version here questie is one of the most popular classic wow addons with over downloads however like almost all wow addons it s built and maintained by a team of volunteers the current questie team is aeroscripts aero discord breakbb thecrux discord drejjmit drejjmit discord dyaxler dyaxler discord gogo discord if you d like to help please consider making a donation you can do so here you can also help as a tester developer or translator please join the questie discord here bug description lots of blacksmith icons on the map named dnd tar pedestal only reinstall fix this for me but they reappear after a while i don t have any profession trainers selected with a tick in questie or in the default tracker screenshots questie version
1
28,085
5,185,297,643
IssuesEvent
2017-01-20 09:57:44
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Oracle 12c support for OFFSET .. FETCH causes ORA-00918 regression on queries with ambiguous column names
C: DB: Oracle C: Functionality P: Urgent R: Fixed T: Defect
jOOQ 3.4 introduced a fix (https://github.com/jOOQ/jOOQ/issues/2335) that works around a limitation in SQL where subqueries are not allowed to produce "ambiguous" column names. jOOQ users may write top-level queries as such: ``` java DSL.using(configuration) .select() .from(BOOK) .join(AUTHOR).on(BOOK.AUTHOR_ID.eq(AUTHOR.ID)) .limit(1) .fetch(); ``` There are ambiguous column names `BOOK.ID` and `AUTHOR.ID`, although from a user perspective, this doesn't matter because the two columns are referenced from a top level select. jOOQ, however, emulates `LIMIT` for Oracle using derived tables and `ROWNUM` filtering. #2335 Renames all columns in the nested select, then applies `ROWNUM` filtering, then renames columns back again to their original names. With support for Oracle 12c `OFFSET .. FETCH` (#2607), there is now a regression as Oracle 12c doesn't support "ambiguous" column names in top level selects that contain this clause: ![image](https://cloud.githubusercontent.com/assets/734593/17094259/6d5c986a-524e-11e6-837a-ac693bde0232.png) We'll have to re-implement the same workaround again
1.0
Oracle 12c support for OFFSET .. FETCH causes ORA-00918 regression on queries with ambiguous column names - jOOQ 3.4 introduced a fix (https://github.com/jOOQ/jOOQ/issues/2335) that works around a limitation in SQL where subqueries are not allowed to produce "ambiguous" column names. jOOQ users may write top-level queries as such: ``` java DSL.using(configuration) .select() .from(BOOK) .join(AUTHOR).on(BOOK.AUTHOR_ID.eq(AUTHOR.ID)) .limit(1) .fetch(); ``` There are ambiguous column names `BOOK.ID` and `AUTHOR.ID`, although from a user perspective, this doesn't matter because the two columns are referenced from a top level select. jOOQ, however, emulates `LIMIT` for Oracle using derived tables and `ROWNUM` filtering. #2335 Renames all columns in the nested select, then applies `ROWNUM` filtering, then renames columns back again to their original names. With support for Oracle 12c `OFFSET .. FETCH` (#2607), there is now a regression as Oracle 12c doesn't support "ambiguous" column names in top level selects that contain this clause: ![image](https://cloud.githubusercontent.com/assets/734593/17094259/6d5c986a-524e-11e6-837a-ac693bde0232.png) We'll have to re-implement the same workaround again
defect
oracle support for offset fetch causes ora regression on queries with ambiguous column names jooq introduced a fix that works around a limitation in sql where subqueries are not allowed to produce ambiguous column names jooq users may write top level queries as such java dsl using configuration select from book join author on book author id eq author id limit fetch there are ambiguous column names book id and author id although from a user perspective this doesn t matter because the two columns are referenced from a top level select jooq however emulates limit for oracle using derived tables and rownum filtering renames all columns in the nested select then applies rownum filtering then renames columns back again to their original names with support for oracle offset fetch there is now a regression as oracle doesn t support ambiguous column names in top level selects that contain this clause we ll have to re implement the same workaround again
1
151,461
23,830,452,223
IssuesEvent
2022-09-05 20:00:10
SynBioHub/synbiohub3
https://api.github.com/repos/SynBioHub/synbiohub3
opened
Redo User Design
Design
- Move off H2 database for user storage to a more standard relational database (mySql most likely) - Add data access layers - Add more tables to describe the connections between classes (refer to UML class diagram schematic) - Integrate with proper authentication layers
1.0
Redo User Design - - Move off H2 database for user storage to a more standard relational database (mySql most likely) - Add data access layers - Add more tables to describe the connections between classes (refer to UML class diagram schematic) - Integrate with proper authentication layers
non_defect
redo user design move off database for user storage to a more standard relational database mysql most likely add data access layers add more tables to describe the connections between classes refer to uml class diagram schematic integrate with proper authentication layers
0
186,428
21,933,228,494
IssuesEvent
2022-05-23 11:39:33
onokatio/blog.katio.net
https://api.github.com/repos/onokatio/blog.katio.net
closed
CVE-2021-23368 (Medium) detected in postcss-7.0.32.tgz
security vulnerability
## CVE-2021-23368 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - css-loader-3.6.0.tgz (Root Library) - :x: **postcss-7.0.32.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/onokatio/blog.katio.net/commit/936580315e62ac99730c0ed7a501c46359f1c0ed">936580315e62ac99730c0ed7a501c46359f1c0ed</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution (postcss): 7.0.36</p> <p>Direct dependency fix Resolution (css-loader): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23368 (Medium) detected in postcss-7.0.32.tgz - ## CVE-2021-23368 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/postcss/package.json</p> <p> Dependency Hierarchy: - css-loader-3.6.0.tgz (Root Library) - :x: **postcss-7.0.32.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/onokatio/blog.katio.net/commit/936580315e62ac99730c0ed7a501c46359f1c0ed">936580315e62ac99730c0ed7a501c46359f1c0ed</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package postcss from 7.0.0 and before 8.2.10 are vulnerable to Regular Expression Denial of Service (ReDoS) during source map parsing. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23368>CVE-2021-23368</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23368</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution (postcss): 7.0.36</p> <p>Direct dependency fix Resolution (css-loader): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in postcss tgz cve medium severity vulnerability vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy css loader tgz root library x postcss tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package postcss from and before are vulnerable to regular expression denial of service redos during source map parsing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss direct dependency fix resolution css loader step up your open source security game with whitesource
0
53,787
13,262,299,331
IssuesEvent
2020-08-20 21:29:52
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[mue] remove `compat` option from muex (Trac #2094)
Migrated from Trac combo reconstruction defect
it causes more confusion than it solves <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2094">https://code.icecube.wisc.edu/projects/icecube/ticket/2094</a>, reported by kjmeagherand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "_ts": "1550067295757382", "description": "it causes more confusion than it solves", "reporter": "kjmeagher", "cc": "desiati", "resolution": "invalid", "time": "2017-09-20T12:55:48", "component": "combo reconstruction", "summary": "[mue] remove `compat` option from muex", "priority": "normal", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
1.0
[mue] remove `compat` option from muex (Trac #2094) - it causes more confusion than it solves <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2094">https://code.icecube.wisc.edu/projects/icecube/ticket/2094</a>, reported by kjmeagherand owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "_ts": "1550067295757382", "description": "it causes more confusion than it solves", "reporter": "kjmeagher", "cc": "desiati", "resolution": "invalid", "time": "2017-09-20T12:55:48", "component": "combo reconstruction", "summary": "[mue] remove `compat` option from muex", "priority": "normal", "keywords": "", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
defect
remove compat option from muex trac it causes more confusion than it solves migrated from json status closed changetime ts description it causes more confusion than it solves reporter kjmeagher cc desiati resolution invalid time component combo reconstruction summary remove compat option from muex priority normal keywords milestone owner kjmeagher type defect
1
5,050
2,610,165,940
IssuesEvent
2015-02-26 18:52:36
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
SFX
auto-migrated Priority-Medium Type-Defect
``` Obi Wan's lock s-foils uses Lukes voice ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 4 May 2011 at 9:41
1.0
SFX - ``` Obi Wan's lock s-foils uses Lukes voice ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 4 May 2011 at 9:41
defect
sfx obi wan s lock s foils uses lukes voice original issue reported on code google com by gmail com on may at
1
282,914
30,889,475,951
IssuesEvent
2023-08-04 02:46:56
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2019-19768 (High) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2019-19768 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel 5.4.0-rc2, there is a use-after-free (read) in the __blk_add_trace function in kernel/trace/blktrace.c (which is used to fill out a blk_io_trace structure and place it in a per-cpu sub-buffer). <p>Publish Date: 2019-12-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19768>CVE-2019-19768</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-19768">https://nvd.nist.gov/vuln/detail/CVE-2019-19768</a></p> <p>Release Date: 2020-06-10</p> <p>Fix Resolution: kernel-doc - 3.10.0-514.76.1,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-327.88.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1;kernel-rt-core - 4.18.0-193.rt13.51;kernel-rt-debug-debuginfo - 4.18.0-193.rt13.51;kernel-abi-whitelists - 3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1;kernel-zfcpdump-modules - 4.18.0-193,4.18.0-147.13.2;kernel-rt-trace-devel - 3.10.0-1127.8.2.rt56.1103;kernel-debug-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;kernel-bootwrapper - 3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1;kernel-rt-debuginfo - 4.18.0-193.rt13.51;kernel-rt-debug-modules - 4.18.0-193.rt13.51;kernel-zfcpdump-devel - 4.18.0-193,4.18.0-147.13.2;perf - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,4.18.0-193,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1;kernel-zfcpdump-modules-extra - 4.18.0-193,4.18.0-147.13.2;kernel-debuginfo - 3.10.0-514.76.1,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1;kernel-debug-devel - 3.10.0-514.76.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-327.88.1,4.18.0-193,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-80.18.1;bpftool - 3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-debug-core - 4.18.0-193.rt13.51;kernel-tools-libs - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;perf-debuginfo - 3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-327.88.1;kernel-cross-headers - 4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-debug-debuginfo - 3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;kernel-debug - 3.10.0-514.76.1,3.10.0-327.88.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-957.54.1,4.18.0-193,4.18.0-193,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2;kernel-devel - 4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-327.88.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2;bpftool-debuginfo - 4.18.0-193,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-80.18.1;kpatch-patch-3_10_0-1062_12_1 - 1-2,1-2;kernel-zfcpdump-core - 4.18.0-147.13.2,4.18.0-193;kernel-debug-core - 4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193;kernel-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;python-perf - 3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-core - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-debug - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-rt-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-debuginfo-common-ppc64 - 3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1062.26.1;python3-perf - 4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2;kernel-tools - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel-debug-modules - 4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-trace-kvm - 3.10.0-1127.8.2.rt56.1103;kernel-rt-debuginfo-common-x86_64 - 4.18.0-193.rt13.51;kernel-tools-libs-devel - 3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-modules - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193;kernel-tools-debuginfo - 3.10.0-1062.26.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-80.18.1,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-693.67.1;kernel-rt-modules - 4.18.0-193.rt13.51;kernel-rt-doc - 3.10.0-1127.8.2.rt56.1103;kernel-rt-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;python-perf-debuginfo - 3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1;kernel-headers - 3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-trace - 3.10.0-1127.8.2.rt56.1103;kernel-debuginfo-common-x86_64 - 3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-514.76.1,4.18.0-193,3.10.0-957.54.1;kernel-rt - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-zfcpdump - 4.18.0-147.13.2,4.18.0-193;kernel-rt-debug-modules-extra - 4.18.0-193.rt13.51;python3-perf-debuginfo - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193;kernel-rt-modules-extra - 4.18.0-193.rt13.51</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19768 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2019-19768 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/blktrace.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel 5.4.0-rc2, there is a use-after-free (read) in the __blk_add_trace function in kernel/trace/blktrace.c (which is used to fill out a blk_io_trace structure and place it in a per-cpu sub-buffer). <p>Publish Date: 2019-12-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19768>CVE-2019-19768</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-19768">https://nvd.nist.gov/vuln/detail/CVE-2019-19768</a></p> <p>Release Date: 2020-06-10</p> <p>Fix Resolution: kernel-doc - 3.10.0-514.76.1,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-327.88.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1;kernel-rt-core - 4.18.0-193.rt13.51;kernel-rt-debug-debuginfo - 4.18.0-193.rt13.51;kernel-abi-whitelists - 3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1;kernel-zfcpdump-modules - 4.18.0-193,4.18.0-147.13.2;kernel-rt-trace-devel - 3.10.0-1127.8.2.rt56.1103;kernel-debug-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;kernel-bootwrapper - 3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1;kernel-rt-debuginfo - 4.18.0-193.rt13.51;kernel-rt-debug-modules - 4.18.0-193.rt13.51;kernel-zfcpdump-devel - 4.18.0-193,4.18.0-147.13.2;perf - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,4.18.0-193,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1;kernel-zfcpdump-modules-extra - 4.18.0-193,4.18.0-147.13.2;kernel-debuginfo - 3.10.0-514.76.1,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1;kernel-debug-devel - 3.10.0-514.76.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-327.88.1,4.18.0-193,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-80.18.1;bpftool - 3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-1062.26.1,3.10.0-1127.8.2,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-debug-core - 4.18.0-193.rt13.51;kernel-tools-libs - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;perf-debuginfo - 3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-327.88.1;kernel-cross-headers - 4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-193,4.18.0-147.13.2;kernel-debug-debuginfo - 3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2;kernel-debug - 3.10.0-514.76.1,3.10.0-327.88.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-957.54.1,4.18.0-193,4.18.0-193,3.10.0-1062.26.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2;kernel-devel - 4.18.0-193,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-957.54.1,4.18.0-147.13.2,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,4.18.0-80.18.1,3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-693.67.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel - 3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-327.88.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-147.13.2,3.10.0-957.54.1,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-193,4.18.0-193,3.10.0-1127.8.2,4.18.0-147.13.2,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-1127.8.2,4.18.0-193,3.10.0-514.76.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2;bpftool-debuginfo - 4.18.0-193,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,4.18.0-80.18.1;kpatch-patch-3_10_0-1062_12_1 - 1-2,1-2;kernel-zfcpdump-core - 4.18.0-147.13.2,4.18.0-193;kernel-debug-core - 4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193;kernel-modules-extra - 4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2;kernel-rt-debug-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;python-perf - 3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-514.76.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-core - 4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-debug - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-rt-devel - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-debuginfo-common-ppc64 - 3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-1062.26.1;python3-perf - 4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2;kernel-tools - 3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-957.54.1,4.18.0-193,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1127.8.2,3.10.0-514.76.1,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-327.88.1,3.10.0-1062.26.1,4.18.0-193,3.10.0-957.54.1,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-1127.8.2;kernel-debug-modules - 4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2;kernel-rt-trace-kvm - 3.10.0-1127.8.2.rt56.1103;kernel-rt-debuginfo-common-x86_64 - 4.18.0-193.rt13.51;kernel-tools-libs-devel - 3.10.0-514.76.1,3.10.0-327.88.1,3.10.0-693.67.1,3.10.0-1062.26.1,3.10.0-1062.26.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-1127.8.2,3.10.0-514.76.1,3.10.0-957.54.1,3.10.0-1062.26.1,3.10.0-957.54.1;kernel-modules - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-193,4.18.0-80.18.1,4.18.0-193,4.18.0-147.13.2,4.18.0-147.13.2,4.18.0-193;kernel-tools-debuginfo - 3.10.0-1062.26.1,4.18.0-193,3.10.0-1127.8.2,4.18.0-80.18.1,3.10.0-327.88.1,4.18.0-147.13.2,3.10.0-1127.8.2,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-693.67.1;kernel-rt-modules - 4.18.0-193.rt13.51;kernel-rt-doc - 3.10.0-1127.8.2.rt56.1103;kernel-rt-kvm - 4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103;python-perf-debuginfo - 3.10.0-693.67.1,3.10.0-1127.8.2,3.10.0-957.54.1,3.10.0-1127.8.2,3.10.0-327.88.1,3.10.0-1062.26.1,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1;kernel-headers - 3.10.0-1062.26.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,4.18.0-193,4.18.0-80.18.1,4.18.0-147.13.2,3.10.0-327.88.1,3.10.0-1127.8.2,4.18.0-147.13.2,4.18.0-193,3.10.0-1062.26.1,3.10.0-693.67.1,4.18.0-193,3.10.0-1127.8.2,3.10.0-693.67.1,4.18.0-147.13.2,3.10.0-957.54.1,3.10.0-514.76.1,3.10.0-1062.26.1,4.18.0-80.18.1,3.10.0-957.54.1,4.18.0-193,3.10.0-1127.8.2;kernel-rt-trace - 3.10.0-1127.8.2.rt56.1103;kernel-debuginfo-common-x86_64 - 3.10.0-1127.8.2,3.10.0-693.67.1,3.10.0-327.88.1,4.18.0-147.13.2,4.18.0-80.18.1,3.10.0-1062.26.1,3.10.0-514.76.1,4.18.0-193,3.10.0-957.54.1;kernel-rt - 3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51,3.10.0-1127.8.2.rt56.1103,4.18.0-193.rt13.51;kernel-zfcpdump - 4.18.0-147.13.2,4.18.0-193;kernel-rt-debug-modules-extra - 4.18.0-193.rt13.51;python3-perf-debuginfo - 4.18.0-147.13.2,4.18.0-80.18.1,4.18.0-193;kernel-rt-modules-extra - 4.18.0-193.rt13.51</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files kernel trace blktrace c kernel trace blktrace c vulnerability details in the linux kernel there is a use after free read in the blk add trace function in kernel trace blktrace c which is used to fill out a blk io trace structure and place it in a per cpu sub buffer publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution kernel doc kernel rt core kernel rt debug debuginfo kernel abi whitelists kernel zfcpdump modules kernel rt trace devel kernel debug modules extra kernel rt debug kvm kernel bootwrapper kernel rt debuginfo kernel rt debug modules kernel zfcpdump devel perf kernel zfcpdump modules extra kernel debuginfo kernel debug devel bpftool kernel rt debug core kernel tools libs perf debuginfo kernel cross headers kernel debug debuginfo kernel debug kernel devel kernel bpftool debuginfo kpatch patch kernel zfcpdump core kernel debug core kernel modules extra kernel rt debug devel python perf kernel core kernel rt debug kernel rt devel kernel debuginfo common perf kernel tools kernel debug modules kernel rt trace kvm kernel rt debuginfo common kernel tools libs devel kernel modules kernel tools debuginfo kernel rt modules kernel rt doc kernel rt kvm python perf debuginfo kernel headers kernel rt trace kernel debuginfo common kernel rt kernel zfcpdump kernel rt debug modules extra perf debuginfo kernel rt modules extra step up your open source security game with mend
0
100,974
21,562,551,057
IssuesEvent
2022-05-01 11:36:46
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.1.x] Cassiopea Registration page Privacy/Terms alignment
New Feature No Code Attached Yet J4 Frontend Template
Hi guys, about the [Cassiopea Registration page Privacy/Terms alignment](https://photos.app.goo.gl/6VZuH7Ja7RQ8yfKt5 "Registration Privacy/Terms"), Should not be better to add by default: .required.radio { display: inline-flex; gap: 1rem; } to align them horizontally and don't waste precious space ?
1.0
[4.1.x] Cassiopea Registration page Privacy/Terms alignment - Hi guys, about the [Cassiopea Registration page Privacy/Terms alignment](https://photos.app.goo.gl/6VZuH7Ja7RQ8yfKt5 "Registration Privacy/Terms"), Should not be better to add by default: .required.radio { display: inline-flex; gap: 1rem; } to align them horizontally and don't waste precious space ?
non_defect
cassiopea registration page privacy terms alignment hi guys about the registration privacy terms should not be better to add by default required radio display inline flex gap to align them horizontally and don t waste precious space
0
75,148
25,557,257,859
IssuesEvent
2022-11-30 07:58:32
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Unread message/mention comes back after closing the browser and re-opening
T-Defect S-Minor A-Notifications O-Occasional
### Steps to reproduce When a message is last sent by a contact with a mention/response after logging in or after closing the client/browser and reopening it, it will still show up as an unread messages even if the message was read before. ### Outcome #### What did you expect? Normal use. #### What happened instead? There is a problem with the display. ### Operating system Ubuntu (Actually all clients have this problem) ### Browser information Edge 107.0 ### URL for webapp _No response_ ### Application version Element Web/Desktop 1.11.15 Element iOS 1.9.12 ### Homeserver Synapse 1.72.0 ### Will you send logs? No
1.0
Unread message/mention comes back after closing the browser and re-opening - ### Steps to reproduce When a message is last sent by a contact with a mention/response after logging in or after closing the client/browser and reopening it, it will still show up as an unread messages even if the message was read before. ### Outcome #### What did you expect? Normal use. #### What happened instead? There is a problem with the display. ### Operating system Ubuntu (Actually all clients have this problem) ### Browser information Edge 107.0 ### URL for webapp _No response_ ### Application version Element Web/Desktop 1.11.15 Element iOS 1.9.12 ### Homeserver Synapse 1.72.0 ### Will you send logs? No
defect
unread message mention comes back after closing the browser and re opening steps to reproduce when a message is last sent by a contact with a mention response after logging in or after closing the client browser and reopening it it will still show up as an unread messages even if the message was read before outcome what did you expect normal use what happened instead there is a problem with the display operating system ubuntu actually all clients have this problem browser information edge url for webapp no response application version element web desktop element ios homeserver synapse will you send logs no
1
10,079
2,618,934,929
IssuesEvent
2015-03-03 00:01:42
chrsmith/open-ig
https://api.github.com/repos/chrsmith/open-ig
closed
pre-hyperdrive bugs
auto-migrated Component-Logic Priority-Medium Type-Defect
``` Game version: 0.95.141 Operating System: xp 32 Java runtime version: jdk1.7.0_04 Installed using the Launcher? yes Game language de 1) Produced destroyers possess a hyperdrive and they have according speed. 2) The cruiser given at the start of level 2 has a hyperdrive and according speed. 3) The destroyer left over from level 1 is shown to have a hyperdrive in fleet equipment screen (box in the middle, actual:1/max:1) but is shown with speed 6 as it should. It looks as if fleets/ships do not always move with the speed they are shown with now since you made the traders "catchable", this confuses me. (Am I faster or is the trader slower than shown ?) The cruiser really moves faster than the leftover destroyer at least. Same for produced destroyers. Attached the start save for level 2 and the last one I still have from before. ``` Original issue reported on code.google.com by `anonymis...@arcor.de` on 27 Aug 2013 at 3:15 Attachments: * [save-2013-08-27-01-32-55-406.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-32-55-406.xml.gz) * [save-2013-08-27-01-55-32-718.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-55-32-718.xml.gz)
1.0
pre-hyperdrive bugs - ``` Game version: 0.95.141 Operating System: xp 32 Java runtime version: jdk1.7.0_04 Installed using the Launcher? yes Game language de 1) Produced destroyers possess a hyperdrive and they have according speed. 2) The cruiser given at the start of level 2 has a hyperdrive and according speed. 3) The destroyer left over from level 1 is shown to have a hyperdrive in fleet equipment screen (box in the middle, actual:1/max:1) but is shown with speed 6 as it should. It looks as if fleets/ships do not always move with the speed they are shown with now since you made the traders "catchable", this confuses me. (Am I faster or is the trader slower than shown ?) The cruiser really moves faster than the leftover destroyer at least. Same for produced destroyers. Attached the start save for level 2 and the last one I still have from before. ``` Original issue reported on code.google.com by `anonymis...@arcor.de` on 27 Aug 2013 at 3:15 Attachments: * [save-2013-08-27-01-32-55-406.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-32-55-406.xml.gz) * [save-2013-08-27-01-55-32-718.xml.gz](https://storage.googleapis.com/google-code-attachments/open-ig/issue-764/comment-0/save-2013-08-27-01-55-32-718.xml.gz)
defect
pre hyperdrive bugs game version operating system xp java runtime version installed using the launcher yes game language de produced destroyers possess a hyperdrive and they have according speed the cruiser given at the start of level has a hyperdrive and according speed the destroyer left over from level is shown to have a hyperdrive in fleet equipment screen box in the middle actual max but is shown with speed as it should it looks as if fleets ships do not always move with the speed they are shown with now since you made the traders catchable this confuses me am i faster or is the trader slower than shown the cruiser really moves faster than the leftover destroyer at least same for produced destroyers attached the start save for level and the last one i still have from before original issue reported on code google com by anonymis arcor de on aug at attachments
1
15,913
2,869,093,867
IssuesEvent
2015-06-05 23:17:04
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Span.getLocationMessage returns incorrect output for lines without trailing newlines
Area-Pkg Pkg-SourceMaps Priority-Unassigned Triaged Type-Defect
If a span points to the final line in a file, and that line doesn't end in a newline, the arrows in getLocationMessage will be rendered on the same line as the source text. For example: &nbsp;&nbsp;&nbsp;&nbsp;Error on line 12, column 28 of pubspec.yaml: Error loading transformer: I hate these settings! &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;- app/src/html_rewriter: {} ^^
1.0
Span.getLocationMessage returns incorrect output for lines without trailing newlines - If a span points to the final line in a file, and that line doesn't end in a newline, the arrows in getLocationMessage will be rendered on the same line as the source text. For example: &nbsp;&nbsp;&nbsp;&nbsp;Error on line 12, column 28 of pubspec.yaml: Error loading transformer: I hate these settings! &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;- app/src/html_rewriter: {} ^^
defect
span getlocationmessage returns incorrect output for lines without trailing newlines if a span points to the final line in a file and that line doesn t end in a newline the arrows in getlocationmessage will be rendered on the same line as the source text for example nbsp nbsp nbsp nbsp error on line column of pubspec yaml error loading transformer i hate these settings nbsp nbsp nbsp nbsp nbsp nbsp app src html rewriter
1
118,302
9,981,595,018
IssuesEvent
2019-07-10 07:54:03
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
reopened
Update 'URI' to 'URL' on the first page of the Connect dialog
:gear: attach :gear: sas 🧪 testing
**Storage Explorer Version:** rel/1.9.0_20190626.4 **Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10 **Architecture:** ia32/x64 **Commit:** 0afd7438 **Regression From:** Not a regression **Steps to reproduce:** 1. Launch storage explorer -> Open Connect dialog. 2. Check the descriptions on the dialog. **Expected Experience:** Show 'Use a shared access signature (SAS) **URL**' **Actual Experience:** Show 'Use a shared access signature (SAS) **URI**' ![image](https://user-images.githubusercontent.com/41351993/60165406-cfe19e00-9831-11e9-893e-ce30e1cf7b5f.png) **More Info:** Show URL when generating share access signature. ![image](https://user-images.githubusercontent.com/41351993/60165549-13d4a300-9832-11e9-84f7-bbde2a163d2c.png)
1.0
Update 'URI' to 'URL' on the first page of the Connect dialog - **Storage Explorer Version:** rel/1.9.0_20190626.4 **Platform/OS:** Linux Ubuntu/macOS High Sierra/Windows 10 **Architecture:** ia32/x64 **Commit:** 0afd7438 **Regression From:** Not a regression **Steps to reproduce:** 1. Launch storage explorer -> Open Connect dialog. 2. Check the descriptions on the dialog. **Expected Experience:** Show 'Use a shared access signature (SAS) **URL**' **Actual Experience:** Show 'Use a shared access signature (SAS) **URI**' ![image](https://user-images.githubusercontent.com/41351993/60165406-cfe19e00-9831-11e9-893e-ce30e1cf7b5f.png) **More Info:** Show URL when generating share access signature. ![image](https://user-images.githubusercontent.com/41351993/60165549-13d4a300-9832-11e9-84f7-bbde2a163d2c.png)
non_defect
update uri to url on the first page of the connect dialog storage explorer version rel platform os linux ubuntu macos high sierra windows architecture commit regression from not a regression steps to reproduce launch storage explorer open connect dialog check the descriptions on the dialog expected experience show use a shared access signature sas url actual experience show use a shared access signature sas uri more info show url when generating share access signature
0
6,732
2,610,274,129
IssuesEvent
2015-02-26 19:27:44
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Window always opens narrow
auto-migrated Priority-Medium Type-Defect
``` What's the problem? every time I click the scribefire icon it opens the scribefire "window" very narrow and tall (unusable) -and I must adjust every single time I open the window again What browser are you using? Firefox - several versions 3.6, 3.6.13 and 4 ... both ob XP, Vista and Windows 7 What version of ScribeFire are you running? How nice it would be if you had an "about" link inside scribefire to make it easier running Scribefire Next 1.4.3.0 (it does not happen in regular Scribefire 3.5.3.3 ``` ----- Original issue reported on code.google.com by `cjcornell2` on 23 Feb 2011 at 9:47
1.0
Window always opens narrow - ``` What's the problem? every time I click the scribefire icon it opens the scribefire "window" very narrow and tall (unusable) -and I must adjust every single time I open the window again What browser are you using? Firefox - several versions 3.6, 3.6.13 and 4 ... both ob XP, Vista and Windows 7 What version of ScribeFire are you running? How nice it would be if you had an "about" link inside scribefire to make it easier running Scribefire Next 1.4.3.0 (it does not happen in regular Scribefire 3.5.3.3 ``` ----- Original issue reported on code.google.com by `cjcornell2` on 23 Feb 2011 at 9:47
defect
window always opens narrow what s the problem every time i click the scribefire icon it opens the scribefire window very narrow and tall unusable and i must adjust every single time i open the window again what browser are you using firefox several versions and both ob xp vista and windows what version of scribefire are you running how nice it would be if you had an about link inside scribefire to make it easier running scribefire next it does not happen in regular scribefire original issue reported on code google com by on feb at
1
16,971
3,587,039,131
IssuesEvent
2016-01-30 02:17:08
dylanaraps/fetch
https://api.github.com/repos/dylanaraps/fetch
closed
GPU support.
Needs Testing
### An experimental version of GPU support is currently in master. Here's the status for GPU support: - [X] Linux - [X] Mac OS X - [X] FreeBSD - [ ] OpenBSD - We could've used `pciutils` to get lspci functionality on OpenBSD but it turns out that it requires root. - [ ] NetBSD - No clue on this one, I haven't been able to find a way to get gpu. - [X] Windows Adding support for OpenBSD and NetBSD is impossible at this point as I can't find a way to get the gpu name without needing root privileges. ### Troubleshooting #### All OS: I'll need to see a screenshot of fetch. #### Linux: I'll need to see the output of this command: ```sh lspci | grep "VGA" ``` If the command above prints something vague instead of your gpu, your pci.ids files may be outdated. See this part of the readme on how to update them. https://github.com/dylanaraps/fetch#getgpu-doesnt-show-my-exact-video-card-name #### Mac OS X: I'll need to see the output of this command: ```sh system_profiler SPDisplaysDataType | awk -F': ' '/^\ *Chipset Model:/ {printf $2}' ``` #### FreeBSD: I'll need to see the output of this command: ```sh pciconf -lv 2>/dev/null | grep -B 4 "VGA" ``` #### Windows: I'll need to see the output of this command: ```sh wmic path Win32_VideoController get caption ```
1.0
GPU support. - ### An experimental version of GPU support is currently in master. Here's the status for GPU support: - [X] Linux - [X] Mac OS X - [X] FreeBSD - [ ] OpenBSD - We could've used `pciutils` to get lspci functionality on OpenBSD but it turns out that it requires root. - [ ] NetBSD - No clue on this one, I haven't been able to find a way to get gpu. - [X] Windows Adding support for OpenBSD and NetBSD is impossible at this point as I can't find a way to get the gpu name without needing root privileges. ### Troubleshooting #### All OS: I'll need to see a screenshot of fetch. #### Linux: I'll need to see the output of this command: ```sh lspci | grep "VGA" ``` If the command above prints something vague instead of your gpu, your pci.ids files may be outdated. See this part of the readme on how to update them. https://github.com/dylanaraps/fetch#getgpu-doesnt-show-my-exact-video-card-name #### Mac OS X: I'll need to see the output of this command: ```sh system_profiler SPDisplaysDataType | awk -F': ' '/^\ *Chipset Model:/ {printf $2}' ``` #### FreeBSD: I'll need to see the output of this command: ```sh pciconf -lv 2>/dev/null | grep -B 4 "VGA" ``` #### Windows: I'll need to see the output of this command: ```sh wmic path Win32_VideoController get caption ```
non_defect
gpu support an experimental version of gpu support is currently in master here s the status for gpu support linux mac os x freebsd openbsd we could ve used pciutils to get lspci functionality on openbsd but it turns out that it requires root netbsd no clue on this one i haven t been able to find a way to get gpu windows adding support for openbsd and netbsd is impossible at this point as i can t find a way to get the gpu name without needing root privileges troubleshooting all os i ll need to see a screenshot of fetch linux i ll need to see the output of this command sh lspci grep vga if the command above prints something vague instead of your gpu your pci ids files may be outdated see this part of the readme on how to update them mac os x i ll need to see the output of this command sh system profiler spdisplaysdatatype awk f chipset model printf freebsd i ll need to see the output of this command sh pciconf lv dev null grep b vga windows i ll need to see the output of this command sh wmic path videocontroller get caption
0
10,115
2,618,937,001
IssuesEvent
2015-03-03 00:02:28
chrsmith/open-ig
https://api.github.com/repos/chrsmith/open-ig
closed
Units not attacks space stations in vinicity
auto-migrated Priority-Medium Type-Defect
``` Game version: 0.95.153 Operating System: Linux x64 Java runtime version: 1.7.0_51 Installed using the Launcher? yes Game language (en, hu, de): hu What steps will reproduce the problem? 1. Attack an enemy planet which has space stations. 2. Move your crusiers or destroyers near a space station, just enough close for the shot range of heavy weapons reach it. What is the expected output? What do you see instead? They should attack the station with the heavy weapons (particle, meson). Instead they let them shot by the station. And when i tell them to attack, they do not shot from they are, they move closer and now the other stations can shot the player units. Please provide any additional information below. Please upload any save before and/or after the problem happened. Please attach the open-ig.log file found in the application's directory. Can give, but any planet attack will do. ``` Original issue reported on code.google.com by `kli...@gmail.com` on 17 Jan 2014 at 11:18 * Merged into: #803
1.0
Units not attacks space stations in vinicity - ``` Game version: 0.95.153 Operating System: Linux x64 Java runtime version: 1.7.0_51 Installed using the Launcher? yes Game language (en, hu, de): hu What steps will reproduce the problem? 1. Attack an enemy planet which has space stations. 2. Move your crusiers or destroyers near a space station, just enough close for the shot range of heavy weapons reach it. What is the expected output? What do you see instead? They should attack the station with the heavy weapons (particle, meson). Instead they let them shot by the station. And when i tell them to attack, they do not shot from they are, they move closer and now the other stations can shot the player units. Please provide any additional information below. Please upload any save before and/or after the problem happened. Please attach the open-ig.log file found in the application's directory. Can give, but any planet attack will do. ``` Original issue reported on code.google.com by `kli...@gmail.com` on 17 Jan 2014 at 11:18 * Merged into: #803
defect
units not attacks space stations in vinicity game version operating system linux java runtime version installed using the launcher yes game language en hu de hu what steps will reproduce the problem attack an enemy planet which has space stations move your crusiers or destroyers near a space station just enough close for the shot range of heavy weapons reach it what is the expected output what do you see instead they should attack the station with the heavy weapons particle meson instead they let them shot by the station and when i tell them to attack they do not shot from they are they move closer and now the other stations can shot the player units please provide any additional information below please upload any save before and or after the problem happened please attach the open ig log file found in the application s directory can give but any planet attack will do original issue reported on code google com by kli gmail com on jan at merged into
1
6,929
3,933,560,306
IssuesEvent
2016-04-25 19:33:39
ghutchis/avogadro
https://api.github.com/repos/ghutchis/avogadro
closed
Require #include <GL/glu.h> on Ubuntu 12.04
auto-migrated Building / Install v_1.1.0
Compiling avogadro 1.0.3 on Ubuntu 12.04 requires the insertion of #include <GL/glu.h> in the following files: libavogadro/src/glpainter_p.cpp libavogadro/src/glwidget.cpp libavogadro/src/tools/selectrotatetool.cpp libavogadro/src/camera.cpp This was not required on Ubuntu 10.04, but I guess the GL headers have changed substantially in that period. Reported by: *anonymous
1.0
Require #include <GL/glu.h> on Ubuntu 12.04 - Compiling avogadro 1.0.3 on Ubuntu 12.04 requires the insertion of #include <GL/glu.h> in the following files: libavogadro/src/glpainter_p.cpp libavogadro/src/glwidget.cpp libavogadro/src/tools/selectrotatetool.cpp libavogadro/src/camera.cpp This was not required on Ubuntu 10.04, but I guess the GL headers have changed substantially in that period. Reported by: *anonymous
non_defect
require include on ubuntu compiling avogadro on ubuntu requires the insertion of include in the following files libavogadro src glpainter p cpp libavogadro src glwidget cpp libavogadro src tools selectrotatetool cpp libavogadro src camera cpp this was not required on ubuntu but i guess the gl headers have changed substantially in that period reported by anonymous
0
9,878
2,616,005,523
IssuesEvent
2015-03-02 00:49:56
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
isIdle and build queue
auto-migrated Priority-Low Type-Defect
``` isIdle does not consider a building that is training a unit. ``` Original issue reported on code.google.com by `AHeinerm` on 29 Jul 2009 at 8:18
1.0
isIdle and build queue - ``` isIdle does not consider a building that is training a unit. ``` Original issue reported on code.google.com by `AHeinerm` on 29 Jul 2009 at 8:18
defect
isidle and build queue isidle does not consider a building that is training a unit original issue reported on code google com by aheinerm on jul at
1
593,518
18,010,100,434
IssuesEvent
2021-09-16 07:34:49
vrchatapi/specification
https://api.github.com/repos/vrchatapi/specification
opened
Missing Select Fallback Avatar endpoint
Priority: Medium Status: Help Wanted Status: Accepted Type: Undocumented Endpoint
PUT `/avatars/<avatarId>/selectfallback` Big :heart: to Natsumi for pointing this out
1.0
Missing Select Fallback Avatar endpoint - PUT `/avatars/<avatarId>/selectfallback` Big :heart: to Natsumi for pointing this out
non_defect
missing select fallback avatar endpoint put avatars selectfallback big heart to natsumi for pointing this out
0
16,754
2,941,602,420
IssuesEvent
2015-07-02 09:04:12
schwabe/ics-openvpn
https://api.github.com/repos/schwabe/ics-openvpn
closed
Add a blacklist of network OpenVPN should not try to connect the VPN.
auto-migrated duplicate Priority-Medium Type-Defect
``` I use OpenVPN to keep my phone connected to my home network while connected to my work WiFi or my friend's Wifi or using my mobile network. ics-openvpn works quite reliably. The problems is when I arrive home: the phone connects to one my home Wifi networks (I have two) and ics-openvpn tries to connect to the VPN but it does not succeed (but, due to the way things are configured it can't connect, so this is ok). The problem is that it keeps trying. I think the solution should be having a blacklist so when connected to my wifi it does not try to connect the VPN. Thanks A. ``` Original issue reported on code.google.com by `besang...@grumulla.net` on 1 Nov 2012 at 3:39
1.0
Add a blacklist of network OpenVPN should not try to connect the VPN. - ``` I use OpenVPN to keep my phone connected to my home network while connected to my work WiFi or my friend's Wifi or using my mobile network. ics-openvpn works quite reliably. The problems is when I arrive home: the phone connects to one my home Wifi networks (I have two) and ics-openvpn tries to connect to the VPN but it does not succeed (but, due to the way things are configured it can't connect, so this is ok). The problem is that it keeps trying. I think the solution should be having a blacklist so when connected to my wifi it does not try to connect the VPN. Thanks A. ``` Original issue reported on code.google.com by `besang...@grumulla.net` on 1 Nov 2012 at 3:39
defect
add a blacklist of network openvpn should not try to connect the vpn i use openvpn to keep my phone connected to my home network while connected to my work wifi or my friend s wifi or using my mobile network ics openvpn works quite reliably the problems is when i arrive home the phone connects to one my home wifi networks i have two and ics openvpn tries to connect to the vpn but it does not succeed but due to the way things are configured it can t connect so this is ok the problem is that it keeps trying i think the solution should be having a blacklist so when connected to my wifi it does not try to connect the vpn thanks a original issue reported on code google com by besang grumulla net on nov at
1
7,795
2,610,636,996
IssuesEvent
2015-02-26 21:33:43
alistairreilly/open-ig
https://api.github.com/repos/alistairreilly/open-ig
closed
UI hangok finomhangolása&elmaradt hangok pótlása
auto-migrated Milestone-0.93.500 Priority-Low Sound-Effects Type-Defect
``` Belső menü: Nem megfelelő hangok: -Hangok-->Gombhangok -Játékmenet-->Épületek automatikus javítása. Ennél a kettőnél click_low_1.wav helyett click_medium_2.wav-ot kellene bejátszani. ``` Original issue reported on code.google.com by `Jozsef.T...@gmail.com` on 25 Aug 2011 at 6:18
1.0
UI hangok finomhangolása&elmaradt hangok pótlása - ``` Belső menü: Nem megfelelő hangok: -Hangok-->Gombhangok -Játékmenet-->Épületek automatikus javítása. Ennél a kettőnél click_low_1.wav helyett click_medium_2.wav-ot kellene bejátszani. ``` Original issue reported on code.google.com by `Jozsef.T...@gmail.com` on 25 Aug 2011 at 6:18
defect
ui hangok finomhangolása elmaradt hangok pótlása belső menü nem megfelelő hangok hangok gombhangok játékmenet épületek automatikus javítása ennél a kettőnél click low wav helyett click medium wav ot kellene bejátszani original issue reported on code google com by jozsef t gmail com on aug at
1
24,034
3,900,663,559
IssuesEvent
2016-04-18 07:29:16
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
simple rms.rmsd() ignores weights option
Component-Analysis defect Difficulty-easy
The `analysis.rms.rmsd` function doesn't weight a calculated rmsd when the weights are given ### Currently version of MDAnalysis: (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.15dev
1.0
simple rms.rmsd() ignores weights option - The `analysis.rms.rmsd` function doesn't weight a calculated rmsd when the weights are given ### Currently version of MDAnalysis: (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.15dev
defect
simple rms rmsd ignores weights option the analysis rms rmsd function doesn t weight a calculated rmsd when the weights are given currently version of mdanalysis run python c import mdanalysis as mda print mda version
1
78,699
3,513,191,475
IssuesEvent
2016-01-11 09:05:52
OCHA-DAP/hdx-ckan
https://api.github.com/repos/OCHA-DAP/hdx-ckan
closed
New Contribute Flow: # Resources
bug Priority-Medium
Currently it is displayed: # Resources (drag and drop to sort) "#" should be updated and display the total number of resources. When adding a new resource, the number will be updated
1.0
New Contribute Flow: # Resources - Currently it is displayed: # Resources (drag and drop to sort) "#" should be updated and display the total number of resources. When adding a new resource, the number will be updated
non_defect
new contribute flow resources currently it is displayed resources drag and drop to sort should be updated and display the total number of resources when adding a new resource the number will be updated
0
57,001
15,591,279,603
IssuesEvent
2021-03-18 10:16:24
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
DataTable: unselect not clearing selection in PFX
10.0.1 defect
**Describe the defect** Testing out PFX. While following the migration guide, I noticed a problem with Datatable when performing a _unselect_. In my datatables, I noticed that _selection_ was not updated even though the rowUnselect event listener was called properly. Also, when I clear the selection from the model in my backing bean and update/re-render the datatable, the row would remain selected. **Environment:** - PF Version: _10.0.0_ - JSF + version: Mojarra 2.2.17 - Affected browsers: _ALL_ **To Reproduce** Steps to reproduce the behavior: 1. Go to the Showcase 2. Navigate to Datatable - Selection 3. Scroll down to 'Single with Row Click' 4. Try to click select a row, and press 'View'. Afterwards, CTRL click the row, thereby unselecting it. When pressing 'View' again, it will still display the previously selected item. **Expected behavior** I expected the selection property of the datatable to be set to null, like in previous versions of PF. **Sample code** I'm not attaching any code since it can be reproduced in the showcase itself. I hope that's ok.
1.0
DataTable: unselect not clearing selection in PFX - **Describe the defect** Testing out PFX. While following the migration guide, I noticed a problem with Datatable when performing a _unselect_. In my datatables, I noticed that _selection_ was not updated even though the rowUnselect event listener was called properly. Also, when I clear the selection from the model in my backing bean and update/re-render the datatable, the row would remain selected. **Environment:** - PF Version: _10.0.0_ - JSF + version: Mojarra 2.2.17 - Affected browsers: _ALL_ **To Reproduce** Steps to reproduce the behavior: 1. Go to the Showcase 2. Navigate to Datatable - Selection 3. Scroll down to 'Single with Row Click' 4. Try to click select a row, and press 'View'. Afterwards, CTRL click the row, thereby unselecting it. When pressing 'View' again, it will still display the previously selected item. **Expected behavior** I expected the selection property of the datatable to be set to null, like in previous versions of PF. **Sample code** I'm not attaching any code since it can be reproduced in the showcase itself. I hope that's ok.
defect
datatable unselect not clearing selection in pfx describe the defect testing out pfx while following the migration guide i noticed a problem with datatable when performing a unselect in my datatables i noticed that selection was not updated even though the rowunselect event listener was called properly also when i clear the selection from the model in my backing bean and update re render the datatable the row would remain selected environment pf version jsf version mojarra affected browsers all to reproduce steps to reproduce the behavior go to the showcase navigate to datatable selection scroll down to single with row click try to click select a row and press view afterwards ctrl click the row thereby unselecting it when pressing view again it will still display the previously selected item expected behavior i expected the selection property of the datatable to be set to null like in previous versions of pf sample code i m not attaching any code since it can be reproduced in the showcase itself i hope that s ok
1
124,824
12,239,674,860
IssuesEvent
2020-05-04 22:11:26
MikeHeiber/Excimontec
https://api.github.com/repos/MikeHeiber/Excimontec
closed
Add recommended reading section to README
documentation/tutorial
List high-quality review papers and book chapters to help new users learn how KMC simulations have been used in the organic electronics field in the past.
1.0
Add recommended reading section to README - List high-quality review papers and book chapters to help new users learn how KMC simulations have been used in the organic electronics field in the past.
non_defect
add recommended reading section to readme list high quality review papers and book chapters to help new users learn how kmc simulations have been used in the organic electronics field in the past
0
25,844
4,472,844,374
IssuesEvent
2016-08-26 00:08:17
zaproxy/zaproxy
https://api.github.com/repos/zaproxy/zaproxy
closed
Performing only a scan via the API can cause an infinite loop
Priority-Medium Type-Defect
``` What steps will reproduce the problem? Start a scan via the API without first proxying any traffic or running a spider. What is the expected output? What do you see instead? I would expect the scan either to fail or to immediately reach 100%. Instead the scan is triggered and the status always returns 0. What version of the product are you using? On what operating system? D-2013-04-08 on Mac OS X ``` Original issue reported on code.google.com by `dave.hunt` on 2013-04-16 10:05:06
1.0
Performing only a scan via the API can cause an infinite loop - ``` What steps will reproduce the problem? Start a scan via the API without first proxying any traffic or running a spider. What is the expected output? What do you see instead? I would expect the scan either to fail or to immediately reach 100%. Instead the scan is triggered and the status always returns 0. What version of the product are you using? On what operating system? D-2013-04-08 on Mac OS X ``` Original issue reported on code.google.com by `dave.hunt` on 2013-04-16 10:05:06
defect
performing only a scan via the api can cause an infinite loop what steps will reproduce the problem start a scan via the api without first proxying any traffic or running a spider what is the expected output what do you see instead i would expect the scan either to fail or to immediately reach instead the scan is triggered and the status always returns what version of the product are you using on what operating system d on mac os x original issue reported on code google com by dave hunt on
1
428,714
12,415,614,769
IssuesEvent
2020-05-22 16:35:18
TykTechnologies/tyk-pump
https://api.github.com/repos/TykTechnologies/tyk-pump
closed
Prometheus metric "tyk_http_status_per_key" not getting incremented
Priority: Low sla wontfix
Version: tyk-pump: 0.8.1 Prometheus metric "tyk_http_status_per_key" not getting incremented Even after making multiple API calls tyk_http_status_per_key{code="201",key="00000000"} 0 tyk_http_status_per_key{code="400",key="00000000"} 0 count remains 0 for all status codes , whereas is should have been incremented accordingly e.g. tyk_http_status_per_key{code="201",key="00000000"} 5 tyk_http_status_per_key{code="400",key="00000000"} 22
1.0
Prometheus metric "tyk_http_status_per_key" not getting incremented - Version: tyk-pump: 0.8.1 Prometheus metric "tyk_http_status_per_key" not getting incremented Even after making multiple API calls tyk_http_status_per_key{code="201",key="00000000"} 0 tyk_http_status_per_key{code="400",key="00000000"} 0 count remains 0 for all status codes , whereas is should have been incremented accordingly e.g. tyk_http_status_per_key{code="201",key="00000000"} 5 tyk_http_status_per_key{code="400",key="00000000"} 22
non_defect
prometheus metric tyk http status per key not getting incremented version tyk pump prometheus metric tyk http status per key not getting incremented even after making multiple api calls tyk http status per key code key tyk http status per key code key count remains for all status codes whereas is should have been incremented accordingly e g tyk http status per key code key tyk http status per key code key
0
7,388
2,601,760,591
IssuesEvent
2015-02-24 00:34:55
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
Const correctness
auto-migrated Component-Persistence Priority-None Type-Task
``` Many functions should be made constant, however it's more of a tedius chore. ``` ----- Original issue reported on code.google.com by `AHeinerm` on 25 Feb 2011 at 10:35
1.0
Const correctness - ``` Many functions should be made constant, however it's more of a tedius chore. ``` ----- Original issue reported on code.google.com by `AHeinerm` on 25 Feb 2011 at 10:35
non_defect
const correctness many functions should be made constant however it s more of a tedius chore original issue reported on code google com by aheinerm on feb at
0
67,130
8,074,999,014
IssuesEvent
2018-08-07 02:48:14
GitHawkApp/GitHawk
https://api.github.com/repos/GitHawkApp/GitHawk
closed
Resign keyboard when tapping manage button
design starter-task
Kind of annoying to have the kb flying around <details> <summary>Bug Report Dump (Auto-generated)</summary> <pre> Version 1.22.0 (1532006598) Device: iPhone X (iOS 11.4.1) TestFlight: true </pre> </details>
1.0
Resign keyboard when tapping manage button - Kind of annoying to have the kb flying around <details> <summary>Bug Report Dump (Auto-generated)</summary> <pre> Version 1.22.0 (1532006598) Device: iPhone X (iOS 11.4.1) TestFlight: true </pre> </details>
non_defect
resign keyboard when tapping manage button kind of annoying to have the kb flying around bug report dump auto generated version device iphone x ios testflight true
0
143,284
21,993,500,761
IssuesEvent
2022-05-26 02:11:45
harryodubhghaill/CI-Portforlio-4-blogsocial
https://api.github.com/repos/harryodubhghaill/CI-Portforlio-4-blogsocial
closed
DESIGN: Group Page
Design
**Full group page** **Page to hold a list of posts specific to that group** **Criteria** - [ ] Group Name - [ ] Group Description - [ ] Follow group button - [ ] Create post button - [ ] Paginated List of posts (similar to home page list #26) - [ ] Number of members
1.0
DESIGN: Group Page - **Full group page** **Page to hold a list of posts specific to that group** **Criteria** - [ ] Group Name - [ ] Group Description - [ ] Follow group button - [ ] Create post button - [ ] Paginated List of posts (similar to home page list #26) - [ ] Number of members
non_defect
design group page full group page page to hold a list of posts specific to that group criteria group name group description follow group button create post button paginated list of posts similar to home page list number of members
0
95,581
27,556,602,026
IssuesEvent
2023-03-07 18:25:35
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Build failure: pkgsCross.raspberryPi.hello (from aarch64-linux)
0.kind: build failure
### Steps To Reproduce See also https://discourse.nixos.org/t/cross-compiling-hello-to-raspberrypi/26093 Steps to reproduce the behavior: 1. `nix build -f '<nixpkgs>' pkgsCross.raspberryPi.hello` ### Build log https://gist.github.com/eliasnaur/b57b5c0337ba4fdf17271708db252b45 ### Additional context This is such as basic cross-compilation that I wonder whether something is wrong in my setup. However, the `exec format` errors are surprising, because I thought most packages were cross-compilation aware enough to not run host binaries on the build platform. ### Metadata Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result. ```console $ nix-shell -p nix-info --run "nix-info -m" - system: `"aarch64-linux"` - host os: `Linux 5.15.97, NixOS, 22.11 (Raccoon), 22.11.20230306.47c0034` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.11.1` - channels(e): `""` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos` ```
1.0
Build failure: pkgsCross.raspberryPi.hello (from aarch64-linux) - ### Steps To Reproduce See also https://discourse.nixos.org/t/cross-compiling-hello-to-raspberrypi/26093 Steps to reproduce the behavior: 1. `nix build -f '<nixpkgs>' pkgsCross.raspberryPi.hello` ### Build log https://gist.github.com/eliasnaur/b57b5c0337ba4fdf17271708db252b45 ### Additional context This is such as basic cross-compilation that I wonder whether something is wrong in my setup. However, the `exec format` errors are surprising, because I thought most packages were cross-compilation aware enough to not run host binaries on the build platform. ### Metadata Please run `nix-shell -p nix-info --run "nix-info -m"` and paste the result. ```console $ nix-shell -p nix-info --run "nix-info -m" - system: `"aarch64-linux"` - host os: `Linux 5.15.97, NixOS, 22.11 (Raccoon), 22.11.20230306.47c0034` - multi-user?: `yes` - sandbox: `yes` - version: `nix-env (Nix) 2.11.1` - channels(e): `""` - nixpkgs: `/nix/var/nix/profiles/per-user/root/channels/nixos` ```
non_defect
build failure pkgscross raspberrypi hello from linux steps to reproduce see also steps to reproduce the behavior nix build f pkgscross raspberrypi hello build log additional context this is such as basic cross compilation that i wonder whether something is wrong in my setup however the exec format errors are surprising because i thought most packages were cross compilation aware enough to not run host binaries on the build platform metadata please run nix shell p nix info run nix info m and paste the result console nix shell p nix info run nix info m system linux host os linux nixos raccoon multi user yes sandbox yes version nix env nix channels e nixpkgs nix var nix profiles per user root channels nixos
0
78,940
27,827,475,883
IssuesEvent
2023-03-19 22:43:07
scoutplan/scoutplan
https://api.github.com/repos/scoutplan/scoutplan
opened
[Scoutplan Production/production] KeyError: key not found: "OPENWEATHER_API_KEY"
defect
## Backtrace line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: fetch line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: api line 21 of [PROJECT_ROOT]/app/services/weather_service.rb: current [View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/94565340)
1.0
[Scoutplan Production/production] KeyError: key not found: "OPENWEATHER_API_KEY" - ## Backtrace line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: fetch line 34 of [PROJECT_ROOT]/app/services/weather_service.rb: api line 21 of [PROJECT_ROOT]/app/services/weather_service.rb: current [View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/94565340)
defect
keyerror key not found openweather api key backtrace line of app services weather service rb fetch line of app services weather service rb api line of app services weather service rb current
1
403,745
11,845,778,247
IssuesEvent
2020-03-24 09:01:02
AY1920S2-CS2103T-F09-4/main
https://api.github.com/repos/AY1920S2-CS2103T-F09-4/main
closed
Create Json adapted objects for data storage
priority.High type.Epic
- [ ] JsonAdaptedActivity - [ ] JsonAdaptedItinerary - [ ] JsonAdaptedEntry - [ ] JsonAdaptedJournal
1.0
Create Json adapted objects for data storage - - [ ] JsonAdaptedActivity - [ ] JsonAdaptedItinerary - [ ] JsonAdaptedEntry - [ ] JsonAdaptedJournal
non_defect
create json adapted objects for data storage jsonadaptedactivity jsonadapteditinerary jsonadaptedentry jsonadaptedjournal
0
684,948
23,439,366,576
IssuesEvent
2022-08-15 13:29:30
Plutonomicon/cardano-transaction-lib
https://api.github.com/repos/Plutonomicon/cardano-transaction-lib
closed
Improve AlwaysSucceeds example
enhancement good first issue lower-priority needs investigation
`countToZero` is a boring way to ensure a tx passes. We can skip some of the waiting by checking utxos available at script address. This code was suggested by Brian: ``` waitForTx :: Int -> ValidatorHash -> TransactionHash -> Contract () (Maybe TransactionInput) waitForTx n vhash txid = do let scriptAddress = scriptHashAddress vhash UtxoM utxos <- fromMaybe (UtxoM Map.empty) <$> utxosAt scriptAddress case fst <$> find hasTransactionId (Map.toUnfoldable utxos :: Array (TransactionInput /\ TransactionOutput)) of Nothing -> if (n <= 0) then do pure Nothing else do logInfo' $ "No tx yet, waiting for: " <> show n <> " more seconds" (liftAff <<< delay <<< wrap) 1000.0 waitForTx (n - 1) vhash txid Just txin -> do logInfo' $ "found tx:" <> show txid pure $ Just txin where hasTransactionId :: forall a. TransactionInput /\ a -> Boolean hasTransactionId (TransactionInput tx /\ _) = tx.transactionId == txid ```
1.0
Improve AlwaysSucceeds example - `countToZero` is a boring way to ensure a tx passes. We can skip some of the waiting by checking utxos available at script address. This code was suggested by Brian: ``` waitForTx :: Int -> ValidatorHash -> TransactionHash -> Contract () (Maybe TransactionInput) waitForTx n vhash txid = do let scriptAddress = scriptHashAddress vhash UtxoM utxos <- fromMaybe (UtxoM Map.empty) <$> utxosAt scriptAddress case fst <$> find hasTransactionId (Map.toUnfoldable utxos :: Array (TransactionInput /\ TransactionOutput)) of Nothing -> if (n <= 0) then do pure Nothing else do logInfo' $ "No tx yet, waiting for: " <> show n <> " more seconds" (liftAff <<< delay <<< wrap) 1000.0 waitForTx (n - 1) vhash txid Just txin -> do logInfo' $ "found tx:" <> show txid pure $ Just txin where hasTransactionId :: forall a. TransactionInput /\ a -> Boolean hasTransactionId (TransactionInput tx /\ _) = tx.transactionId == txid ```
non_defect
improve alwayssucceeds example counttozero is a boring way to ensure a tx passes we can skip some of the waiting by checking utxos available at script address this code was suggested by brian waitfortx int validatorhash transactionhash contract maybe transactioninput waitfortx n vhash txid do let scriptaddress scripthashaddress vhash utxom utxos utxosat scriptaddress case fst find hastransactionid map tounfoldable utxos array transactioninput transactionoutput of nothing if n then do pure nothing else do loginfo no tx yet waiting for show n more seconds liftaff delay wrap waitfortx n vhash txid just txin do loginfo found tx show txid pure just txin where hastransactionid forall a transactioninput a boolean hastransactionid transactioninput tx tx transactionid txid
0
46,689
13,055,959,860
IssuesEvent
2020-07-30 03:14:26
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
trigger-sim needs specially keyed volume trigger for each instance (Trac #1717)
Incomplete Migration Migrated from Trac combo simulation defect
Migrated from https://code.icecube.wisc.edu/ticket/1717 ```json { "status": "closed", "changetime": "2019-02-13T14:12:58", "description": "When running the trigger-sim segment on hitspool data younger than May 9th 2016, I get the following error: \n`RuntimeError: Tray already contains module named \"InIceRawdata_triggersim_cyl\" of type CylinderTrigger (in I3Tray::param_setter I3Tray::AddModule(boost::python::api::object, std::string))`\n\nWhich is caused (I guess) by multiple volume trigger instances in the frame. Would be nice to have the same behaviour as we have for multiple SMTs already as well for multiple clyinder triggers.", "reporter": "dheereman", "cc": "", "resolution": "fixed", "_ts": "1550067178841456", "component": "combo simulation", "summary": "trigger-sim needs specially keyed volume trigger for each instance", "priority": "critical", "keywords": "trigger-sim, multiple volume rigger instances", "time": "2016-05-30T20:05:37", "milestone": "", "owner": "olivas", "type": "defect" } ```
1.0
trigger-sim needs specially keyed volume trigger for each instance (Trac #1717) - Migrated from https://code.icecube.wisc.edu/ticket/1717 ```json { "status": "closed", "changetime": "2019-02-13T14:12:58", "description": "When running the trigger-sim segment on hitspool data younger than May 9th 2016, I get the following error: \n`RuntimeError: Tray already contains module named \"InIceRawdata_triggersim_cyl\" of type CylinderTrigger (in I3Tray::param_setter I3Tray::AddModule(boost::python::api::object, std::string))`\n\nWhich is caused (I guess) by multiple volume trigger instances in the frame. Would be nice to have the same behaviour as we have for multiple SMTs already as well for multiple clyinder triggers.", "reporter": "dheereman", "cc": "", "resolution": "fixed", "_ts": "1550067178841456", "component": "combo simulation", "summary": "trigger-sim needs specially keyed volume trigger for each instance", "priority": "critical", "keywords": "trigger-sim, multiple volume rigger instances", "time": "2016-05-30T20:05:37", "milestone": "", "owner": "olivas", "type": "defect" } ```
defect
trigger sim needs specially keyed volume trigger for each instance trac migrated from json status closed changetime description when running the trigger sim segment on hitspool data younger than may i get the following error n runtimeerror tray already contains module named inicerawdata triggersim cyl of type cylindertrigger in param setter addmodule boost python api object std string n nwhich is caused i guess by multiple volume trigger instances in the frame would be nice to have the same behaviour as we have for multiple smts already as well for multiple clyinder triggers reporter dheereman cc resolution fixed ts component combo simulation summary trigger sim needs specially keyed volume trigger for each instance priority critical keywords trigger sim multiple volume rigger instances time milestone owner olivas type defect
1
51,547
13,207,523,854
IssuesEvent
2020-08-14 23:26:36
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
examples project can not be built without phys-services (Trac #604)
IceTray Incomplete Migration Migrated from Trac defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/604">https://code.icecube.wisc.edu/projects/icecube/ticket/604</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T20:52:58", "_ts": "1351716778000000", "description": "The examples project (now in icetray svn)\ncan not be build or cmake'd with out the icecube phys-services\nproject.\n\nNeed to cleave these better to examples can serve all by itself.\n\nSeveral of the modules in the examples project use the phys-services code (I3Calculator) and the scripts load phys-services libraries.", "reporter": "blaufuss", "cc": "", "resolution": "wont or cant fix", "time": "2010-03-11T15:28:56", "component": "IceTray", "summary": "examples project can not be built without phys-services", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
examples project can not be built without phys-services (Trac #604) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/604">https://code.icecube.wisc.edu/projects/icecube/ticket/604</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T20:52:58", "_ts": "1351716778000000", "description": "The examples project (now in icetray svn)\ncan not be build or cmake'd with out the icecube phys-services\nproject.\n\nNeed to cleave these better to examples can serve all by itself.\n\nSeveral of the modules in the examples project use the phys-services code (I3Calculator) and the scripts load phys-services libraries.", "reporter": "blaufuss", "cc": "", "resolution": "wont or cant fix", "time": "2010-03-11T15:28:56", "component": "IceTray", "summary": "examples project can not be built without phys-services", "priority": "normal", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
defect
examples project can not be built without phys services trac migrated from json status closed changetime ts description the examples project now in icetray svn ncan not be build or cmake d with out the icecube phys services nproject n nneed to cleave these better to examples can serve all by itself n nseveral of the modules in the examples project use the phys services code and the scripts load phys services libraries reporter blaufuss cc resolution wont or cant fix time component icetray summary examples project can not be built without phys services priority normal keywords milestone owner troy type defect
1
687,192
23,516,387,243
IssuesEvent
2022-08-18 21:55:42
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
closed
Don't allow merging a work with more than 50 editions into another work
Type: Bug Priority: 2 Affects: Librarians Needs: Investigation Lead: @cdrini
Sub-task of #6556 The Merge UI currently allows this, but only moves over the first 50 editions! Leaving the rest in a sort of limbo. Prevent this from being allowed, or investigate if we can increase it to a larger limit, but we'll probably still need to block at whatever limit that is. ### Stakeholders @mheiman @dcapillae
1.0
Don't allow merging a work with more than 50 editions into another work - Sub-task of #6556 The Merge UI currently allows this, but only moves over the first 50 editions! Leaving the rest in a sort of limbo. Prevent this from being allowed, or investigate if we can increase it to a larger limit, but we'll probably still need to block at whatever limit that is. ### Stakeholders @mheiman @dcapillae
non_defect
don t allow merging a work with more than editions into another work sub task of the merge ui currently allows this but only moves over the first editions leaving the rest in a sort of limbo prevent this from being allowed or investigate if we can increase it to a larger limit but we ll probably still need to block at whatever limit that is stakeholders mheiman dcapillae
0
583,615
17,393,960,522
IssuesEvent
2021-08-02 11:03:54
JonasMuehlmann/BIB_VPR
https://api.github.com/repos/JonasMuehlmann/BIB_VPR
opened
Role Permissions
bug frontend priority: medium
You can not save the changes for a Role when you add more then 2 Permissions.
1.0
Role Permissions - You can not save the changes for a Role when you add more then 2 Permissions.
non_defect
role permissions you can not save the changes for a role when you add more then permissions
0
10,858
2,622,193,654
IssuesEvent
2015-03-04 00:24:12
byzhang/cudpp
https://api.github.com/repos/byzhang/cudpp
opened
cudppPlan execution time
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. run cudppPlan in any function What is the expected output? What do you see instead? The output is correct (all tests in cudpp_testrig passed), but the first call of the function cudppPlan is slow (about a second). What version of the product are you using? On what operating system? Motherboard Rampage III GENE Intel(R) Core(TM) i7 CPU X 980 @ 3.33GHz DIMM 1333 MHz 6GiB GeForce GTX680 Ubuntu 12.04 Cudpp 2.0 Cuda 5.0 Nvidia Driver version 310.19 Please provide any additional information below. I mentioned an "infinite loop" issue in cudppPlan here: https://groups.google.com/forum/?fromgroups=#!topic/cudpp/J4AEijEFzW4 The two problems seem to be related somehow: if I make an exit(2) just before the cudppPlan call (in simpleCUDPP.cu for example), the execution time is normal. If I make it on the first line of cudppPlan, it takes about 0.8 second to make the exit on my GTX680 card. I also have a GT240 card. The same problem happens but it takes about 0.5 second. I just made a new compilation of the cudpp library and the execution time of simpleCUDPP was about 15 seconds. A second run and this is about one second again. ``` Original issue reported on code.google.com by `nicolas....@gmail.com` on 10 Dec 2012 at 6:46
1.0
cudppPlan execution time - ``` What steps will reproduce the problem? 1. run cudppPlan in any function What is the expected output? What do you see instead? The output is correct (all tests in cudpp_testrig passed), but the first call of the function cudppPlan is slow (about a second). What version of the product are you using? On what operating system? Motherboard Rampage III GENE Intel(R) Core(TM) i7 CPU X 980 @ 3.33GHz DIMM 1333 MHz 6GiB GeForce GTX680 Ubuntu 12.04 Cudpp 2.0 Cuda 5.0 Nvidia Driver version 310.19 Please provide any additional information below. I mentioned an "infinite loop" issue in cudppPlan here: https://groups.google.com/forum/?fromgroups=#!topic/cudpp/J4AEijEFzW4 The two problems seem to be related somehow: if I make an exit(2) just before the cudppPlan call (in simpleCUDPP.cu for example), the execution time is normal. If I make it on the first line of cudppPlan, it takes about 0.8 second to make the exit on my GTX680 card. I also have a GT240 card. The same problem happens but it takes about 0.5 second. I just made a new compilation of the cudpp library and the execution time of simpleCUDPP was about 15 seconds. A second run and this is about one second again. ``` Original issue reported on code.google.com by `nicolas....@gmail.com` on 10 Dec 2012 at 6:46
defect
cudppplan execution time what steps will reproduce the problem run cudppplan in any function what is the expected output what do you see instead the output is correct all tests in cudpp testrig passed but the first call of the function cudppplan is slow about a second what version of the product are you using on what operating system motherboard rampage iii gene intel r core tm cpu x dimm mhz geforce ubuntu cudpp cuda nvidia driver version please provide any additional information below i mentioned an infinite loop issue in cudppplan here the two problems seem to be related somehow if i make an exit just before the cudppplan call in simplecudpp cu for example the execution time is normal if i make it on the first line of cudppplan it takes about second to make the exit on my card i also have a card the same problem happens but it takes about second i just made a new compilation of the cudpp library and the execution time of simplecudpp was about seconds a second run and this is about one second again original issue reported on code google com by nicolas gmail com on dec at
1