Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
54,741 | 3,071,164,643 | IssuesEvent | 2015-08-19 10:13:37 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Проблемы с распознаванием URL'ом | bug Component-Logic Component-UI imported Priority-Medium Usability | _From [sa.stol...@gmail.com](https://code.google.com/u/103855734279594150839/) on April 19, 2010 10:40:16_
При распозновании URL'ов используется конечный символ - пробел. Необходимо
также считать конечным символом - ковычку ("), скобки ((,),[]) - может
что-то еще?
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=80_ | 1.0 | Проблемы с распознаванием URL'ом - _From [sa.stol...@gmail.com](https://code.google.com/u/103855734279594150839/) on April 19, 2010 10:40:16_
При распозновании URL'ов используется конечный символ - пробел. Необходимо
также считать конечным символом - ковычку ("), скобки ((,),[]) - может
что-то еще?
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=80_ | priority | проблемы с распознаванием url ом from on april при распозновании url ов используется конечный символ пробел необходимо также считать конечным символом ковычку скобки может что то еще original issue | 1 |
364,292 | 10,761,791,035 | IssuesEvent | 2019-10-31 21:37:37 | ix/notewell | https://api.github.com/repos/ix/notewell | opened | Markdown image support | medium priority | Notewell should _probably_ support images. It's not pivotal in my opinion and it'd be quite a lot of hassle to support, but it's not exactly complete unless it does. | 1.0 | Markdown image support - Notewell should _probably_ support images. It's not pivotal in my opinion and it'd be quite a lot of hassle to support, but it's not exactly complete unless it does. | priority | markdown image support notewell should probably support images it s not pivotal in my opinion and it d be quite a lot of hassle to support but it s not exactly complete unless it does | 1 |
315,476 | 9,621,072,863 | IssuesEvent | 2019-05-14 09:46:30 | cms-gem-daq-project/cmsgemos | https://api.github.com/repos/cms-gem-daq-project/cmsgemos | opened | [feature] Send calibration parameters to manager applications | Priority: Medium Type: Feature Request | ## Brief summary of issue
In order to have the hardware receive correct scan configuration, the parameters need to be sent from the `CalibrationSuite` to the appropriate `<HW>Manager` applications.
### Types of issue
- [x] Feature request (request for change which adds functionality)
## Expected Behavior
This can happen by doing the following:
* Define a `CalibrationParameters` object in `CalibrationSuite`
* Should contain `xdata::<Type>` objects (see, e.g., [`OptoHybridManager.h`](https://github.com/cms-gem-daq-project/cmsgemos/blob/feature/gemcalib/gemhardware/include/gem/hw/optohybrid/OptoHybridManager.h#L64-L93)
* Define a SOAP callback in the manager applications `xoap::bind()`, (see, e.g., [`GEMSupervisor`](https://github.com/cms-gem-daq-project/cmsgemos/blob/1fdff44b5089203c526de06419de766c46cddd1b/gemsupervisor/src/common/GEMSupervisor.cc#L65) [and](https://github.com/cms-gem-daq-project/cmsgemos/blob/1fdff44b5089203c526de06419de766c46cddd1b/gemsupervisor/src/common/GEMSupervisor.cc#L1291-L1370))
* This callback should parse the calibration parameters object and set the necessary ones in the manager
* For a first test, you can print out the parameters received in the manager log message
* From the `CalibrationSuite` send the SOAP message to the appropriate application(s) using: [`gem::utils::soap::GEMSOAPToolBox::sendCommandWithParameterBag`](https://github.com/jsturdy/cmsgemos/blob/subfeature/hw-db-interface/gemutils/src/common/soap/GEMSOAPToolBox.cc#L344-L434)
* You will have to translate the `CalibrationParameters` into the appropriate type for serialization (it *may* be possible to do this as an `xdata::Bag` object) | 1.0 | [feature] Send calibration parameters to manager applications - ## Brief summary of issue
In order to have the hardware receive correct scan configuration, the parameters need to be sent from the `CalibrationSuite` to the appropriate `<HW>Manager` applications.
### Types of issue
- [x] Feature request (request for change which adds functionality)
## Expected Behavior
This can happen by doing the following:
* Define a `CalibrationParameters` object in `CalibrationSuite`
* Should contain `xdata::<Type>` objects (see, e.g., [`OptoHybridManager.h`](https://github.com/cms-gem-daq-project/cmsgemos/blob/feature/gemcalib/gemhardware/include/gem/hw/optohybrid/OptoHybridManager.h#L64-L93)
* Define a SOAP callback in the manager applications `xoap::bind()`, (see, e.g., [`GEMSupervisor`](https://github.com/cms-gem-daq-project/cmsgemos/blob/1fdff44b5089203c526de06419de766c46cddd1b/gemsupervisor/src/common/GEMSupervisor.cc#L65) [and](https://github.com/cms-gem-daq-project/cmsgemos/blob/1fdff44b5089203c526de06419de766c46cddd1b/gemsupervisor/src/common/GEMSupervisor.cc#L1291-L1370))
* This callback should parse the calibration parameters object and set the necessary ones in the manager
* For a first test, you can print out the parameters received in the manager log message
* From the `CalibrationSuite` send the SOAP message to the appropriate application(s) using: [`gem::utils::soap::GEMSOAPToolBox::sendCommandWithParameterBag`](https://github.com/jsturdy/cmsgemos/blob/subfeature/hw-db-interface/gemutils/src/common/soap/GEMSOAPToolBox.cc#L344-L434)
* You will have to translate the `CalibrationParameters` into the appropriate type for serialization (it *may* be possible to do this as an `xdata::Bag` object) | priority | send calibration parameters to manager applications brief summary of issue in order to have the hardware receive correct scan configuration the parameters need to be sent from the calibrationsuite to the appropriate manager applications types of issue feature request request for change which adds functionality expected behavior this can happen by doing the following define a calibrationparameters object in calibrationsuite should contain xdata objects see e g define a soap callback in the manager applications xoap bind see e g this callback should parse the calibration parameters object and set the necessary ones in the manager for a first test you can print out the parameters received in the manager log message from the calibrationsuite send the soap message to the appropriate application s using you will have to translate the calibrationparameters into the appropriate type for serialization it may be possible to do this as an xdata bag object | 1 |
537,284 | 15,726,489,421 | IssuesEvent | 2021-03-29 11:23:15 | AY2021S2-CS2103T-T11-4/tp | https://api.github.com/repos/AY2021S2-CS2103T-T11-4/tp | closed | FindCommand - Modify to cater to more functionalities | priority.Medium type.Enhancement | Aligning to the purpose of CakeCollate to track orders, enable the FindCommand to be flexible and have more fields to find keywords from so that user can properly track specific orders.
Flexibility - to allow for finding of substrings instead of having to key in a full keyword. Case-insensitive maybe.
More field - to allow for finding of keywords in fields other than n/NAME. For instance, user may key in "find a/Orchard" to find orders with address that contains "Orchard". | 1.0 | FindCommand - Modify to cater to more functionalities - Aligning to the purpose of CakeCollate to track orders, enable the FindCommand to be flexible and have more fields to find keywords from so that user can properly track specific orders.
Flexibility - to allow for finding of substrings instead of having to key in a full keyword. Case-insensitive maybe.
More field - to allow for finding of keywords in fields other than n/NAME. For instance, user may key in "find a/Orchard" to find orders with address that contains "Orchard". | priority | findcommand modify to cater to more functionalities aligning to the purpose of cakecollate to track orders enable the findcommand to be flexible and have more fields to find keywords from so that user can properly track specific orders flexibility to allow for finding of substrings instead of having to key in a full keyword case insensitive maybe more field to allow for finding of keywords in fields other than n name for instance user may key in find a orchard to find orders with address that contains orchard | 1 |
176,125 | 6,556,785,188 | IssuesEvent | 2017-09-06 15:11:13 | buttercup/buttercup-mobile | https://api.github.com/repos/buttercup/buttercup-mobile | opened | Remote explorer item icons | Effort: Low Priority: Medium Status: Available Type: Enhancement | The remote explorer, when selecting or creating an archive, needs **local** icons. | 1.0 | Remote explorer item icons - The remote explorer, when selecting or creating an archive, needs **local** icons. | priority | remote explorer item icons the remote explorer when selecting or creating an archive needs local icons | 1 |
108,188 | 4,328,764,814 | IssuesEvent | 2016-07-26 14:56:48 | DigitalCampus/django-oppia | https://api.github.com/repos/DigitalCampus/django-oppia | closed | Summary cron - oppia.models.DoesNotExist: Tracker matching query does not exist | bug medium priority | Get the following error when running the summary cron task against an empty tracker table...
Starting Oppia Summary cron...
Traceback (most recent call last):
File "/home/oppiamobile/django-oppia/oppia/summary/cron.py", line 104, in <module>
run()
File "/home/oppiamobile/django-oppia/oppia/summary/cron.py", line 24, in run
newest_tracker_pk = Tracker.objects.latest('id').id
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/manager.py", line 92, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/query.py", line 502, in latest
return self._earliest_or_latest(field_name=field_name, direction="-")
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/query.py", line 496, in _earliest_or_latest
return obj.get()
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/query.py", line 357, in get
self.model._meta.object_name)
oppia.models.DoesNotExist: Tracker matching query does not exist.
| 1.0 | Summary cron - oppia.models.DoesNotExist: Tracker matching query does not exist - Get the following error when running the summary cron task against an empty tracker table...
Starting Oppia Summary cron...
Traceback (most recent call last):
File "/home/oppiamobile/django-oppia/oppia/summary/cron.py", line 104, in <module>
run()
File "/home/oppiamobile/django-oppia/oppia/summary/cron.py", line 24, in run
newest_tracker_pk = Tracker.objects.latest('id').id
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/manager.py", line 92, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/query.py", line 502, in latest
return self._earliest_or_latest(field_name=field_name, direction="-")
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/query.py", line 496, in _earliest_or_latest
return obj.get()
File "/home/oppiamobile/env/local/lib/python2.7/site-packages/django/db/models/query.py", line 357, in get
self.model._meta.object_name)
oppia.models.DoesNotExist: Tracker matching query does not exist.
| priority | summary cron oppia models doesnotexist tracker matching query does not exist get the following error when running the summary cron task against an empty tracker table starting oppia summary cron traceback most recent call last file home oppiamobile django oppia oppia summary cron py line in run file home oppiamobile django oppia oppia summary cron py line in run newest tracker pk tracker objects latest id id file home oppiamobile env local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file home oppiamobile env local lib site packages django db models query py line in latest return self earliest or latest field name field name direction file home oppiamobile env local lib site packages django db models query py line in earliest or latest return obj get file home oppiamobile env local lib site packages django db models query py line in get self model meta object name oppia models doesnotexist tracker matching query does not exist | 1 |
80,043 | 3,549,870,549 | IssuesEvent | 2016-01-20 19:42:38 | bbengfort/commis | https://api.github.com/repos/bbengfort/commis | opened | Label Command | priority: medium type: feature | A management command which takes one or more arbitrary arguments (labels) on the command line, and does something with each of them.
Rather than implementing `handle()`, subclasses must implement `handle_label()`, which will be called once for each label. | 1.0 | Label Command - A management command which takes one or more arbitrary arguments (labels) on the command line, and does something with each of them.
Rather than implementing `handle()`, subclasses must implement `handle_label()`, which will be called once for each label. | priority | label command a management command which takes one or more arbitrary arguments labels on the command line and does something with each of them rather than implementing handle subclasses must implement handle label which will be called once for each label | 1 |
420,924 | 12,246,010,307 | IssuesEvent | 2020-05-05 13:51:06 | ansible/galaxy_ng | https://api.github.com/repos/ansible/galaxy_ng | closed | TASK [pulp-database : Run database auth migrations] in playbook install.yml fails. | area/installer priority/medium status/new type/bug | TASK [pulp-database : Run database auth migrations] fails when running playbook install.yml
TASK [pulp-database : Run database auth migrations] ******************************************************************************************************************************
fatal: [35.178.172.220]: FAILED! => {"changed": true, "cmd": ["/usr/local/lib/pulp/bin/django-admin", "migrate", "auth", "--no-input"], "delta": "0:00:00.658146", "end": "2020-04-19 11:08:35.107148", "msg": "non-zero return code", "rc": 1, "start": "2020-04-19 11:08:34.449002", "stderr": "Traceback (most recent call last):\n File \"/usr/local/lib/pulp/bin/django-admin\", line 8, in <module>\n sys.exit(execute_from_command_line())\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 381, in execute_from_command_line\n utility.execute()\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 325, in execute\n settings.INSTALLED_APPS\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 79, in __getattr__\n self._setup(name)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 66, in _setup\n self._wrapped = Settings(settings_module)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 157, in __init__\n mod = importlib.import_module(self.SETTINGS_MODULE)\n File \"/usr/lib/python3.6/importlib/__init__.py\", line 126, in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\n File \"/usr/local/lib/pulp/src/pulpcore/pulpcore/app/settings.py\", line 73, in <module>\n plugin_app_config = entry_point.load()\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2449, in load\n self.require(*args, **kwargs)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2472, in require\n items = working_set.resolve(reqs, env, installer, extras=self.extras)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 792, in resolve\n raise VersionConflict(dist, req).with_context(dependent_req)\npkg_resources.VersionConflict: (pulpcore 3.4.0.dev0 (/usr/local/lib/pulp/src/pulpcore), Requirement.parse('pulpcore<3.3,>=3.0'))", "stderr_lines": ["Traceback (most recent call last):", " File \"/usr/local/lib/pulp/bin/django-admin\", line 8, in <module>", " sys.exit(execute_from_command_line())", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 381, in execute_from_command_line", " utility.execute()", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 325, in execute", " settings.INSTALLED_APPS", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 79, in __getattr__", " self._setup(name)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 66, in _setup", " self._wrapped = Settings(settings_module)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 157, in __init__", " mod = importlib.import_module(self.SETTINGS_MODULE)", " File \"/usr/lib/python3.6/importlib/__init__.py\", line 126, in import_module", " return _bootstrap._gcd_import(name[level:], package, level)", " File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import", " File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load", " File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked", " File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked", " File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module", " File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed", " File \"/usr/local/lib/pulp/src/pulpcore/pulpcore/app/settings.py\", line 73, in <module>", " plugin_app_config = entry_point.load()", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2449, in load", " self.require(*args, **kwargs)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2472, in require", " items = working_set.resolve(reqs, env, installer, extras=self.extras)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 792, in resolve", " raise VersionConflict(dist, req).with_context(dependent_req)", "pkg_resources.VersionConflict: (pulpcore 3.4.0.dev0 (/usr/local/lib/pulp/src/pulpcore), Requirement.parse('pulpcore<3.3,>=3.0'))"], "stdout": "", "stdout_lines": []}
| 1.0 | TASK [pulp-database : Run database auth migrations] in playbook install.yml fails. - TASK [pulp-database : Run database auth migrations] fails when running playbook install.yml
TASK [pulp-database : Run database auth migrations] ******************************************************************************************************************************
fatal: [35.178.172.220]: FAILED! => {"changed": true, "cmd": ["/usr/local/lib/pulp/bin/django-admin", "migrate", "auth", "--no-input"], "delta": "0:00:00.658146", "end": "2020-04-19 11:08:35.107148", "msg": "non-zero return code", "rc": 1, "start": "2020-04-19 11:08:34.449002", "stderr": "Traceback (most recent call last):\n File \"/usr/local/lib/pulp/bin/django-admin\", line 8, in <module>\n sys.exit(execute_from_command_line())\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 381, in execute_from_command_line\n utility.execute()\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 325, in execute\n settings.INSTALLED_APPS\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 79, in __getattr__\n self._setup(name)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 66, in _setup\n self._wrapped = Settings(settings_module)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 157, in __init__\n mod = importlib.import_module(self.SETTINGS_MODULE)\n File \"/usr/lib/python3.6/importlib/__init__.py\", line 126, in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import\n File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load\n File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked\n File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked\n File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module\n File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed\n File \"/usr/local/lib/pulp/src/pulpcore/pulpcore/app/settings.py\", line 73, in <module>\n plugin_app_config = entry_point.load()\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2449, in load\n self.require(*args, **kwargs)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2472, in require\n items = working_set.resolve(reqs, env, installer, extras=self.extras)\n File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 792, in resolve\n raise VersionConflict(dist, req).with_context(dependent_req)\npkg_resources.VersionConflict: (pulpcore 3.4.0.dev0 (/usr/local/lib/pulp/src/pulpcore), Requirement.parse('pulpcore<3.3,>=3.0'))", "stderr_lines": ["Traceback (most recent call last):", " File \"/usr/local/lib/pulp/bin/django-admin\", line 8, in <module>", " sys.exit(execute_from_command_line())", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 381, in execute_from_command_line", " utility.execute()", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/core/management/__init__.py\", line 325, in execute", " settings.INSTALLED_APPS", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 79, in __getattr__", " self._setup(name)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 66, in _setup", " self._wrapped = Settings(settings_module)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/django/conf/__init__.py\", line 157, in __init__", " mod = importlib.import_module(self.SETTINGS_MODULE)", " File \"/usr/lib/python3.6/importlib/__init__.py\", line 126, in import_module", " return _bootstrap._gcd_import(name[level:], package, level)", " File \"<frozen importlib._bootstrap>\", line 994, in _gcd_import", " File \"<frozen importlib._bootstrap>\", line 971, in _find_and_load", " File \"<frozen importlib._bootstrap>\", line 955, in _find_and_load_unlocked", " File \"<frozen importlib._bootstrap>\", line 665, in _load_unlocked", " File \"<frozen importlib._bootstrap_external>\", line 678, in exec_module", " File \"<frozen importlib._bootstrap>\", line 219, in _call_with_frames_removed", " File \"/usr/local/lib/pulp/src/pulpcore/pulpcore/app/settings.py\", line 73, in <module>", " plugin_app_config = entry_point.load()", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2449, in load", " self.require(*args, **kwargs)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 2472, in require", " items = working_set.resolve(reqs, env, installer, extras=self.extras)", " File \"/usr/local/lib/pulp/lib/python3.6/site-packages/pkg_resources/__init__.py\", line 792, in resolve", " raise VersionConflict(dist, req).with_context(dependent_req)", "pkg_resources.VersionConflict: (pulpcore 3.4.0.dev0 (/usr/local/lib/pulp/src/pulpcore), Requirement.parse('pulpcore<3.3,>=3.0'))"], "stdout": "", "stdout_lines": []}
| priority | task in playbook install yml fails task fails when running playbook install yml task fatal failed changed true cmd delta end msg non zero return code rc start stderr traceback most recent call last n file usr local lib pulp bin django admin line in n sys exit execute from command line n file usr local lib pulp lib site packages django core management init py line in execute from command line n utility execute n file usr local lib pulp lib site packages django core management init py line in execute n settings installed apps n file usr local lib pulp lib site packages django conf init py line in getattr n self setup name n file usr local lib pulp lib site packages django conf init py line in setup n self wrapped settings settings module n file usr local lib pulp lib site packages django conf init py line in init n mod importlib import module self settings module n file usr lib importlib init py line in import module n return bootstrap gcd import name package level n file line in gcd import n file line in find and load n file line in find and load unlocked n file line in load unlocked n file line in exec module n file line in call with frames removed n file usr local lib pulp src pulpcore pulpcore app settings py line in n plugin app config entry point load n file usr local lib pulp lib site packages pkg resources init py line in load n self require args kwargs n file usr local lib pulp lib site packages pkg resources init py line in require n items working set resolve reqs env installer extras self extras n file usr local lib pulp lib site packages pkg resources init py line in resolve n raise versionconflict dist req with context dependent req npkg resources versionconflict pulpcore usr local lib pulp src pulpcore requirement parse pulpcore stderr lines package level file line in gcd import file line in find and load file line in find and load unlocked file line in load unlocked file line in exec module file line in call with frames removed file usr local lib pulp src pulpcore pulpcore app settings py line in plugin app config entry point load file usr local lib pulp lib site packages pkg resources init py line in load self require args kwargs file usr local lib pulp lib site packages pkg resources init py line in require items working set resolve reqs env installer extras self extras file usr local lib pulp lib site packages pkg resources init py line in resolve raise versionconflict dist req with context dependent req pkg resources versionconflict pulpcore usr local lib pulp src pulpcore requirement parse pulpcore stdout stdout lines | 1 |
435,999 | 12,543,918,058 | IssuesEvent | 2020-06-05 16:21:07 | graknlabs/grakn | https://api.github.com/repos/graknlabs/grakn | opened | Incorrect Graql behaviour in some scenarios (major issues) | priority: medium type: bug | ## Description
A number of minor issues have been found while crafting BDD scenarios, where the actual behaviour of Graql does not match the expected behaviour.
## Environment
1. OS (where Grakn server runs): Mac OS 10
2. Grakn version: Grakn Core 1.7.2
3. Grakn client: client-java
## Scenarios
### Scenario: define relation subtype inherits 'relates' from supertypes without role subtyping
#### expected behaviour
given
```
define
employment sub relation, relates employee;
part-time-employment sub employment;
```
then `part-time-employment` should `relates employee`.
#### actual behaviour
The roleplayer `employee` is not inherited.
### Scenario: define additional 'key' on a type throws if it is not added to existing instances prior to commit
#### expected behaviour
given
```
define
name sub attribute, value string;
barcode sub attribute, value string;
product sub entity, has name;
insert
$x isa product, has name "Cheese";
$y isa product, has name "Ham";
```
then `define product key barcode;` should throw on commit if values for `barcode` have not been added to the existing products.
#### actual behaviour
It doesn't throw, and we now have instances that don't have the correct keys. | 1.0 | Incorrect Graql behaviour in some scenarios (major issues) - ## Description
A number of minor issues have been found while crafting BDD scenarios, where the actual behaviour of Graql does not match the expected behaviour.
## Environment
1. OS (where Grakn server runs): Mac OS 10
2. Grakn version: Grakn Core 1.7.2
3. Grakn client: client-java
## Scenarios
### Scenario: define relation subtype inherits 'relates' from supertypes without role subtyping
#### expected behaviour
given
```
define
employment sub relation, relates employee;
part-time-employment sub employment;
```
then `part-time-employment` should `relates employee`.
#### actual behaviour
The roleplayer `employee` is not inherited.
### Scenario: define additional 'key' on a type throws if it is not added to existing instances prior to commit
#### expected behaviour
given
```
define
name sub attribute, value string;
barcode sub attribute, value string;
product sub entity, has name;
insert
$x isa product, has name "Cheese";
$y isa product, has name "Ham";
```
then `define product key barcode;` should throw on commit if values for `barcode` have not been added to the existing products.
#### actual behaviour
It doesn't throw, and we now have instances that don't have the correct keys. | priority | incorrect graql behaviour in some scenarios major issues description a number of minor issues have been found while crafting bdd scenarios where the actual behaviour of graql does not match the expected behaviour environment os where grakn server runs mac os grakn version grakn core grakn client client java scenarios scenario define relation subtype inherits relates from supertypes without role subtyping expected behaviour given define employment sub relation relates employee part time employment sub employment then part time employment should relates employee actual behaviour the roleplayer employee is not inherited scenario define additional key on a type throws if it is not added to existing instances prior to commit expected behaviour given define name sub attribute value string barcode sub attribute value string product sub entity has name insert x isa product has name cheese y isa product has name ham then define product key barcode should throw on commit if values for barcode have not been added to the existing products actual behaviour it doesn t throw and we now have instances that don t have the correct keys | 1 |
89,080 | 3,789,517,527 | IssuesEvent | 2016-03-21 18:12:10 | PolarisSS13/Polaris | https://api.github.com/repos/PolarisSS13/Polaris | closed | Changeling armblade and flavortext fuckery | Bug Priority: Medium | Getting stunned or otherwise disarmed as a ling will instantly remove the armblade you just shelled 20 chems for, leaving lings to constantly reform it in combat.
The armblade does not seem deflect any projectile weapons. Not sure if this is an issue or if its just intended to work with lasers.
Changing identities while having snowflake text will retain the previous snowflake text, making lings a dead giveaway to the powergaymen in red suits. | 1.0 | Changeling armblade and flavortext fuckery - Getting stunned or otherwise disarmed as a ling will instantly remove the armblade you just shelled 20 chems for, leaving lings to constantly reform it in combat.
The armblade does not seem deflect any projectile weapons. Not sure if this is an issue or if its just intended to work with lasers.
Changing identities while having snowflake text will retain the previous snowflake text, making lings a dead giveaway to the powergaymen in red suits. | priority | changeling armblade and flavortext fuckery getting stunned or otherwise disarmed as a ling will instantly remove the armblade you just shelled chems for leaving lings to constantly reform it in combat the armblade does not seem deflect any projectile weapons not sure if this is an issue or if its just intended to work with lasers changing identities while having snowflake text will retain the previous snowflake text making lings a dead giveaway to the powergaymen in red suits | 1 |
576,415 | 17,086,634,503 | IssuesEvent | 2021-07-08 12:41:53 | CCAFS/MARLO | https://api.github.com/repos/CCAFS/MARLO | closed | [KT] (AICCRA) Cluster tags (Management, Flagship, Regional and Country) | AICCRA Priority - Medium Type - Enhancement | In AICCRA, we have 4 types of projects which would be really great we can tag somehow.
1. Flagship led, 2. Regional led, 3. Country led, 4. PMU led (Cluster Type)

- [x] Database
- [x] Manager
- [x] Migration
- [x] Front end componets
| 1.0 | [KT] (AICCRA) Cluster tags (Management, Flagship, Regional and Country) - In AICCRA, we have 4 types of projects which would be really great we can tag somehow.
1. Flagship led, 2. Regional led, 3. Country led, 4. PMU led (Cluster Type)

- [x] Database
- [x] Manager
- [x] Migration
- [x] Front end componets
| priority | aiccra cluster tags management flagship regional and country in aiccra we have types of projects which would be really great we can tag somehow flagship led regional led country led pmu led cluster type database manager migration front end componets | 1 |
387,662 | 11,464,132,137 | IssuesEvent | 2020-02-07 17:23:06 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | opened | MultiSelect on adding a new item the dataSource requestEnd event args return "read" type of request | Bug C: MultiSelect FP: Planned Kendo2 Priority 5 SEV: Medium | ### Bug report
When a new item is added and the dataSource's sync method is called, the requestEnd event handler data (**arg.type**) returns the type of request as "read", instead of "create".
As a result the [Add new item](https://demos.telerik.com/kendo-ui/multiselect/addnewitem) demo, does not work as expected, because it has a check for the type of the request in the requestEnd handler, and expects the request to be "create". Since the request type comes out as "read" the logic for selecting the newly added item is not executed.
In previous versions the request has been correctly identified as "create". The issue is exhibited only in the MultiSelect. The ComboBox and the DropDownList return the request as "create".
This behavior has been introduced in R3 2017. Reproducible in Chrome, Firefox and Chromium Edge. Not reproducible in IE11 and Spartan Edge.
As a **workaround** the addNew function can be modified as shown below:
```
function addNew(widgetId, value) {
var widget = $("#" + widgetId).getKendoMultiSelect();
var dataSource = widget.dataSource;
if (confirm("Are you sure?")) {
dataSource.add({
ProductID: 0,
ProductName: value
});
dataSource.one("sync", function() {
var index = dataSource.view().length - 1;
var newValue = dataSource.at(index).ProductID;
widget.value(widget.value().concat([newValue]));
});
dataSource.sync();
}
}
```
### Reproduction of the problem
[Dojo ](https://dojo.telerik.com/eSohATUz) example.
1. Open the browser's console.
2. Focus the input and type in some random text.
3. Click the button in the popup to add a new item.
4. The type of the request returned by the requestEnd event data is logged in the console.
### Current behavior
The event data returns "read" as the type of the request.
### Expected/desired behavior
The event data returns "create" as the type of the request.
### Environment
* **Kendo UI version:** 2020.1.114
* **jQuery version:** x.y
* **Browser:** [Chrome XX | Firefox XX | Chromium Edge ]
| 1.0 | MultiSelect on adding a new item the dataSource requestEnd event args return "read" type of request - ### Bug report
When a new item is added and the dataSource's sync method is called, the requestEnd event handler data (**arg.type**) returns the type of request as "read", instead of "create".
As a result the [Add new item](https://demos.telerik.com/kendo-ui/multiselect/addnewitem) demo, does not work as expected, because it has a check for the type of the request in the requestEnd handler, and expects the request to be "create". Since the request type comes out as "read" the logic for selecting the newly added item is not executed.
In previous versions the request has been correctly identified as "create". The issue is exhibited only in the MultiSelect. The ComboBox and the DropDownList return the request as "create".
This behavior has been introduced in R3 2017. Reproducible in Chrome, Firefox and Chromium Edge. Not reproducible in IE11 and Spartan Edge.
As a **workaround** the addNew function can be modified as shown below:
```
function addNew(widgetId, value) {
var widget = $("#" + widgetId).getKendoMultiSelect();
var dataSource = widget.dataSource;
if (confirm("Are you sure?")) {
dataSource.add({
ProductID: 0,
ProductName: value
});
dataSource.one("sync", function() {
var index = dataSource.view().length - 1;
var newValue = dataSource.at(index).ProductID;
widget.value(widget.value().concat([newValue]));
});
dataSource.sync();
}
}
```
### Reproduction of the problem
[Dojo ](https://dojo.telerik.com/eSohATUz) example.
1. Open the browser's console.
2. Focus the input and type in some random text.
3. Click the button in the popup to add a new item.
4. The type of the request returned by the requestEnd event data is logged in the console.
### Current behavior
The event data returns "read" as the type of the request.
### Expected/desired behavior
The event data returns "create" as the type of the request.
### Environment
* **Kendo UI version:** 2020.1.114
* **jQuery version:** x.y
* **Browser:** [Chrome XX | Firefox XX | Chromium Edge ]
| priority | multiselect on adding a new item the datasource requestend event args return read type of request bug report when a new item is added and the datasource s sync method is called the requestend event handler data arg type returns the type of request as read instead of create as a result the demo does not work as expected because it has a check for the type of the request in the requestend handler and expects the request to be create since the request type comes out as read the logic for selecting the newly added item is not executed in previous versions the request has been correctly identified as create the issue is exhibited only in the multiselect the combobox and the dropdownlist return the request as create this behavior has been introduced in reproducible in chrome firefox and chromium edge not reproducible in and spartan edge as a workaround the addnew function can be modified as shown below function addnew widgetid value var widget widgetid getkendomultiselect var datasource widget datasource if confirm are you sure datasource add productid productname value datasource one sync function var index datasource view length var newvalue datasource at index productid widget value widget value concat datasource sync reproduction of the problem example open the browser s console focus the input and type in some random text click the button in the popup to add a new item the type of the request returned by the requestend event data is logged in the console current behavior the event data returns read as the type of the request expected desired behavior the event data returns create as the type of the request environment kendo ui version jquery version x y browser | 1 |
56,366 | 3,079,497,047 | IssuesEvent | 2015-08-21 16:34:53 | pavel-pimenov/flylinkdc-r5xx | https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx | closed | Свойства избранного хаба -> Прямое соединение -> IP (разрешить DNS name) | Component-Logic enhancement imported Priority-Medium Usability | _From [vts...@gmail.com](https://code.google.com/u/103957436465684378630/) on November 06, 2011 13:41:11_
Добавить возможность вводить DNS в поле адреса для прямого соединения.
Подробно:
Есть серая сеть провайдера для которой работают основные настройки
и есть VPN подключение с реальным IP. Реальный IP каждый раз меняется, но настроен DynDns который может по DNS name вернуть текущий IP.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=587_ | 1.0 | Свойства избранного хаба -> Прямое соединение -> IP (разрешить DNS name) - _From [vts...@gmail.com](https://code.google.com/u/103957436465684378630/) on November 06, 2011 13:41:11_
Добавить возможность вводить DNS в поле адреса для прямого соединения.
Подробно:
Есть серая сеть провайдера для которой работают основные настройки
и есть VPN подключение с реальным IP. Реальный IP каждый раз меняется, но настроен DynDns который может по DNS name вернуть текущий IP.
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=587_ | priority | свойства избранного хаба прямое соединение ip разрешить dns name from on november добавить возможность вводить dns в поле адреса для прямого соединения подробно есть серая сеть провайдера для которой работают основные настройки и есть vpn подключение с реальным ip реальный ip каждый раз меняется но настроен dyndns который может по dns name вернуть текущий ip original issue | 1 |
749,646 | 26,172,936,477 | IssuesEvent | 2023-01-02 04:34:44 | battlecode/galaxy | https://api.github.com/repos/battlecode/galaxy | closed | error_handling.js can obscure error messages; dump in console | type: bug type: feature module: frontend priority: p2 medium | error_handling.js is good at parsing many error messages, and turning them into human-readable form.
However, sometimes the error message we get is uncommon, or the result of a deeper error. In this case, our parser will parse it...but because it's not getting expected input, it gives bad output. This can confuse devs, make for extra digging, etc.
We _could_ have a more robust handler. But this is really hard -- it's hard to balance human-readability with covering all cases. There's a reason we have a custom parser, after all.
The better idea is to simply **dump the full error message/array in the console.** (Not insecure -- since the network response in console can give it to us, and motivated ppl would just check that anyways) And it's much more convenient.
Side-note: from dev experience, you should prob learn that error_handling misparses.
**Should we include a note in readme? "If response looks weird, open your console and retry and see what's printed".** | 1.0 | error_handling.js can obscure error messages; dump in console - error_handling.js is good at parsing many error messages, and turning them into human-readable form.
However, sometimes the error message we get is uncommon, or the result of a deeper error. In this case, our parser will parse it...but because it's not getting expected input, it gives bad output. This can confuse devs, make for extra digging, etc.
We _could_ have a more robust handler. But this is really hard -- it's hard to balance human-readability with covering all cases. There's a reason we have a custom parser, after all.
The better idea is to simply **dump the full error message/array in the console.** (Not insecure -- since the network response in console can give it to us, and motivated ppl would just check that anyways) And it's much more convenient.
Side-note: from dev experience, you should prob learn that error_handling misparses.
**Should we include a note in readme? "If response looks weird, open your console and retry and see what's printed".** | priority | error handling js can obscure error messages dump in console error handling js is good at parsing many error messages and turning them into human readable form however sometimes the error message we get is uncommon or the result of a deeper error in this case our parser will parse it but because it s not getting expected input it gives bad output this can confuse devs make for extra digging etc we could have a more robust handler but this is really hard it s hard to balance human readability with covering all cases there s a reason we have a custom parser after all the better idea is to simply dump the full error message array in the console not insecure since the network response in console can give it to us and motivated ppl would just check that anyways and it s much more convenient side note from dev experience you should prob learn that error handling misparses should we include a note in readme if response looks weird open your console and retry and see what s printed | 1 |
663,298 | 22,172,297,832 | IssuesEvent | 2022-06-06 03:11:29 | authzed/spicedb | https://api.github.com/repos/authzed/spicedb | closed | Migration functions should have a context.Context parameter | priority/2 medium area/datastore | Also, we should probably be able to execute the migration and the part that writes the version in a single transaction. | 1.0 | Migration functions should have a context.Context parameter - Also, we should probably be able to execute the migration and the part that writes the version in a single transaction. | priority | migration functions should have a context context parameter also we should probably be able to execute the migration and the part that writes the version in a single transaction | 1 |
354,050 | 10,562,347,131 | IssuesEvent | 2019-10-04 18:06:00 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | Occultism events are not adapted | :beetle: bug :beetle: :grey_exclamation: priority medium | **Mod Version**
46bfcd20
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
Occultism events (69040-69047) can fire for any religion
**Steps to reproduce the issue:**
`---`
**Upload an attachment below: .zip of your save, or screenshots:**
<details>
<summary>Click to expand</summary>

</details> | 1.0 | Occultism events are not adapted - **Mod Version**
46bfcd20
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
Occultism events (69040-69047) can fire for any religion
**Steps to reproduce the issue:**
`---`
**Upload an attachment below: .zip of your save, or screenshots:**
<details>
<summary>Click to expand</summary>

</details> | priority | occultism events are not adapted mod version what expansions do you have installed all please explain your issue in as much detail as possible occultism events can fire for any religion steps to reproduce the issue upload an attachment below zip of your save or screenshots click to expand | 1 |
621,295 | 19,582,557,928 | IssuesEvent | 2022-01-04 23:56:00 | bats-core/bats-core | https://api.github.com/repos/bats-core/bats-core | closed | bats waits for bg processes in `setup_file` even when FD3 is closed | Type: Bug Component: Docs Priority: Medium | I'm following the indications in [https://bats-core.readthedocs.io/en/stable/writing-tests.html\#file-descriptor-3-read-this-if-bats-hangs](https://bats-core.readthedocs.io/en/stable/writing-tests.html#file-descriptor-3-read-this-if-bats-hangs) and that seems to work when bg processes like `sleep 5s 3>- &` are launched from the test body or the (per-test) `setup` function. But when launched from the (per-file) `setup_file` function, `sleep 5s 3>- &` is still blocking the bats execution after all tests are done.
To reproduce:
```
@test "bats does not hang on bg process in setup_file" {
cd "$BATS_TEST_TMPDIR"
cat <<EOF >foo.bats
setup_file (){
sleep 5s 3>- &
}
$(echo @test) "foo" {
true
}
EOF
SECONDS=0
run bats foo.bats
test $SECONDS -lt 2
}
```
Is this a known issue? I have seen very recent work around this topic in [https://github.com/bats-core/bats-core/pull/525](https://github.com/bats-core/bats-core/pull/525) but not sure if that will apply to this particular issue.
**Environment:**
- Bats 1.5.0
- OS: Fedora 33
- Bash version: GNU bash, version 5.0.17(1)-release (x86\_64-redhat-linux-gnu)
| 1.0 | bats waits for bg processes in `setup_file` even when FD3 is closed - I'm following the indications in [https://bats-core.readthedocs.io/en/stable/writing-tests.html\#file-descriptor-3-read-this-if-bats-hangs](https://bats-core.readthedocs.io/en/stable/writing-tests.html#file-descriptor-3-read-this-if-bats-hangs) and that seems to work when bg processes like `sleep 5s 3>- &` are launched from the test body or the (per-test) `setup` function. But when launched from the (per-file) `setup_file` function, `sleep 5s 3>- &` is still blocking the bats execution after all tests are done.
To reproduce:
```
@test "bats does not hang on bg process in setup_file" {
cd "$BATS_TEST_TMPDIR"
cat <<EOF >foo.bats
setup_file (){
sleep 5s 3>- &
}
$(echo @test) "foo" {
true
}
EOF
SECONDS=0
run bats foo.bats
test $SECONDS -lt 2
}
```
Is this a known issue? I have seen very recent work around this topic in [https://github.com/bats-core/bats-core/pull/525](https://github.com/bats-core/bats-core/pull/525) but not sure if that will apply to this particular issue.
**Environment:**
- Bats 1.5.0
- OS: Fedora 33
- Bash version: GNU bash, version 5.0.17(1)-release (x86\_64-redhat-linux-gnu)
| priority | bats waits for bg processes in setup file even when is closed i m following the indications in and that seems to work when bg processes like sleep are launched from the test body or the per test setup function but when launched from the per file setup file function sleep is still blocking the bats execution after all tests are done to reproduce test bats does not hang on bg process in setup file cd bats test tmpdir cat foo bats setup file sleep echo test foo true eof seconds run bats foo bats test seconds lt is this a known issue i have seen very recent work around this topic in but not sure if that will apply to this particular issue environment bats os fedora bash version gnu bash version release redhat linux gnu | 1 |
424,604 | 12,313,533,367 | IssuesEvent | 2020-05-12 15:26:25 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | Unable to reconfigure existing filter | Priority: Medium Type: Bug | **Describe the bug**
I'm trying to reconfigure an existing filter and change its logic but moving objects around is glitchy
**To Reproduce**
1. Showed in live to @satkunas
2. Use the following config:
```
[02_no_employees_on_student_ssid]
status=enabled
condition=node_info.category != "Eleves" && ssid == "pi" || ssid == "niz"
role=REJECT
scopes=RegisteredRole
```
| 1.0 | Unable to reconfigure existing filter - **Describe the bug**
I'm trying to reconfigure an existing filter and change its logic but moving objects around is glitchy
**To Reproduce**
1. Showed in live to @satkunas
2. Use the following config:
```
[02_no_employees_on_student_ssid]
status=enabled
condition=node_info.category != "Eleves" && ssid == "pi" || ssid == "niz"
role=REJECT
scopes=RegisteredRole
```
| priority | unable to reconfigure existing filter describe the bug i m trying to reconfigure an existing filter and change its logic but moving objects around is glitchy to reproduce showed in live to satkunas use the following config status enabled condition node info category eleves ssid pi ssid niz role reject scopes registeredrole | 1 |
721,194 | 24,820,835,674 | IssuesEvent | 2022-10-25 16:18:00 | AY2223S1-CS2103-F14-3/tp | https://api.github.com/repos/AY2223S1-CS2103-F14-3/tp | closed | As a user, I can easily view upcoming interviews happening within 1 week from now | type.Story priority.Medium | ... so that I can prepare for them accordingly. | 1.0 | As a user, I can easily view upcoming interviews happening within 1 week from now - ... so that I can prepare for them accordingly. | priority | as a user i can easily view upcoming interviews happening within week from now so that i can prepare for them accordingly | 1 |
809,861 | 30,215,177,959 | IssuesEvent | 2023-07-05 15:07:43 | fmv1001/F1RacePredictor | https://api.github.com/repos/fmv1001/F1RacePredictor | closed | Desarrollar la lógica de la aplicación para la interacción con el modelo | medium priority | En esta tarea se crearán los medios necesarios para que la aplicación interactúe con el modelo y devuelva los resultados a la interfaz. | 1.0 | Desarrollar la lógica de la aplicación para la interacción con el modelo - En esta tarea se crearán los medios necesarios para que la aplicación interactúe con el modelo y devuelva los resultados a la interfaz. | priority | desarrollar la lógica de la aplicación para la interacción con el modelo en esta tarea se crearán los medios necesarios para que la aplicación interactúe con el modelo y devuelva los resultados a la interfaz | 1 |
204,049 | 7,079,882,942 | IssuesEvent | 2018-01-10 11:19:17 | gluster/glusterd2 | https://api.github.com/repos/gluster/glusterd2 | closed | New format for Volume create request | FW: ReST priority: medium | Current Volume Create request format is very limited since it converts
the list of bricks into sub volumes based on other parameters and
order of bricks. With new format, each sub volume can have its own type
and list of bricks.
For example, to create distribute replicate volume,
Existing,
{
"name": "gv1",
"replica": 2,
"bricks": [
"b4e2a4a5-103f-4ae7-8545-706b7c5039e9:/bricks/b1",
"72b66b1e-e29c-450a-a4af-bfc8499aae54:/bricks/b2",
"b0e0b26c-fecc-4e24-8edc-f2f328258301:/bricks/b3",
"9fd10082-84b7-431d-a58c-9a6d060fd7cd:/bricks/b4"
],
"transport": "tcp"
}
Proposed,
{
"name": "gv1",
"subvols": [
{
"name": "s1",
"type": "replicate",
"bricks": [
{"nodeid": "b4e2a4a5-103f-4ae7-8545-706b7c5039e9", "path": "/bricks/b1"},
{"nodeid": "72b66b1e-e29c-450a-a4af-bfc8499aae54", "path": "/bricks/b2"}
]
},
{
"name": "s2",
"type": "replicate",
"bricks": [
{"nodeid": "b0e0b26c-fecc-4e24-8edc-f2f328258301", "path": "/bricks/b3"},
{"nodeid": "9fd10082-84b7-431d-a58c-9a6d060fd7cd", "path": "/bricks/b4"}
]
}
],
"transport": "tcp"
}
In case of Arbiter volume,(Note, bricks order is not strict)
{
"name": "gv1",
"subvols": [
{
"name": "s1",
"type": "arbiter",
"bricks": [
{"nodeid": "b4e2a4a5-103f-4ae7-8545-706b7c5039e9", "path": "/bricks/b1"},
{"nodeid": "72b66b1e-e29c-450a-a4af-bfc8499aae54", "path": "/bricks/b2"},
{"nodeid": "59a6033e-07c3-49bb-92a0-f38c8dfbdb0a", "path": "/bricks/b3", "arbiter": true}
]
},
{
"name": "s2",
"type": "arbiter",
"bricks": [
{"nodeid": "b0e0b26c-fecc-4e24-8edc-f2f328258301", "path": "/bricks/b4"},
{"nodeid": "9fd10082-84b7-431d-a58c-9a6d060fd7cd", "path": "/bricks/b5", "arbiter": true},
{"nodeid": "1b9f7313-a85b-46ba-a2ff-e9620f0f06bb", "path": "/bricks/b6"}
]
}
],
"transport": "tcp"
}
Let me know if any changes required to the format or any new fields
required.
## Expected changes
- REST Client package, CLI to REST Client communication remains same
- CLI changes while displaying volinfo
- Volinfo struct changes
- Volume commands changes while accessing information
- REST API documentation changes
| 1.0 | New format for Volume create request - Current Volume Create request format is very limited since it converts
the list of bricks into sub volumes based on other parameters and
order of bricks. With new format, each sub volume can have its own type
and list of bricks.
For example, to create distribute replicate volume,
Existing,
{
"name": "gv1",
"replica": 2,
"bricks": [
"b4e2a4a5-103f-4ae7-8545-706b7c5039e9:/bricks/b1",
"72b66b1e-e29c-450a-a4af-bfc8499aae54:/bricks/b2",
"b0e0b26c-fecc-4e24-8edc-f2f328258301:/bricks/b3",
"9fd10082-84b7-431d-a58c-9a6d060fd7cd:/bricks/b4"
],
"transport": "tcp"
}
Proposed,
{
"name": "gv1",
"subvols": [
{
"name": "s1",
"type": "replicate",
"bricks": [
{"nodeid": "b4e2a4a5-103f-4ae7-8545-706b7c5039e9", "path": "/bricks/b1"},
{"nodeid": "72b66b1e-e29c-450a-a4af-bfc8499aae54", "path": "/bricks/b2"}
]
},
{
"name": "s2",
"type": "replicate",
"bricks": [
{"nodeid": "b0e0b26c-fecc-4e24-8edc-f2f328258301", "path": "/bricks/b3"},
{"nodeid": "9fd10082-84b7-431d-a58c-9a6d060fd7cd", "path": "/bricks/b4"}
]
}
],
"transport": "tcp"
}
In case of Arbiter volume,(Note, bricks order is not strict)
{
"name": "gv1",
"subvols": [
{
"name": "s1",
"type": "arbiter",
"bricks": [
{"nodeid": "b4e2a4a5-103f-4ae7-8545-706b7c5039e9", "path": "/bricks/b1"},
{"nodeid": "72b66b1e-e29c-450a-a4af-bfc8499aae54", "path": "/bricks/b2"},
{"nodeid": "59a6033e-07c3-49bb-92a0-f38c8dfbdb0a", "path": "/bricks/b3", "arbiter": true}
]
},
{
"name": "s2",
"type": "arbiter",
"bricks": [
{"nodeid": "b0e0b26c-fecc-4e24-8edc-f2f328258301", "path": "/bricks/b4"},
{"nodeid": "9fd10082-84b7-431d-a58c-9a6d060fd7cd", "path": "/bricks/b5", "arbiter": true},
{"nodeid": "1b9f7313-a85b-46ba-a2ff-e9620f0f06bb", "path": "/bricks/b6"}
]
}
],
"transport": "tcp"
}
Let me know if any changes required to the format or any new fields
required.
## Expected changes
- REST Client package, CLI to REST Client communication remains same
- CLI changes while displaying volinfo
- Volinfo struct changes
- Volume commands changes while accessing information
- REST API documentation changes
| priority | new format for volume create request current volume create request format is very limited since it converts the list of bricks into sub volumes based on other parameters and order of bricks with new format each sub volume can have its own type and list of bricks for example to create distribute replicate volume existing name replica bricks bricks bricks fecc bricks bricks transport tcp proposed name subvols name type replicate bricks nodeid path bricks nodeid path bricks name type replicate bricks nodeid fecc path bricks nodeid path bricks transport tcp in case of arbiter volume note bricks order is not strict name subvols name type arbiter bricks nodeid path bricks nodeid path bricks nodeid path bricks arbiter true name type arbiter bricks nodeid fecc path bricks nodeid path bricks arbiter true nodeid path bricks transport tcp let me know if any changes required to the format or any new fields required expected changes rest client package cli to rest client communication remains same cli changes while displaying volinfo volinfo struct changes volume commands changes while accessing information rest api documentation changes | 1 |
272,343 | 8,507,509,779 | IssuesEvent | 2018-10-30 19:16:27 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | opened | bazel: Use `try-import %workspace%/user.bazelrc` | configuration: bazel priority: medium team: kitware | Follow-up from #9864; super useful to not have to mutate `$HOME` dotfiles.
Can use once we have Bazel 0.18.
\cc @jamiesnape @sammy-tri | 1.0 | bazel: Use `try-import %workspace%/user.bazelrc` - Follow-up from #9864; super useful to not have to mutate `$HOME` dotfiles.
Can use once we have Bazel 0.18.
\cc @jamiesnape @sammy-tri | priority | bazel use try import workspace user bazelrc follow up from super useful to not have to mutate home dotfiles can use once we have bazel cc jamiesnape sammy tri | 1 |
274,762 | 8,565,045,582 | IssuesEvent | 2018-11-09 18:36:40 | sdss/marvin | https://api.github.com/repos/sdss/marvin | closed | VACs break pickling | marvin-tools priority-medium | <!-- **NEVER INCLUDE PLAINTEXT PASSWORDS OR PRIVATE INFORMATION IN THE BUG REPORT** -->
**Describe the bug**
Pickling seems broken everywhere with the introduction of the VACs.
**Additional context**
One example:
```python
test_spaxel.py:383:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../../tools/spaxel.py:249: in save
return marvin.core.marvin_pickle.save(self, path=path, overwrite=overwrite)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
obj = <Marvin Spaxel (plateifu=8485-1901, x=25, y=15; x_cen=8, y_cen=-2, loaded=maps/modelcube)>
path = '/private/var/folders/gp/vsl4fq4d47s9mpywmcbrsj7c0000gn/T/pytest-of-albireo/pytest-3/scratch7/test_spaxel.mpf'
overwrite = True
def save(obj, path=None, overwrite=False):
"""Pickles the object.
If ``path=None``, uses the default location of the file in the tree
but changes the extension of the file to ``.mpf`` (MaNGA Pickle File).
Returns the path of the saved pickle file.
Parameters:
obj:
Marvin object to pickle.
path (str):
Path of saved file. Default is ``None``.
overwrite (bool):
If ``True``, overwrite existing file. Default is ``False``.
Returns:
str:
Path of saved file.
"""
from ..tools.core import MarvinToolsClass
if path is None:
assert isinstance(obj, MarvinToolsClass), 'path=None is only allowed for core objects.'
path = obj._getFullPath()
assert isinstance(path, string_types), 'path must be a string.'
if path is None:
raise MarvinError('cannot determine the default path in the '
'tree for this file. You can overcome this '
'by calling save with a path keyword with '
'the path to which the file should be saved.')
# Replaces the extension (normally fits) with mpf
if '.fits.gz' in path:
path = path.strip('.fits.gz')
else:
path = os.path.splitext(path)[0]
path += '.mpf'
path = os.path.realpath(os.path.expanduser(path))
if os.path.isdir(path):
raise MarvinError('path must be a full route, including the filename.')
if os.path.exists(path) and not overwrite:
warnings.warn('file already exists. Not overwriting.', MarvinUserWarning)
return
dirname = os.path.dirname(path)
if not os.path.exists(dirname):
os.makedirs(dirname)
try:
with open(path, 'wb') as fout:
pickle.dump(obj, fout, protocol=-1)
except Exception as ee:
if os.path.exists(path):
os.remove(path)
> raise MarvinError('error found while pickling: {0}'.format(str(ee)))
E marvin.core.exceptions.MarvinError: error found while pickling: Can't pickle local object 'VACMixIn.get_vacs.<locals>.VACContainer'.
```
| 1.0 | VACs break pickling - <!-- **NEVER INCLUDE PLAINTEXT PASSWORDS OR PRIVATE INFORMATION IN THE BUG REPORT** -->
**Describe the bug**
Pickling seems broken everywhere with the introduction of the VACs.
**Additional context**
One example:
```python
test_spaxel.py:383:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../../tools/spaxel.py:249: in save
return marvin.core.marvin_pickle.save(self, path=path, overwrite=overwrite)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
obj = <Marvin Spaxel (plateifu=8485-1901, x=25, y=15; x_cen=8, y_cen=-2, loaded=maps/modelcube)>
path = '/private/var/folders/gp/vsl4fq4d47s9mpywmcbrsj7c0000gn/T/pytest-of-albireo/pytest-3/scratch7/test_spaxel.mpf'
overwrite = True
def save(obj, path=None, overwrite=False):
"""Pickles the object.
If ``path=None``, uses the default location of the file in the tree
but changes the extension of the file to ``.mpf`` (MaNGA Pickle File).
Returns the path of the saved pickle file.
Parameters:
obj:
Marvin object to pickle.
path (str):
Path of saved file. Default is ``None``.
overwrite (bool):
If ``True``, overwrite existing file. Default is ``False``.
Returns:
str:
Path of saved file.
"""
from ..tools.core import MarvinToolsClass
if path is None:
assert isinstance(obj, MarvinToolsClass), 'path=None is only allowed for core objects.'
path = obj._getFullPath()
assert isinstance(path, string_types), 'path must be a string.'
if path is None:
raise MarvinError('cannot determine the default path in the '
'tree for this file. You can overcome this '
'by calling save with a path keyword with '
'the path to which the file should be saved.')
# Replaces the extension (normally fits) with mpf
if '.fits.gz' in path:
path = path.strip('.fits.gz')
else:
path = os.path.splitext(path)[0]
path += '.mpf'
path = os.path.realpath(os.path.expanduser(path))
if os.path.isdir(path):
raise MarvinError('path must be a full route, including the filename.')
if os.path.exists(path) and not overwrite:
warnings.warn('file already exists. Not overwriting.', MarvinUserWarning)
return
dirname = os.path.dirname(path)
if not os.path.exists(dirname):
os.makedirs(dirname)
try:
with open(path, 'wb') as fout:
pickle.dump(obj, fout, protocol=-1)
except Exception as ee:
if os.path.exists(path):
os.remove(path)
> raise MarvinError('error found while pickling: {0}'.format(str(ee)))
E marvin.core.exceptions.MarvinError: error found while pickling: Can't pickle local object 'VACMixIn.get_vacs.<locals>.VACContainer'.
```
| priority | vacs break pickling describe the bug pickling seems broken everywhere with the introduction of the vacs additional context one example python test spaxel py tools spaxel py in save return marvin core marvin pickle save self path path overwrite overwrite obj path private var folders gp t pytest of albireo pytest test spaxel mpf overwrite true def save obj path none overwrite false pickles the object if path none uses the default location of the file in the tree but changes the extension of the file to mpf manga pickle file returns the path of the saved pickle file parameters obj marvin object to pickle path str path of saved file default is none overwrite bool if true overwrite existing file default is false returns str path of saved file from tools core import marvintoolsclass if path is none assert isinstance obj marvintoolsclass path none is only allowed for core objects path obj getfullpath assert isinstance path string types path must be a string if path is none raise marvinerror cannot determine the default path in the tree for this file you can overcome this by calling save with a path keyword with the path to which the file should be saved replaces the extension normally fits with mpf if fits gz in path path path strip fits gz else path os path splitext path path mpf path os path realpath os path expanduser path if os path isdir path raise marvinerror path must be a full route including the filename if os path exists path and not overwrite warnings warn file already exists not overwriting marvinuserwarning return dirname os path dirname path if not os path exists dirname os makedirs dirname try with open path wb as fout pickle dump obj fout protocol except exception as ee if os path exists path os remove path raise marvinerror error found while pickling format str ee e marvin core exceptions marvinerror error found while pickling can t pickle local object vacmixin get vacs vaccontainer | 1 |
338,445 | 10,229,415,212 | IssuesEvent | 2019-08-17 12:30:45 | ifedchankau/trading-risk-manager | https://api.github.com/repos/ifedchankau/trading-risk-manager | opened | Position size calculation | Priority: Medium Type: feature | Realize position size calculation based on:
- stop-loss level (calculated in `find_order_levels` function) and position open price
- exchange or broker comissions (fetching by API or stored in file: consider both realizations)
- slippage of market orders and stop-losses (stored in file for every market) | 1.0 | Position size calculation - Realize position size calculation based on:
- stop-loss level (calculated in `find_order_levels` function) and position open price
- exchange or broker comissions (fetching by API or stored in file: consider both realizations)
- slippage of market orders and stop-losses (stored in file for every market) | priority | position size calculation realize position size calculation based on stop loss level calculated in find order levels function and position open price exchange or broker comissions fetching by api or stored in file consider both realizations slippage of market orders and stop losses stored in file for every market | 1 |
362,305 | 10,725,978,661 | IssuesEvent | 2019-10-28 08:17:30 | she-code-africa/SCA-Website | https://api.github.com/repos/she-code-africa/SCA-Website | opened | Convert blog post editor to markdown editor | back-end priority: medium | This is to convert the current WYSIWYG editor used when creating or editing a blog post to a markdown equivalent which is also provided by Keystone - https://v4.keystonejs.com/api/field/markdown
The markdown section of this article may also be helpful with implementation - https://modernweb.com/building-blog-keystone-cms-node-js/ | 1.0 | Convert blog post editor to markdown editor - This is to convert the current WYSIWYG editor used when creating or editing a blog post to a markdown equivalent which is also provided by Keystone - https://v4.keystonejs.com/api/field/markdown
The markdown section of this article may also be helpful with implementation - https://modernweb.com/building-blog-keystone-cms-node-js/ | priority | convert blog post editor to markdown editor this is to convert the current wysiwyg editor used when creating or editing a blog post to a markdown equivalent which is also provided by keystone the markdown section of this article may also be helpful with implementation | 1 |
700,558 | 24,064,417,839 | IssuesEvent | 2022-09-17 09:10:23 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Group preset for `@tanstack/react-query` and `@tanstack/react-query-devtools` | type:feature priority-3-medium status:in-progress core:config | ### What would you like Renovate to be able to do?
I think you're supposed to update the `@tanstack/react-query` and `@tanstack/react-query-devtools` packages in one go together.
### If you have any ideas on how this should be implemented, please tell us here.
Create a new `group:tanstack` preset, like this:
```json
{
"packageRules": [
{
"matchDatasources": [
"npm"
],
"matchPackageNames": [
"@tanstack/react-query",
"@tanstack/react-query-devtools"
],
"groupName": "tanstack react-query packages"
}
]
}
```
### Is this a feature you are interested in implementing yourself?
Maybe | 1.0 | Group preset for `@tanstack/react-query` and `@tanstack/react-query-devtools` - ### What would you like Renovate to be able to do?
I think you're supposed to update the `@tanstack/react-query` and `@tanstack/react-query-devtools` packages in one go together.
### If you have any ideas on how this should be implemented, please tell us here.
Create a new `group:tanstack` preset, like this:
```json
{
"packageRules": [
{
"matchDatasources": [
"npm"
],
"matchPackageNames": [
"@tanstack/react-query",
"@tanstack/react-query-devtools"
],
"groupName": "tanstack react-query packages"
}
]
}
```
### Is this a feature you are interested in implementing yourself?
Maybe | priority | group preset for tanstack react query and tanstack react query devtools what would you like renovate to be able to do i think you re supposed to update the tanstack react query and tanstack react query devtools packages in one go together if you have any ideas on how this should be implemented please tell us here create a new group tanstack preset like this json packagerules matchdatasources npm matchpackagenames tanstack react query tanstack react query devtools groupname tanstack react query packages is this a feature you are interested in implementing yourself maybe | 1 |
276,952 | 8,614,824,748 | IssuesEvent | 2018-11-19 18:39:42 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | AMA Intake | Don't show ineligible issues as having been added to an appeal | Ready-for-Dev bug-medium-priority caseflow-intake sierra | Currently, ineligible issues are shown as having been added to an appeal. Even though we save them to the database, they should not considered to be issues on the appeal (similar to HLRs and SCs, where we do not create a contention)
## Acceptance Criteria
- Ineligible issues should not show up on the confirmation page of an appeal intake
## Screenshot
In this screen shot, one of these two issues was ineligible

| 1.0 | AMA Intake | Don't show ineligible issues as having been added to an appeal - Currently, ineligible issues are shown as having been added to an appeal. Even though we save them to the database, they should not considered to be issues on the appeal (similar to HLRs and SCs, where we do not create a contention)
## Acceptance Criteria
- Ineligible issues should not show up on the confirmation page of an appeal intake
## Screenshot
In this screen shot, one of these two issues was ineligible

| priority | ama intake don t show ineligible issues as having been added to an appeal currently ineligible issues are shown as having been added to an appeal even though we save them to the database they should not considered to be issues on the appeal similar to hlrs and scs where we do not create a contention acceptance criteria ineligible issues should not show up on the confirmation page of an appeal intake screenshot in this screen shot one of these two issues was ineligible | 1 |
154,232 | 5,916,172,221 | IssuesEvent | 2017-05-22 09:49:02 | harryshipton/secsplit | https://api.github.com/repos/harryshipton/secsplit | closed | Add a migration command | feature request medium priority | You should be able to migrate to the latest shard version by a single command, rather than having to merge and then resplit.
Extension to #8 (which is a prerequisite). | 1.0 | Add a migration command - You should be able to migrate to the latest shard version by a single command, rather than having to merge and then resplit.
Extension to #8 (which is a prerequisite). | priority | add a migration command you should be able to migrate to the latest shard version by a single command rather than having to merge and then resplit extension to which is a prerequisite | 1 |
520,369 | 15,085,248,554 | IssuesEvent | 2021-02-05 18:22:29 | dtcenter/MET | https://api.github.com/repos/dtcenter/MET | closed | Update the plotting R-scripts to handle output from different versions of MET. | alert: NEED ACCOUNT KEY component: user support component: utility scripts priority: medium requestor: Community type: enhancement | ## Describe the Enhancement ##
This issue is based on a request that came in via met-help:
https://rt.rap.ucar.edu/rt/Ticket/Display.html?id=98517
I actually already posted updates for plot_cnt.R and plot_mpr.R prior to realizing that these scripts actually do live in the MET repository. So this task is to make the same updates in the develop branch. Also, check if the same changes are needed in any of the other R-scripts. The change is this...
- add a -met_base command line argument
- if not specified there, check for a MET_BASE environment variable
- if neither are set, error out
- read the data and extract the MET version number from the first column of the first line
- read the corresponding header for that line type/version number from MET_BASE/table_files/met_header_columns_vX.Y.txt
- remove the hard-coded headers from the top of these scripts
### Time Estimate ###
4 hours.
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
No sub-issues required.
### Relevant Deadlines ###
None.
### Funding Source ###
?
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required: John HG
- [x] Select **scientist(s)** or **no scientist** required: no scientist required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [x] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
No impacts.
## Enhancement Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Update the plotting R-scripts to handle output from different versions of MET. - ## Describe the Enhancement ##
This issue is based on a request that came in via met-help:
https://rt.rap.ucar.edu/rt/Ticket/Display.html?id=98517
I actually already posted updates for plot_cnt.R and plot_mpr.R prior to realizing that these scripts actually do live in the MET repository. So this task is to make the same updates in the develop branch. Also, check if the same changes are needed in any of the other R-scripts. The change is this...
- add a -met_base command line argument
- if not specified there, check for a MET_BASE environment variable
- if neither are set, error out
- read the data and extract the MET version number from the first column of the first line
- read the corresponding header for that line type/version number from MET_BASE/table_files/met_header_columns_vX.Y.txt
- remove the hard-coded headers from the top of these scripts
### Time Estimate ###
4 hours.
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
No sub-issues required.
### Relevant Deadlines ###
None.
### Funding Source ###
?
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required: John HG
- [x] Select **scientist(s)** or **no scientist** required: no scientist required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [x] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
No impacts.
## Enhancement Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | update the plotting r scripts to handle output from different versions of met describe the enhancement this issue is based on a request that came in via met help i actually already posted updates for plot cnt r and plot mpr r prior to realizing that these scripts actually do live in the met repository so this task is to make the same updates in the develop branch also check if the same changes are needed in any of the other r scripts the change is this add a met base command line argument if not specified there check for a met base environment variable if neither are set error out read the data and extract the met version number from the first column of the first line read the corresponding header for that line type version number from met base table files met header columns vx y txt remove the hard coded headers from the top of these scripts time estimate hours sub issues consider breaking the enhancement down into sub issues no sub issues required relevant deadlines none funding source define the metadata assignee select engineer s or no engineer required john hg select scientist s or no scientist required no scientist required labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones or add alert need project assignment label select milestone to next major version milestone or future versions define related issue s consider the impact to the other metplus components no impacts enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project s milestone and linked issues iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
239,845 | 7,800,089,054 | IssuesEvent | 2018-06-09 04:37:32 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | USER ISSUE: Currency Exchange as World Leader | Medium Priority | **Version:** 0.7.2.5 beta
**Steps to Reproduce:**
1. Become World Leader
2. Make a Mint.
3. Make a Currency Exchange.
4. Set all the exchange rates.
5. Now, as a world leader, try to exchange some currency.
**Expected behavior:**
There should be a way for the world leader to exchange currencies they aquired earlier in the game.
**Actual behavior:**
The world leader should be able to exchange currencies like all the other players. Although I can see potential abuse with this... Perhaps a way to lock the rates by voting? | 1.0 | USER ISSUE: Currency Exchange as World Leader - **Version:** 0.7.2.5 beta
**Steps to Reproduce:**
1. Become World Leader
2. Make a Mint.
3. Make a Currency Exchange.
4. Set all the exchange rates.
5. Now, as a world leader, try to exchange some currency.
**Expected behavior:**
There should be a way for the world leader to exchange currencies they aquired earlier in the game.
**Actual behavior:**
The world leader should be able to exchange currencies like all the other players. Although I can see potential abuse with this... Perhaps a way to lock the rates by voting? | priority | user issue currency exchange as world leader version beta steps to reproduce become world leader make a mint make a currency exchange set all the exchange rates now as a world leader try to exchange some currency expected behavior there should be a way for the world leader to exchange currencies they aquired earlier in the game actual behavior the world leader should be able to exchange currencies like all the other players although i can see potential abuse with this perhaps a way to lock the rates by voting | 1 |
54,907 | 3,071,589,913 | IssuesEvent | 2015-08-19 13:03:33 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | HorizontalScrollView scroll to right | bug imported invalid Priority-Medium | _From [malliknm...@gmail.com](https://code.google.com/u/106657980405433549546/) on August 30, 2010 11:27:54_
What steps will reproduce the problem? 1.Create a HorizontalScrollView and views like buttons which spance more than screen size
2.From solo getViews and select the invisible button of the HorizontalScrollView
3.And do clickOnView(hiddenButton) expected is action from button
but there was no action performed What is the expected output? What do you see instead? Hidden buttons click action has to be performed. I can see the button from getViews but click dont give any output. What version of the product are you using? On what operating system? 1.7 API, Android 2.2, windows XP, on Droid device Please provide any additional information below.
_Original issue: http://code.google.com/p/robotium/issues/detail?id=15_ | 1.0 | HorizontalScrollView scroll to right - _From [malliknm...@gmail.com](https://code.google.com/u/106657980405433549546/) on August 30, 2010 11:27:54_
What steps will reproduce the problem? 1.Create a HorizontalScrollView and views like buttons which spance more than screen size
2.From solo getViews and select the invisible button of the HorizontalScrollView
3.And do clickOnView(hiddenButton) expected is action from button
but there was no action performed What is the expected output? What do you see instead? Hidden buttons click action has to be performed. I can see the button from getViews but click dont give any output. What version of the product are you using? On what operating system? 1.7 API, Android 2.2, windows XP, on Droid device Please provide any additional information below.
_Original issue: http://code.google.com/p/robotium/issues/detail?id=15_ | priority | horizontalscrollview scroll to right from on august what steps will reproduce the problem create a horizontalscrollview and views like buttons which spance more than screen size from solo getviews and select the invisible button of the horizontalscrollview and do clickonview hiddenbutton expected is action from button but there was no action performed what is the expected output what do you see instead hidden buttons click action has to be performed i can see the button from getviews but click dont give any output what version of the product are you using on what operating system api android windows xp on droid device please provide any additional information below original issue | 1 |
213,754 | 7,255,919,752 | IssuesEvent | 2018-02-16 16:22:00 | Z3r0byte/Magistify | https://api.github.com/repos/Z3r0byte/Magistify | closed | Notification when homework gets modified | Medium Priority enhancement | Get a notification when homework is changed, not only when there is new homework. | 1.0 | Notification when homework gets modified - Get a notification when homework is changed, not only when there is new homework. | priority | notification when homework gets modified get a notification when homework is changed not only when there is new homework | 1 |
795,611 | 28,079,067,750 | IssuesEvent | 2023-03-30 03:57:43 | AY2223S2-CS2103T-T13-2/tp | https://api.github.com/repos/AY2223S2-CS2103T-T13-2/tp | closed | V1.3: Refactor regex and parsing pipeline | type.Chore priority.Medium | Here is a sequence diagram of how a single `RecipePortion` object is parsed:

This one sequence involves repeated calls to `RecipePortion::isValidRecipePortion`, multiple parses of the same string, and may throw different exceptions at different stages of the parsing process.
To do:
- [ ] Refactor regex to parse using capture groups instead of string splitting
- [ ] Improve parsing pipeline to remove redundant calls | 1.0 | V1.3: Refactor regex and parsing pipeline - Here is a sequence diagram of how a single `RecipePortion` object is parsed:

This one sequence involves repeated calls to `RecipePortion::isValidRecipePortion`, multiple parses of the same string, and may throw different exceptions at different stages of the parsing process.
To do:
- [ ] Refactor regex to parse using capture groups instead of string splitting
- [ ] Improve parsing pipeline to remove redundant calls | priority | refactor regex and parsing pipeline here is a sequence diagram of how a single recipeportion object is parsed this one sequence involves repeated calls to recipeportion isvalidrecipeportion multiple parses of the same string and may throw different exceptions at different stages of the parsing process to do refactor regex to parse using capture groups instead of string splitting improve parsing pipeline to remove redundant calls | 1 |
108,116 | 4,327,503,089 | IssuesEvent | 2016-07-26 10:46:12 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | closed | Nodes on ways: "double selection" to get to tags editor | Enhancement Medium Priority | The described process can be seen more clearly here:
http://bjohas.de/wiki/Vespucci/selection_and_creation
1. Find a way and look at a node on it.
2. Now tap (short tap) on the node. You'll see the "selection" popup
3. Choose the node and node is selected.
4. Tap the node again, and the selector appears again
5. If you select the node, you'll get to the tag editor.
Step 4, with the selector re-appearing is potentially confusing. Moreover, if in step 4, you select the way, the way will be selected. You can the tap the node (way stays selected), then tap the way again, and you'll get to the tag editor for the way.
My suggestion is to either:
(A) Change the text in the 1st/2nd popup (steps 2 or 4), so that the user knows it's different.
E.g. in Step 2, it could say "Select one of the objects in this location:".
In Step 4, it could say: "Re-select or edit tags..."
(B) Alternatively re-implement this as the following rule: "Tapping a selected object, opens the tag editor for that object." This would mean that to re-select, one would have to tap elsewhere to remove the selection.
Thanks! | 1.0 | Nodes on ways: "double selection" to get to tags editor - The described process can be seen more clearly here:
http://bjohas.de/wiki/Vespucci/selection_and_creation
1. Find a way and look at a node on it.
2. Now tap (short tap) on the node. You'll see the "selection" popup
3. Choose the node and node is selected.
4. Tap the node again, and the selector appears again
5. If you select the node, you'll get to the tag editor.
Step 4, with the selector re-appearing is potentially confusing. Moreover, if in step 4, you select the way, the way will be selected. You can the tap the node (way stays selected), then tap the way again, and you'll get to the tag editor for the way.
My suggestion is to either:
(A) Change the text in the 1st/2nd popup (steps 2 or 4), so that the user knows it's different.
E.g. in Step 2, it could say "Select one of the objects in this location:".
In Step 4, it could say: "Re-select or edit tags..."
(B) Alternatively re-implement this as the following rule: "Tapping a selected object, opens the tag editor for that object." This would mean that to re-select, one would have to tap elsewhere to remove the selection.
Thanks! | priority | nodes on ways double selection to get to tags editor the described process can be seen more clearly here find a way and look at a node on it now tap short tap on the node you ll see the selection popup choose the node and node is selected tap the node again and the selector appears again if you select the node you ll get to the tag editor step with the selector re appearing is potentially confusing moreover if in step you select the way the way will be selected you can the tap the node way stays selected then tap the way again and you ll get to the tag editor for the way my suggestion is to either a change the text in the popup steps or so that the user knows it s different e g in step it could say select one of the objects in this location in step it could say re select or edit tags b alternatively re implement this as the following rule tapping a selected object opens the tag editor for that object this would mean that to re select one would have to tap elsewhere to remove the selection thanks | 1 |
432,120 | 12,489,303,341 | IssuesEvent | 2020-05-31 18:10:52 | OrangeJuice7/SDL-OpenGL-Game-Framework | https://api.github.com/repos/OrangeJuice7/SDL-OpenGL-Game-Framework | opened | Graphic editor for Widgets | area.Widgets priority.medium work.high | Currently, widget parameters are defined in code, which is rather bulky, messy and design is by trial-and-error.
A graphic editor can help define and provide more immediate visual feedback for widgets, while being cleaner and less prone to programming errors.
The graphic editor should be an auxiliary program / dev tool (i.e. not released to the end user).
The editor should start up two windows: One to display Widgets as they would appear in-game, and one to display the explicit Widget properties (name, Widget subclass, position, alignment, sprite etc.) when that Widget is clicked.
(Editing the properties directly in the second window might be out of scope. Rather,) the editor should load and save the Widget properties into its own file (borrow the Settings code and use its own `.ini`s?), and has an option to reload Widgets when their properties are modified externally (i.e. through a text editor). It may also allow Widgets to be resized by dragging on their edges.
Widgets can have their own text name. Used internally by the Scene to map to specific Widgets to manipulate in certain functions.
Widgets need particular functions as well (click, draw etc.). These functions should be defined in a .cpp file (or maybe .hpp file?) (that also #includes all the dependencies like the UiManager, Widgets and Model), and referred to by their symbolic name in the editor. They don't need to be functional in the editor. (They might need to be methods of the Scene...)
On compilation of the actual game, a Scene will ask for a specific set of Widgets from the editor that corresponds to that Scene. The editor translates its Widget properties into actual Widgets for the game. | 1.0 | Graphic editor for Widgets - Currently, widget parameters are defined in code, which is rather bulky, messy and design is by trial-and-error.
A graphic editor can help define and provide more immediate visual feedback for widgets, while being cleaner and less prone to programming errors.
The graphic editor should be an auxiliary program / dev tool (i.e. not released to the end user).
The editor should start up two windows: One to display Widgets as they would appear in-game, and one to display the explicit Widget properties (name, Widget subclass, position, alignment, sprite etc.) when that Widget is clicked.
(Editing the properties directly in the second window might be out of scope. Rather,) the editor should load and save the Widget properties into its own file (borrow the Settings code and use its own `.ini`s?), and has an option to reload Widgets when their properties are modified externally (i.e. through a text editor). It may also allow Widgets to be resized by dragging on their edges.
Widgets can have their own text name. Used internally by the Scene to map to specific Widgets to manipulate in certain functions.
Widgets need particular functions as well (click, draw etc.). These functions should be defined in a .cpp file (or maybe .hpp file?) (that also #includes all the dependencies like the UiManager, Widgets and Model), and referred to by their symbolic name in the editor. They don't need to be functional in the editor. (They might need to be methods of the Scene...)
On compilation of the actual game, a Scene will ask for a specific set of Widgets from the editor that corresponds to that Scene. The editor translates its Widget properties into actual Widgets for the game. | priority | graphic editor for widgets currently widget parameters are defined in code which is rather bulky messy and design is by trial and error a graphic editor can help define and provide more immediate visual feedback for widgets while being cleaner and less prone to programming errors the graphic editor should be an auxiliary program dev tool i e not released to the end user the editor should start up two windows one to display widgets as they would appear in game and one to display the explicit widget properties name widget subclass position alignment sprite etc when that widget is clicked editing the properties directly in the second window might be out of scope rather the editor should load and save the widget properties into its own file borrow the settings code and use its own ini s and has an option to reload widgets when their properties are modified externally i e through a text editor it may also allow widgets to be resized by dragging on their edges widgets can have their own text name used internally by the scene to map to specific widgets to manipulate in certain functions widgets need particular functions as well click draw etc these functions should be defined in a cpp file or maybe hpp file that also includes all the dependencies like the uimanager widgets and model and referred to by their symbolic name in the editor they don t need to be functional in the editor they might need to be methods of the scene on compilation of the actual game a scene will ask for a specific set of widgets from the editor that corresponds to that scene the editor translates its widget properties into actual widgets for the game | 1 |
804,856 | 29,504,082,103 | IssuesEvent | 2023-06-03 04:45:41 | lipskii198/XQGame- | https://api.github.com/repos/lipskii198/XQGame- | opened | Player movement | Priority: Medium | find a speed and a drag which will make the player movement feel more snappy, maybe even redo the comands | 1.0 | Player movement - find a speed and a drag which will make the player movement feel more snappy, maybe even redo the comands | priority | player movement find a speed and a drag which will make the player movement feel more snappy maybe even redo the comands | 1 |
264,507 | 8,315,771,478 | IssuesEvent | 2018-09-25 06:52:54 | FlowzPlatform/Sprint-User-Story-Board | https://api.github.com/repos/FlowzPlatform/Sprint-User-Story-Board | opened | QA - Dashboard - My Plan - UI needs to be proper on My Plan Screen. | Medium Priority New | QA - Dashboard - My Plan - UI needs to be proper on My Plan Screen.
Refer Snapshot.

| 1.0 | QA - Dashboard - My Plan - UI needs to be proper on My Plan Screen. - QA - Dashboard - My Plan - UI needs to be proper on My Plan Screen.
Refer Snapshot.

| priority | qa dashboard my plan ui needs to be proper on my plan screen qa dashboard my plan ui needs to be proper on my plan screen refer snapshot | 1 |
89,160 | 3,790,289,930 | IssuesEvent | 2016-03-21 20:55:56 | EyeSeeTea/QAApp | https://api.github.com/repos/EyeSeeTea/QAApp | closed | Drop down menus icon : people will click on the… | android buddybug complexity - med (1-5hr) priority - medium type - feature | Feedback from clussiana@psi.org : Drop down menus are not clear: people will click on the arrow instead of on 'All assessments' and 'Org units'. Can we expand the "click area" to include the drop down menu buttons?
[Link to buddybuild feedback from build 6](https://dashboard.buddybuild.com/apps/56b408ad65c5670100adf4df/feedback?fid=56dc334dba38dc0100fe7764&bnum=6)
<img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/56b408ad65c5670100adf4df/56d774842ebeba010067f8b1/fee09221-e119-4c88-a0df-c7b06b919d6c.jpg" width="33%" height="33%" /><table><tr><td>Build</td><td>6</td></tr><tr><td>Device type</td><td>SM-T531</td></tr><tr><td>Screen size</td><td>800</td></tr><tr><td>Screen size</td><td>800px by 1280px</td></tr><tr><td>Battery</td><td>39% Unplugged</td></tr><tr><td>Memory free</td><td>442 MB / 1376 MB</td></tr></table> | 1.0 | Drop down menus icon : people will click on the… - Feedback from clussiana@psi.org : Drop down menus are not clear: people will click on the arrow instead of on 'All assessments' and 'Org units'. Can we expand the "click area" to include the drop down menu buttons?
[Link to buddybuild feedback from build 6](https://dashboard.buddybuild.com/apps/56b408ad65c5670100adf4df/feedback?fid=56dc334dba38dc0100fe7764&bnum=6)
<img src="https://s3-us-west-2.amazonaws.com/buddybuild-screenshots/56b408ad65c5670100adf4df/56d774842ebeba010067f8b1/fee09221-e119-4c88-a0df-c7b06b919d6c.jpg" width="33%" height="33%" /><table><tr><td>Build</td><td>6</td></tr><tr><td>Device type</td><td>SM-T531</td></tr><tr><td>Screen size</td><td>800</td></tr><tr><td>Screen size</td><td>800px by 1280px</td></tr><tr><td>Battery</td><td>39% Unplugged</td></tr><tr><td>Memory free</td><td>442 MB / 1376 MB</td></tr></table> | priority | drop down menus icon people will click on the… feedback from clussiana psi org drop down menus are not clear people will click on the arrow instead of on all assessments and org units can we expand the click area to include the drop down menu buttons build device type sm screen size screen size by battery unplugged memory free mb mb | 1 |
449,759 | 12,974,774,886 | IssuesEvent | 2020-07-21 15:55:47 | BCcampus/edehr | https://api.github.com/repos/BCcampus/edehr | closed | Clear seed data | Epic - Form Epic - Instructor Priority - Medium | ## User story
As a seed creator user,
I want to be able to clear seed data for an entire seed, or on a page by page basis
Because it is faster and more dependable than needing to delete data on a field by field basis
## Description
Clear seed data for entire seed or on a page by page basis.
## Expected behaviour
- Clear seed data for entire seed
- Clear data functionality for each app page
- Delete documents associated with seed
**Button options:**
- Entire seed - `Delete all seed data` and `Cancel`
- Each app page - `Delete all page data` and `Cancel` - On the dialog give the user options, `Confirm for just this page` `Confirm for all pages` So that they don't have to keep hitting the confirm button.
- Documents - `Delete all seed documents` and `Cancel`
## Error states/messaging
Show the user a "are you sure?" dialog to confirm that they want to remove the seed data. Use the message:
Title - "Are you sure you want to delete this data?"
Body - "This action cannot be undone and you will be left with an empty [page/seed - depending on what they're deleting] | 1.0 | Clear seed data - ## User story
As a seed creator user,
I want to be able to clear seed data for an entire seed, or on a page by page basis
Because it is faster and more dependable than needing to delete data on a field by field basis
## Description
Clear seed data for entire seed or on a page by page basis.
## Expected behaviour
- Clear seed data for entire seed
- Clear data functionality for each app page
- Delete documents associated with seed
**Button options:**
- Entire seed - `Delete all seed data` and `Cancel`
- Each app page - `Delete all page data` and `Cancel` - On the dialog give the user options, `Confirm for just this page` `Confirm for all pages` So that they don't have to keep hitting the confirm button.
- Documents - `Delete all seed documents` and `Cancel`
## Error states/messaging
Show the user a "are you sure?" dialog to confirm that they want to remove the seed data. Use the message:
Title - "Are you sure you want to delete this data?"
Body - "This action cannot be undone and you will be left with an empty [page/seed - depending on what they're deleting] | priority | clear seed data user story as a seed creator user i want to be able to clear seed data for an entire seed or on a page by page basis because it is faster and more dependable than needing to delete data on a field by field basis description clear seed data for entire seed or on a page by page basis expected behaviour clear seed data for entire seed clear data functionality for each app page delete documents associated with seed button options entire seed delete all seed data and cancel each app page delete all page data and cancel on the dialog give the user options confirm for just this page confirm for all pages so that they don t have to keep hitting the confirm button documents delete all seed documents and cancel error states messaging show the user a are you sure dialog to confirm that they want to remove the seed data use the message title are you sure you want to delete this data body this action cannot be undone and you will be left with an empty | 1 |
617,984 | 19,411,795,564 | IssuesEvent | 2021-12-20 10:24:14 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Discussion > Pagination and post count break when you merge discussion. | bug priority: medium bug: needs validating | **Describe the bug**
When we merge any discussion that time pagination and the post count is broken.
**To Reproduce**
Steps to reproduce the behavior:
1. Set a limit to 3 replies per page. http://i.imgur.com/F8Pl4Xk.png
2. create the two different discussions with repliers.
3. try to merge the discussion.
4. See the error of the pagination count once the merge is completed.
**Expected behavior**
Pagination and pagination count should not be broken.
**Screenshots**
Dev:
https://user-images.githubusercontent.com/83228316/146747490-68638651-1f43-46c0-a7d3-1ea0f09f3f66.mp4
Customer:
https://user-images.githubusercontent.com/83228316/146751775-ad5ed1c7-5161-47a9-bc3f-1f2cd11c61e6.mp4

**Environment**
dev, local
**Support ticket links**
https://secure.helpscout.net/conversation/1701567391/174577
| 1.0 | Discussion > Pagination and post count break when you merge discussion. - **Describe the bug**
When we merge any discussion that time pagination and the post count is broken.
**To Reproduce**
Steps to reproduce the behavior:
1. Set a limit to 3 replies per page. http://i.imgur.com/F8Pl4Xk.png
2. create the two different discussions with repliers.
3. try to merge the discussion.
4. See the error of the pagination count once the merge is completed.
**Expected behavior**
Pagination and pagination count should not be broken.
**Screenshots**
Dev:
https://user-images.githubusercontent.com/83228316/146747490-68638651-1f43-46c0-a7d3-1ea0f09f3f66.mp4
Customer:
https://user-images.githubusercontent.com/83228316/146751775-ad5ed1c7-5161-47a9-bc3f-1f2cd11c61e6.mp4

**Environment**
dev, local
**Support ticket links**
https://secure.helpscout.net/conversation/1701567391/174577
| priority | discussion pagination and post count break when you merge discussion describe the bug when we merge any discussion that time pagination and the post count is broken to reproduce steps to reproduce the behavior set a limit to replies per page create the two different discussions with repliers try to merge the discussion see the error of the pagination count once the merge is completed expected behavior pagination and pagination count should not be broken screenshots dev customer environment dev local support ticket links | 1 |
473,438 | 13,642,378,253 | IssuesEvent | 2020-09-25 15:26:46 | AY2021S1-CS2103T-T17-3/tp | https://api.github.com/repos/AY2021S1-CS2103T-T17-3/tp | opened | As a new user I want to see sample data | priority.Medium type.Story | ... so that I can see how the application would work when it is in use | 1.0 | As a new user I want to see sample data - ... so that I can see how the application would work when it is in use | priority | as a new user i want to see sample data so that i can see how the application would work when it is in use | 1 |
808,350 | 30,056,805,347 | IssuesEvent | 2023-06-28 07:28:30 | devhausleipzigacademy/camp10-midterm | https://api.github.com/repos/devhausleipzigacademy/camp10-midterm | opened | Setup Routing | medium priority:high | Go through the user journey. Make notes on which routes are needed. Find out which will be static routes and which ones will be dynamic. We recommend doing this as a small team or a pair. Routes should be simple and informative, e.g.
- /login
- /movies
- /favorites
- /blahblahblah
- [ ] setup provider for react-router-dom and all needed config
- [ ] setup the layouts with proper nesting
- [ ] setup route definitions (be mindful about dynamic routing)
- [ ] create empty pages for each route and render "Hello ${route}" on each page
| 1.0 | Setup Routing - Go through the user journey. Make notes on which routes are needed. Find out which will be static routes and which ones will be dynamic. We recommend doing this as a small team or a pair. Routes should be simple and informative, e.g.
- /login
- /movies
- /favorites
- /blahblahblah
- [ ] setup provider for react-router-dom and all needed config
- [ ] setup the layouts with proper nesting
- [ ] setup route definitions (be mindful about dynamic routing)
- [ ] create empty pages for each route and render "Hello ${route}" on each page
| priority | setup routing go through the user journey make notes on which routes are needed find out which will be static routes and which ones will be dynamic we recommend doing this as a small team or a pair routes should be simple and informative e g login movies favorites blahblahblah setup provider for react router dom and all needed config setup the layouts with proper nesting setup route definitions be mindful about dynamic routing create empty pages for each route and render hello route on each page | 1 |
28,141 | 2,699,985,672 | IssuesEvent | 2015-04-03 21:26:33 | starteam/starcellbio_html | https://api.github.com/repos/starteam/starcellbio_html | closed | Two similar javascript libraries swipe / coinslider | Medium Priority | They seem to do the same thing (making sliders)
https://github.com/starteam/starcellbio_html/tree/develop/html_app
- Are they both in use ?
- Can we just use 1 of them
- Only keep the min version of css/js files and delete the rest (http://goo.gl/S5AeDr) | 1.0 | Two similar javascript libraries swipe / coinslider - They seem to do the same thing (making sliders)
https://github.com/starteam/starcellbio_html/tree/develop/html_app
- Are they both in use ?
- Can we just use 1 of them
- Only keep the min version of css/js files and delete the rest (http://goo.gl/S5AeDr) | priority | two similar javascript libraries swipe coinslider they seem to do the same thing making sliders are they both in use can we just use of them only keep the min version of css js files and delete the rest | 1 |
717,429 | 24,674,810,430 | IssuesEvent | 2022-10-18 16:08:18 | AY2223S1-CS2103T-W15-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-W15-1/tp | closed | Implement Todo - Priority functionalities | enhancement priority.Medium | Priority
1. High
2. Medium
3. Low
Commands
1. update `todo add` command with compulsory priority field
2. update `todo edit` command with priority field
3. update `todo show` command with priority field - to show all todos of a certain priority
4. auto sort todo list by decreasing priority | 1.0 | Implement Todo - Priority functionalities - Priority
1. High
2. Medium
3. Low
Commands
1. update `todo add` command with compulsory priority field
2. update `todo edit` command with priority field
3. update `todo show` command with priority field - to show all todos of a certain priority
4. auto sort todo list by decreasing priority | priority | implement todo priority functionalities priority high medium low commands update todo add command with compulsory priority field update todo edit command with priority field update todo show command with priority field to show all todos of a certain priority auto sort todo list by decreasing priority | 1 |
621,142 | 19,578,751,265 | IssuesEvent | 2022-01-04 18:19:24 | E3SM-Project/scream | https://api.github.com/repos/E3SM-Project/scream | closed | [AD] create a new field manager registration for "internal" variables | enhancement Atmosphere Driver Development Task priority:medium | There are a handful of variables in SCREAM that must be in the field manager because they are both a) persistent across time steps and b) needed for BFB restarts. However, they are only needed by one process and should not be changed by any other process.
An example of this sort of variable would be `T_prev_micro_step` which is needed by P3 microphysics, but nowhere else.
After some discussion with @bartgol we decided there was a need for a new class of field managed variable, other than the "computed" and "required" registrations we have now. A possible name might be "internal" or "private" or "local".
By registering a field managed variable as "internal" we can do a number of things.
1. We can check that no other process is registering a variable with the same name, to protect against inadvertent changing of local persistent variables.
2. We can treat these differently at initialization. Note this was the primary motivation for the correspondence between @AaronDonahue and @bartgol. In other words, instead of requiring internal variables to be initialized by the AD, we can have the FM set the variable to NaN's and allow the init_impl of the process itself to set the value.
- Note, we could add even more functionality, such as checking to see if initialization is possible using the init netCDF file, but not throwing an error if it isn't found. This would be required for restart runs.
| 1.0 | [AD] create a new field manager registration for "internal" variables - There are a handful of variables in SCREAM that must be in the field manager because they are both a) persistent across time steps and b) needed for BFB restarts. However, they are only needed by one process and should not be changed by any other process.
An example of this sort of variable would be `T_prev_micro_step` which is needed by P3 microphysics, but nowhere else.
After some discussion with @bartgol we decided there was a need for a new class of field managed variable, other than the "computed" and "required" registrations we have now. A possible name might be "internal" or "private" or "local".
By registering a field managed variable as "internal" we can do a number of things.
1. We can check that no other process is registering a variable with the same name, to protect against inadvertent changing of local persistent variables.
2. We can treat these differently at initialization. Note this was the primary motivation for the correspondence between @AaronDonahue and @bartgol. In other words, instead of requiring internal variables to be initialized by the AD, we can have the FM set the variable to NaN's and allow the init_impl of the process itself to set the value.
- Note, we could add even more functionality, such as checking to see if initialization is possible using the init netCDF file, but not throwing an error if it isn't found. This would be required for restart runs.
| priority | create a new field manager registration for internal variables there are a handful of variables in scream that must be in the field manager because they are both a persistent across time steps and b needed for bfb restarts however they are only needed by one process and should not be changed by any other process an example of this sort of variable would be t prev micro step which is needed by microphysics but nowhere else after some discussion with bartgol we decided there was a need for a new class of field managed variable other than the computed and required registrations we have now a possible name might be internal or private or local by registering a field managed variable as internal we can do a number of things we can check that no other process is registering a variable with the same name to protect against inadvertent changing of local persistent variables we can treat these differently at initialization note this was the primary motivation for the correspondence between aarondonahue and bartgol in other words instead of requiring internal variables to be initialized by the ad we can have the fm set the variable to nan s and allow the init impl of the process itself to set the value note we could add even more functionality such as checking to see if initialization is possible using the init netcdf file but not throwing an error if it isn t found this would be required for restart runs | 1 |
816,541 | 30,602,110,199 | IssuesEvent | 2023-07-22 14:17:22 | codersforcauses/repair-labs | https://api.github.com/repos/codersforcauses/repair-labs | closed | Generate seed data with faker | backend difficulty:easy priority:medium | ## Basic Information
- Automatically generate data in seed.ts with faker
## Other Information
- Make sure the random data created is always from the same starting seed to ensure the data is the same after each run | 1.0 | Generate seed data with faker - ## Basic Information
- Automatically generate data in seed.ts with faker
## Other Information
- Make sure the random data created is always from the same starting seed to ensure the data is the same after each run | priority | generate seed data with faker basic information automatically generate data in seed ts with faker other information make sure the random data created is always from the same starting seed to ensure the data is the same after each run | 1 |
585,580 | 17,500,863,769 | IssuesEvent | 2021-08-10 09:14:09 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | opened | [Челенджі. Навчай українською] "Допомогти проєкту" button and social contacts are not present below the banner | bug UI Priority: Medium | **Environment:** Windows 10, version 92.0.4515.107, (64)
**Reproducible:** always
**Build found:** last commit
**Steps to reproduce**
Go to https://speak-ukrainian.org.ua/dev/
Go to Tab 'Челендж' > 'Навчай українською'
**Actual result**
- Social media links are not present below the banner.
- "Допомогти проєкту" button is not present.

**Expected result**
- Social media links are located below the banner.
- "Допомогти проєкту" button is present.

| 1.0 | [Челенджі. Навчай українською] "Допомогти проєкту" button and social contacts are not present below the banner - **Environment:** Windows 10, version 92.0.4515.107, (64)
**Reproducible:** always
**Build found:** last commit
**Steps to reproduce**
Go to https://speak-ukrainian.org.ua/dev/
Go to Tab 'Челендж' > 'Навчай українською'
**Actual result**
- Social media links are not present below the banner.
- "Допомогти проєкту" button is not present.

**Expected result**
- Social media links are located below the banner.
- "Допомогти проєкту" button is present.

| priority | допомогти проєкту button and social contacts are not present below the banner environment windows version reproducible always build found last commit steps to reproduce go to go to tab челендж навчай українською actual result social media links are not present below the banner допомогти проєкту button is not present expected result social media links are located below the banner допомогти проєкту button is present | 1 |
480,719 | 13,865,839,928 | IssuesEvent | 2020-10-16 05:29:22 | ethereum/sourcify | https://api.github.com/repos/ethereum/sourcify | closed | Separate staging environment | medium priority | Deploy the staging environment to another server to reduce the load on the main one.
It will also be easier to test new features that way.
Create a CI to do that automatically. | 1.0 | Separate staging environment - Deploy the staging environment to another server to reduce the load on the main one.
It will also be easier to test new features that way.
Create a CI to do that automatically. | priority | separate staging environment deploy the staging environment to another server to reduce the load on the main one it will also be easier to test new features that way create a ci to do that automatically | 1 |
159,953 | 6,065,095,527 | IssuesEvent | 2017-06-14 15:32:51 | saulgreenberg/Timelapse | https://api.github.com/repos/saulgreenberg/Timelapse | closed | When .tdb and .ddb templates don't match, Timelapse should ask the user to resolve the differences | enhancement medium priority | Some cases:
- both match except there is a new column in the template .tdb
- both match except there is a deleted column in the template .tdb
- columns added / deleted in both, which may also be a renamed column | 1.0 | When .tdb and .ddb templates don't match, Timelapse should ask the user to resolve the differences - Some cases:
- both match except there is a new column in the template .tdb
- both match except there is a deleted column in the template .tdb
- columns added / deleted in both, which may also be a renamed column | priority | when tdb and ddb templates don t match timelapse should ask the user to resolve the differences some cases both match except there is a new column in the template tdb both match except there is a deleted column in the template tdb columns added deleted in both which may also be a renamed column | 1 |
15,484 | 2,611,485,784 | IssuesEvent | 2015-02-27 05:27:13 | chrsmith/switchlist | https://api.github.com/repos/chrsmith/switchlist | closed | Support multi-page switchlists in general. | auto-migrated Priority-Medium Type-Enhancement | ```
Some switchlists (SP-style and B-7) support multi-page. That code should be
generalized so any multi-page form style switchlist can be printed.
```
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 24 Apr 2011 at 5:27 | 1.0 | Support multi-page switchlists in general. - ```
Some switchlists (SP-style and B-7) support multi-page. That code should be
generalized so any multi-page form style switchlist can be printed.
```
Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 24 Apr 2011 at 5:27 | priority | support multi page switchlists in general some switchlists sp style and b support multi page that code should be generalized so any multi page form style switchlist can be printed original issue reported on code google com by rwbowdi gmail com on apr at | 1 |
459,648 | 13,196,421,277 | IssuesEvent | 2020-08-13 20:37:09 | Energy-Innovation/eps-us | https://api.github.com/repos/Energy-Innovation/eps-us | closed | Data Logging Script adding Time rows within a single variable output | medium priority | The data logging script appears to be adding some extra Time lines for certain variables when saving out data. See attached for an example
[DataLoggingExample.xlsx](https://github.com/Energy-Innovation/eps-us/files/5065996/DataLoggingExample.xlsx)
| 1.0 | Data Logging Script adding Time rows within a single variable output - The data logging script appears to be adding some extra Time lines for certain variables when saving out data. See attached for an example
[DataLoggingExample.xlsx](https://github.com/Energy-Innovation/eps-us/files/5065996/DataLoggingExample.xlsx)
| priority | data logging script adding time rows within a single variable output the data logging script appears to be adding some extra time lines for certain variables when saving out data see attached for an example | 1 |
293,178 | 8,973,209,101 | IssuesEvent | 2019-01-29 20:22:30 | CameraKit/camerakit-android | https://api.github.com/repos/CameraKit/camerakit-android | closed | [Feature] Check if Flash hardware exists on device | Priority: Medium Status: Pending Type: Enhancement | This is a Request For Enhancement (RFE), rather than a bug report.
I'm finding a lot of devices don't have a flash.
It would be very handy if I could find out if i should show the flash button with a simple call to `<cameraview>.hasFlash()`.
Some information about how this is done can be found in this Stack Overflow link.
https://stackoverflow.com/questions/13413938/how-to-check-if-device-has-flash-light-led-android
At this point, the codebase seems to be a little out of sync, with a pending 1.0.0 merge request, so I don't want to add another merge request on top of that.
However, if the main team can get the merge requests cleared up, I don't mind putting in the work, to get this RFE done. | 1.0 | [Feature] Check if Flash hardware exists on device - This is a Request For Enhancement (RFE), rather than a bug report.
I'm finding a lot of devices don't have a flash.
It would be very handy if I could find out if i should show the flash button with a simple call to `<cameraview>.hasFlash()`.
Some information about how this is done can be found in this Stack Overflow link.
https://stackoverflow.com/questions/13413938/how-to-check-if-device-has-flash-light-led-android
At this point, the codebase seems to be a little out of sync, with a pending 1.0.0 merge request, so I don't want to add another merge request on top of that.
However, if the main team can get the merge requests cleared up, I don't mind putting in the work, to get this RFE done. | priority | check if flash hardware exists on device this is a request for enhancement rfe rather than a bug report i m finding a lot of devices don t have a flash it would be very handy if i could find out if i should show the flash button with a simple call to hasflash some information about how this is done can be found in this stack overflow link at this point the codebase seems to be a little out of sync with a pending merge request so i don t want to add another merge request on top of that however if the main team can get the merge requests cleared up i don t mind putting in the work to get this rfe done | 1 |
674,714 | 23,063,637,029 | IssuesEvent | 2022-07-25 12:09:53 | bitsongofficial/wallet-mobile | https://api.github.com/repos/bitsongofficial/wallet-mobile | opened | Disable Screenshot ability | bug medium priority | As per a home banking application, since there is a lot of sensible data, we need to disable the ability to take screenshots on the mobile app.
| 1.0 | Disable Screenshot ability - As per a home banking application, since there is a lot of sensible data, we need to disable the ability to take screenshots on the mobile app.
| priority | disable screenshot ability as per a home banking application since there is a lot of sensible data we need to disable the ability to take screenshots on the mobile app | 1 |
56,932 | 3,081,219,180 | IssuesEvent | 2015-08-22 14:04:06 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Bot body shapes | 016a enhancement imported Priority-Medium | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on January 29, 2012 10:50:33_
Let bots pick from one of several body shapes; see examples created by blackbird and/or little apple in the forums
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=158_ | 1.0 | Bot body shapes - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on January 29, 2012 10:50:33_
Let bots pick from one of several body shapes; see examples created by blackbird and/or little apple in the forums
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=158_ | priority | bot body shapes from on january let bots pick from one of several body shapes see examples created by blackbird and or little apple in the forums original issue | 1 |
213,986 | 7,262,377,673 | IssuesEvent | 2018-02-19 05:32:28 | joshherkness/AR-Top | https://api.github.com/repos/joshherkness/AR-Top | closed | Deploy Unity App to Mobile | Effort: 4 Priority: Medium | Getting a start on Deploying the Unity App to Android Phones.
Note that iOS phones require a Mac with Xcode v7.3 or greater to deploy to them. | 1.0 | Deploy Unity App to Mobile - Getting a start on Deploying the Unity App to Android Phones.
Note that iOS phones require a Mac with Xcode v7.3 or greater to deploy to them. | priority | deploy unity app to mobile getting a start on deploying the unity app to android phones note that ios phones require a mac with xcode or greater to deploy to them | 1 |
613,069 | 19,072,410,936 | IssuesEvent | 2021-11-27 05:42:36 | CMPUT301F21T21/detes | https://api.github.com/repos/CMPUT301F21T21/detes | closed | US 02.03.02 Habit Events | Clarification needed High Risk Medium Priority New | As a doer, I want the photograph for a habit event to be taken with the camera within the app.
Story Points: 3.5 | 1.0 | US 02.03.02 Habit Events - As a doer, I want the photograph for a habit event to be taken with the camera within the app.
Story Points: 3.5 | priority | us habit events as a doer i want the photograph for a habit event to be taken with the camera within the app story points | 1 |
259,817 | 8,200,328,300 | IssuesEvent | 2018-09-01 02:38:41 | LessWrong2/Lesswrong2 | https://api.github.com/repos/LessWrong2/Lesswrong2 | closed | Remove URL from the user's profile and replace bio with ory-editor content field | 2. Medium Priority (Hard) 7. Feature | The URL field strikes me as fairly useless, and I don't think it make super much sense for us to keep it. But I am hesitant to just remove it since we don't yet have rich-text capabilities in the bio field (i.e. it's currently just plaintext). So before I remove it, I would want to make sure we switch over towards the full editor on people's bio's. | 1.0 | Remove URL from the user's profile and replace bio with ory-editor content field - The URL field strikes me as fairly useless, and I don't think it make super much sense for us to keep it. But I am hesitant to just remove it since we don't yet have rich-text capabilities in the bio field (i.e. it's currently just plaintext). So before I remove it, I would want to make sure we switch over towards the full editor on people's bio's. | priority | remove url from the user s profile and replace bio with ory editor content field the url field strikes me as fairly useless and i don t think it make super much sense for us to keep it but i am hesitant to just remove it since we don t yet have rich text capabilities in the bio field i e it s currently just plaintext so before i remove it i would want to make sure we switch over towards the full editor on people s bio s | 1 |
622,916 | 19,658,816,825 | IssuesEvent | 2022-01-10 15:07:09 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | closed | The avatar of an invited organizer is displayed on the header of the group before he accepts to join. | bug priority-medium t3-needs-review Stale | **Describe the bug**
When you invite a member to a group and promote him before he accepts the request to join, his avatar is already displayed in the header of the group page (in the "organizers" section)
In the example below, I've invited Ivan to join the group and promoted him as organizer. You can see that there's only one member (= me) but there're 2 avatars on the header (Organizers (2)).

**To Reproduce**
1. Create a new group
2. Invite some members
3. Promote one of the member to organizer
4. Go to the group
EDIT : I do this using REST API calls.
**Expected behavior**
Do not show "pending" organizers
**Screenshots**
above
**Support ticket links**
none
**Jira issue** : [PROD-846]
[PROD-846]: https://buddyboss.atlassian.net/browse/PROD-846?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | The avatar of an invited organizer is displayed on the header of the group before he accepts to join. - **Describe the bug**
When you invite a member to a group and promote him before he accepts the request to join, his avatar is already displayed in the header of the group page (in the "organizers" section)
In the example below, I've invited Ivan to join the group and promoted him as organizer. You can see that there's only one member (= me) but there're 2 avatars on the header (Organizers (2)).

**To Reproduce**
1. Create a new group
2. Invite some members
3. Promote one of the member to organizer
4. Go to the group
EDIT : I do this using REST API calls.
**Expected behavior**
Do not show "pending" organizers
**Screenshots**
above
**Support ticket links**
none
**Jira issue** : [PROD-846]
[PROD-846]: https://buddyboss.atlassian.net/browse/PROD-846?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | the avatar of an invited organizer is displayed on the header of the group before he accepts to join describe the bug when you invite a member to a group and promote him before he accepts the request to join his avatar is already displayed in the header of the group page in the organizers section in the example below i ve invited ivan to join the group and promoted him as organizer you can see that there s only one member me but there re avatars on the header organizers to reproduce create a new group invite some members promote one of the member to organizer go to the group edit i do this using rest api calls expected behavior do not show pending organizers screenshots above support ticket links none jira issue | 1 |
13,727 | 2,610,295,466 | IssuesEvent | 2015-02-26 19:34:57 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | opened | The "Quality" presets in versions 0.9.14+ are suboptimal | auto-migrated Priority-Medium Type-Enhancement | ```
What steps will reproduce the problem?
1. Run Hedgewars on a computer with an intergated Intel GMA
2. Set to any quality other than 0 (lowest)
3. Experience extreme lag (~6 fps)
What is the expected output? What do you see instead?
Major improvements can be gained from setting rqClampLess, rqNoBackground, and
rqSlowMenu
Unfortunately, the only way to change them is by editing the source code, so
finer-grained controls are needed for those of us using lower-end GPUs. (I am
using Intel Core i7 / Arrendale)
What version of the product are you using? On what operating system?
Hedgewars 0.9.15 on Linux 2.6.37
Please provide any additional information below.
Unfortunately, it's not as easy to fix this as just adding a bunch of options
for the users to set, since I've experienced engine crashes under with some
combinations.
The configuration I use - which gives me a steady 30 fps is:
rqClampLess | rqNoBackground | rqSlowMenu | rqKillFlakes | rqPlainSplash |
rq2DWater | rqFancyBoom
```
-----
Original issue reported on code.google.com by `Ale...@gmail.com` on 10 Jan 2011 at 2:59
* Blocking: #822 | 1.0 | The "Quality" presets in versions 0.9.14+ are suboptimal - ```
What steps will reproduce the problem?
1. Run Hedgewars on a computer with an intergated Intel GMA
2. Set to any quality other than 0 (lowest)
3. Experience extreme lag (~6 fps)
What is the expected output? What do you see instead?
Major improvements can be gained from setting rqClampLess, rqNoBackground, and
rqSlowMenu
Unfortunately, the only way to change them is by editing the source code, so
finer-grained controls are needed for those of us using lower-end GPUs. (I am
using Intel Core i7 / Arrendale)
What version of the product are you using? On what operating system?
Hedgewars 0.9.15 on Linux 2.6.37
Please provide any additional information below.
Unfortunately, it's not as easy to fix this as just adding a bunch of options
for the users to set, since I've experienced engine crashes under with some
combinations.
The configuration I use - which gives me a steady 30 fps is:
rqClampLess | rqNoBackground | rqSlowMenu | rqKillFlakes | rqPlainSplash |
rq2DWater | rqFancyBoom
```
-----
Original issue reported on code.google.com by `Ale...@gmail.com` on 10 Jan 2011 at 2:59
* Blocking: #822 | priority | the quality presets in versions are suboptimal what steps will reproduce the problem run hedgewars on a computer with an intergated intel gma set to any quality other than lowest experience extreme lag fps what is the expected output what do you see instead major improvements can be gained from setting rqclampless rqnobackground and rqslowmenu unfortunately the only way to change them is by editing the source code so finer grained controls are needed for those of us using lower end gpus i am using intel core arrendale what version of the product are you using on what operating system hedgewars on linux please provide any additional information below unfortunately it s not as easy to fix this as just adding a bunch of options for the users to set since i ve experienced engine crashes under with some combinations the configuration i use which gives me a steady fps is rqclampless rqnobackground rqslowmenu rqkillflakes rqplainsplash rqfancyboom original issue reported on code google com by ale gmail com on jan at blocking | 1 |
604,758 | 18,718,233,360 | IssuesEvent | 2021-11-03 08:45:33 | kheeyaa/Who-ate-my-fish | https://api.github.com/repos/kheeyaa/Who-ate-my-fish | closed | JailUsers 비활성화 로직 구현 | enhancement priority: medium status: pending | - [x] 내가 죽은 고양이인지 확인
- [x] 입력창, 투표창 비활성화
- [x] 프로필 감옥으로 변경
- [x] 다른 유저들에게도 감옥 고양이로 변경
- [x] 투표시 감옥 고양이 선택 못하게 변경 | 1.0 | JailUsers 비활성화 로직 구현 - - [x] 내가 죽은 고양이인지 확인
- [x] 입력창, 투표창 비활성화
- [x] 프로필 감옥으로 변경
- [x] 다른 유저들에게도 감옥 고양이로 변경
- [x] 투표시 감옥 고양이 선택 못하게 변경 | priority | jailusers 비활성화 로직 구현 내가 죽은 고양이인지 확인 입력창 투표창 비활성화 프로필 감옥으로 변경 다른 유저들에게도 감옥 고양이로 변경 투표시 감옥 고양이 선택 못하게 변경 | 1 |
314,053 | 9,592,338,125 | IssuesEvent | 2019-05-09 08:41:48 | LifeMC/LifeSkript | https://api.github.com/repos/LifeMC/LifeSkript | opened | Fix eclipse compatibility with Kotlin | bug enhancement medium priority | **Describe the bug**
Kotlin classes currently gives error on Eclipse.
**To Reproduce**
Open the Project with Eclipse.
**Expected behavior**
Since we are supporting both IDEs, Eclipse should _not_ give errors or warnings.
**Screenshots**
N/A
**Version information:**
- Minecraft Version: N/A
- Skript Version: 2.2-V13b (dev-build)
**Additional information**
N/A
**Possible solutions**
Add Kotlin nature to Project in Eclipse.
| 1.0 | Fix eclipse compatibility with Kotlin - **Describe the bug**
Kotlin classes currently gives error on Eclipse.
**To Reproduce**
Open the Project with Eclipse.
**Expected behavior**
Since we are supporting both IDEs, Eclipse should _not_ give errors or warnings.
**Screenshots**
N/A
**Version information:**
- Minecraft Version: N/A
- Skript Version: 2.2-V13b (dev-build)
**Additional information**
N/A
**Possible solutions**
Add Kotlin nature to Project in Eclipse.
| priority | fix eclipse compatibility with kotlin describe the bug kotlin classes currently gives error on eclipse to reproduce open the project with eclipse expected behavior since we are supporting both ides eclipse should not give errors or warnings screenshots n a version information minecraft version n a skript version dev build additional information n a possible solutions add kotlin nature to project in eclipse | 1 |
603,665 | 18,670,417,889 | IssuesEvent | 2021-10-30 15:52:59 | AY2122S1-CS2113T-T10-3/tp | https://api.github.com/repos/AY2122S1-CS2113T-T10-3/tp | closed | If file i/o error occurs during recipe creation/push/pop, only the error message is printed | type.Bug priority.Medium | Steps to reproduce
1) Have a directory called `recipe`.
2) Create a **folder** called `recipe.txt` in the `recipe` directory.
3) Run the program.
4) Enter `recipe new recipe`.
Actual result:

Expected result:

Cause:
Because when a file i/o error occurs during the `recipe new, recipe push etc` command, the program immediately returns to `Decodex`, therefore messages for the commands are not printed and the recipe is not getting selected even though the changes to `RecipeManager` are indeed applied. | 1.0 | If file i/o error occurs during recipe creation/push/pop, only the error message is printed - Steps to reproduce
1) Have a directory called `recipe`.
2) Create a **folder** called `recipe.txt` in the `recipe` directory.
3) Run the program.
4) Enter `recipe new recipe`.
Actual result:

Expected result:

Cause:
Because when a file i/o error occurs during the `recipe new, recipe push etc` command, the program immediately returns to `Decodex`, therefore messages for the commands are not printed and the recipe is not getting selected even though the changes to `RecipeManager` are indeed applied. | priority | if file i o error occurs during recipe creation push pop only the error message is printed steps to reproduce have a directory called recipe create a folder called recipe txt in the recipe directory run the program enter recipe new recipe actual result expected result cause because when a file i o error occurs during the recipe new recipe push etc command the program immediately returns to decodex therefore messages for the commands are not printed and the recipe is not getting selected even though the changes to recipemanager are indeed applied | 1 |
256,869 | 8,129,937,503 | IssuesEvent | 2018-08-17 16:39:50 | exercism/exercism | https://api.github.com/repos/exercism/exercism | closed | Some Profile Tags Missing | priority/medium type/bug | On the profiles page (https://exercism.io/profiles/jpreese), it does not look like all of the tags are there. A couple examples:
Mentors page itself:

My profile:

In this case, I'm missing the F# (and Go) mentor tag.
It also doesn't look like maintainers are given tags at all (e.g. "C# Maintainer). I'm not sure if that's intended or not. | 1.0 | Some Profile Tags Missing - On the profiles page (https://exercism.io/profiles/jpreese), it does not look like all of the tags are there. A couple examples:
Mentors page itself:

My profile:

In this case, I'm missing the F# (and Go) mentor tag.
It also doesn't look like maintainers are given tags at all (e.g. "C# Maintainer). I'm not sure if that's intended or not. | priority | some profile tags missing on the profiles page it does not look like all of the tags are there a couple examples mentors page itself my profile in this case i m missing the f and go mentor tag it also doesn t look like maintainers are given tags at all e g c maintainer i m not sure if that s intended or not | 1 |
198,099 | 6,969,880,081 | IssuesEvent | 2017-12-11 08:06:11 | pmem/issues | https://api.github.com/repos/pmem/issues | closed | obj ctl: allocation from custom class fails with ENOMEM when did after alloc&free from class of id 0 | Exposure: Medium OS: Linux Priority: 3 medium Type: Bug | Steps to reproduce:
```
1. Create obj pool
2.
PMEMoid oid = OID_NULL;
pobj_alloc_class_desc read_arg;
// allocate from class of id 0
int ret = pmemobj_ctl_get(pop, "heap.alloc_class.0.desc", &read_arg);
assert(ret == 0);
ret = pmemobj_xalloc(pop, &oid, read_arg.unit_size, 0, POBJ_CLASS_ID(0), nullptr, nullptr);
assert(ret == 0);
// release object
pmemobj_free(&oid);
// create custom allocation class
pobj_alloc_class_desc write_arg;
write_arg.unit_size = 1024;
write_arg.units_per_block = 1024;
write_arg.header_type = POBJ_HEADER_COMPACT;
ret = pmemobj_ctl_set(pop, "heap.alloc_class.new.desc", &write_arg);
assert(ret == 0)
string entry_point = "heap.alloc_class." + to_string(write_arg.class_id) + ".desc";
ret = pmemobj_ctl_get(pop, entry_point.c_str(), &read_arg));
assert(ret == 0)
//allocate from custom allocation class
ret = pmemobj_xalloc(pop, &oid, read_arg.unit_size - 16, 0,
POBJ_CLASS_ID(write_arg.class_id), nullptr, nullptr)
assert(ret == 0)
```
Actual result:
- Last assertion was not satisfied, ret = -1
- object allocation failed with ENOMEM
Expected result:
- allocation succeeded
Found in 1.3+b2-276-g335828a | 1.0 | obj ctl: allocation from custom class fails with ENOMEM when did after alloc&free from class of id 0 - Steps to reproduce:
```
1. Create obj pool
2.
PMEMoid oid = OID_NULL;
pobj_alloc_class_desc read_arg;
// allocate from class of id 0
int ret = pmemobj_ctl_get(pop, "heap.alloc_class.0.desc", &read_arg);
assert(ret == 0);
ret = pmemobj_xalloc(pop, &oid, read_arg.unit_size, 0, POBJ_CLASS_ID(0), nullptr, nullptr);
assert(ret == 0);
// release object
pmemobj_free(&oid);
// create custom allocation class
pobj_alloc_class_desc write_arg;
write_arg.unit_size = 1024;
write_arg.units_per_block = 1024;
write_arg.header_type = POBJ_HEADER_COMPACT;
ret = pmemobj_ctl_set(pop, "heap.alloc_class.new.desc", &write_arg);
assert(ret == 0)
string entry_point = "heap.alloc_class." + to_string(write_arg.class_id) + ".desc";
ret = pmemobj_ctl_get(pop, entry_point.c_str(), &read_arg));
assert(ret == 0)
//allocate from custom allocation class
ret = pmemobj_xalloc(pop, &oid, read_arg.unit_size - 16, 0,
POBJ_CLASS_ID(write_arg.class_id), nullptr, nullptr)
assert(ret == 0)
```
Actual result:
- Last assertion was not satisfied, ret = -1
- object allocation failed with ENOMEM
Expected result:
- allocation succeeded
Found in 1.3+b2-276-g335828a | priority | obj ctl allocation from custom class fails with enomem when did after alloc free from class of id steps to reproduce create obj pool pmemoid oid oid null pobj alloc class desc read arg allocate from class of id int ret pmemobj ctl get pop heap alloc class desc read arg assert ret ret pmemobj xalloc pop oid read arg unit size pobj class id nullptr nullptr assert ret release object pmemobj free oid create custom allocation class pobj alloc class desc write arg write arg unit size write arg units per block write arg header type pobj header compact ret pmemobj ctl set pop heap alloc class new desc write arg assert ret string entry point heap alloc class to string write arg class id desc ret pmemobj ctl get pop entry point c str read arg assert ret allocate from custom allocation class ret pmemobj xalloc pop oid read arg unit size pobj class id write arg class id nullptr nullptr assert ret actual result last assertion was not satisfied ret object allocation failed with enomem expected result allocation succeeded found in | 1 |
95,861 | 3,961,338,609 | IssuesEvent | 2016-05-02 12:32:58 | Victoire/victoire | https://api.github.com/repos/Victoire/victoire | opened | [1.6] When I delete a blog article, I'm redirected to a 404 | Priority : Medium Type : Bug | I used a multilingual website, with Quantums (or CustomContent).
I created a blog article in the EN version but deleted it as I wanted to write it under the FR version.
At deletion, i've been redirected to a 404 (with the url www.www.com/en/articles instead of www.www.com/fr/news) and all the EN version of the website was 404. | 1.0 | [1.6] When I delete a blog article, I'm redirected to a 404 - I used a multilingual website, with Quantums (or CustomContent).
I created a blog article in the EN version but deleted it as I wanted to write it under the FR version.
At deletion, i've been redirected to a 404 (with the url www.www.com/en/articles instead of www.www.com/fr/news) and all the EN version of the website was 404. | priority | when i delete a blog article i m redirected to a i used a multilingual website with quantums or customcontent i created a blog article in the en version but deleted it as i wanted to write it under the fr version at deletion i ve been redirected to a with the url instead of and all the en version of the website was | 1 |
760,526 | 26,645,597,869 | IssuesEvent | 2023-01-25 09:44:02 | blockscout/blockscout | https://api.github.com/repos/blockscout/blockscout | closed | Implement fetching internal transactions from tracer: callTracer | priority: medium client: geth indexing | Some forks of Geth remove JS tracer. For instance, avalanchego client did it, and here is the related issue https://github.com/blockscout/blockscout/issues/5729.
In this case, on request with JS tracer, the node responds with:
```
{"jsonrpc":"2.0","id":1,"error":{"code":-32000,"message":"tracer not found"}}
```
At the same time, `callTracer` option is still available:
```
curl -H "Content-type:application/json" --data '{"jsonrpc":"2.0","method":"debug_traceTransaction","params":["0x716d3cba6c5d892d6651cb9a29e90d55b23193b9490d0e35547ed9842922f648", {"tracer": "callTracer"}],"id":1}'
```
I suggest adding the environment variable `INDEXER_INTERNAL_TRANSACTIONS_TRACER_TYPE` with the default value `js`, which means using JS tracer. In the case of a disabled js tracer, the instance maintainer should set `call_tracer` in order to fetch internal transactions using `callTracer` tracer. This should be applied only for `ETHEREUM_JSONRPC_VARIANT=geth`.
<details>
<summary>An example of a response using `callTracer` tracer</summary>
```json
{
"jsonrpc": "2.0",
"id": 1,
"result": {
"type": "CALL",
"from": "0x278bb8e32ce12fd56f44facaaaf28101c6587186",
"to": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"value": "0x0",
"gas": "0x3cb6c4",
"gasUsed": "0x248cb",
"input": "0x1c00ab5b",
"output": "0x",
"calls": [{
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3bb30c",
"gasUsed": "0x5b88",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b6000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000132e746fb71a5f79430000000000000000000000000000000000000000000000000000000000d561900000000000000000000000000000000000000000000000000000000000d36a47",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x3ab21f",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc8f6c6a51050e22fdc1b44d550a52464c6675b12",
"gas": "0x3a95f7",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000006b04e9042e63abf3c4e5d0000000000000000000000000000000000000000000000000000004aa26ecdb70000000000000000000000000000000000000000000000000000000063769de2"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x3a79c5",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3b4e96",
"gasUsed": "0x26ae",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b6000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000000000000000d64f6300000000000000000000000000000000000000000000001325dbd6dc20aac1b70000000000000000000000000000000000000000000000000000000000d3f3e4",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x3a58eb",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc8f6c6a51050e22fdc1b44d550a52464c6675b12",
"gas": "0x3a4e08",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000006b04e9042e63abf3c4e5d0000000000000000000000000000000000000000000000000000004aa26ecdb70000000000000000000000000000000000000000000000000000000063769de2"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x3a432d",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3b1e26",
"gasUsed": "0x5b6a",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000818ec0a7fe18ff94269904fced6ae3dae6d6dc0b000000000000000000000000b44a9b6905af7c801311e8f4e76932ee959c663c000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000296323e4c25b2200000000000000000000000000000000000000000000000000000000000147fc0000000000000000000000000000000000000000000000000000000000d38ff7",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xedf8a513c0e1fc89491e03946f908432d589327b",
"gas": "0x3a1f8d",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000dd5ff4cf680c6b0b90000000000000000000000000000000000000000000000000000004732919c94000000000000000000000000000000000000000000000000000000006376944e"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x6e835ebf3a4c98f62b06f36823f6432a77a4291a",
"gas": "0x3a0365",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000d4c9e03ea249048100000000000000000000000000000000000000000000000000000000069b763860000000000000000000000000000000000000000000000000000000063769c62"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x9117372745df764d2f1feb6372b358de0f856b46",
"gas": "0x39e746",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000000000000a573b7840000000000000000000000000000000000000000000000000000006b0c14e1a50000000000000000000000000000000000000000000000000000000063765c61"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3ab9cc",
"gasUsed": "0x26cc",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000b44a9b6905af7c801311e8f4e76932ee959c663c000000000000000000000000818ec0a7fe18ff94269904fced6ae3dae6d6dc0b000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f80000000000000000000000000000000000000000000000000000000000001492c000000000000000000000000000000000000000000000000002948be8c580bb80000000000000000000000000000000000000000000000000000000000d3cb80",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x9117372745df764d2f1feb6372b358de0f856b46",
"gas": "0x39c661",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000000000000a573b7840000000000000000000000000000000000000000000000000000006b0c14e1a50000000000000000000000000000000000000000000000000000000063765c61"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x6e835ebf3a4c98f62b06f36823f6432a77a4291a",
"gas": "0x39bb73",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000d4c9e03ea249048100000000000000000000000000000000000000000000000000000000069b763860000000000000000000000000000000000000000000000000000000063769c62"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xedf8a513c0e1fc89491e03946f908432d589327b",
"gas": "0x39b098",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000dd5ff4cf680c6b0b90000000000000000000000000000000000000000000000000000004732919c94000000000000000000000000000000000000000000000000000000006376944e"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3a893c",
"gasUsed": "0x49d6",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000efaeee334f0fd1712f9a8cc375f427d9cdd40d73000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000b9ab1b33740de20000000000000000000000000000000000000000000000000000000000d4fac30000000000000000000000000000000000000000000000000000000000d3046d",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc07a896d2e57d60a66f04d4ff893216e6432e18b",
"gas": "0x398d0a",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000446829ce9400000000000000000000000000000000000000000000003ba30cc3ce91e5f4c40000000000000000000000000000000000000000000000000000000063768629"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x669b2884e1e0941f97a44ecc29a8e2a39863a65e",
"gas": "0x3970ec",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000003e393149eadac9656600000000000000000000000000000000000000000000000000000047982230ee00000000000000000000000000000000000000000000000000000000637683d5"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x395e57",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3a362e",
"gasUsed": "0x26cc",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000efaeee334f0fd1712f9a8cc375f427d9cdd40d73000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000000000000000d64f6300000000000000000000000000000000000000000000000000b9b130e49473330000000000000000000000000000000000000000000000000000000000d459ab",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x3944e5",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x669b2884e1e0941f97a44ecc29a8e2a39863a65e",
"gas": "0x393a01",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000003e393149eadac9656600000000000000000000000000000000000000000000000000000047982230ee00000000000000000000000000000000000000000000000000000000637683d5"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc07a896d2e57d60a66f04d4ff893216e6432e18b",
"gas": "0x392f13",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000446829ce9400000000000000000000000000000000000000000000003ba30cc3ce91e5f4c40000000000000000000000000000000000000000000000000000000063768629"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3a059b",
"gasUsed": "0x49f4",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b600000000000000000000000068b2dfc494362aae300f2c401019205d8960226b000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000132e746fb71a5f79430000000000000000000000000000000000000000000000005e42fe1fdc659dc80000000000000000000000000000000000000000000000000000000000d581c7",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x391501",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x75d65d46bb2cf345df7b7d005c0102a589a83034",
"gas": "0x390076",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000764251ac24516a7664d000000000000000000000000000000000000000000017fce4b065c9f5d9ddcde000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x3ceedd917d1f16f0404c03220056e96637b06f07",
"gas": "0x38e44e",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000004d41d5a00ba34c18d00a000000000000000000000000000000000000000000000000000000af8550620a000000000000000000000000000000000000000000000000000000006376a157"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x39b26d",
"gasUsed": "0x26ae",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d00000000000000000000000068b2dfc494362aae300f2c401019205d8960226b000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b6000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000005dbe937c14416d3f000000000000000000000000000000000000000000000012f45f049954c7ddb20000000000000000000000000000000000000000000000000000000000d1d020",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x3ceedd917d1f16f0404c03220056e96637b06f07",
"gas": "0x38c31f",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000004d41d5a00ba34c18d00a000000000000000000000000000000000000000000000000000000af8550620a000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x75d65d46bb2cf345df7b7d005c0102a589a83034",
"gas": "0x38b845",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000764251ac24516a7664d000000000000000000000000000000000000000000017fce4b065c9f5d9ddcde000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x38ad74",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}]
}]
}
}
```
</details>
<details>
<summary>
An example of call with Blockscout's JS tracer
</summary>
```json
{
"jsonrpc": "2.0",
"id": 1,
"result": [{
"type": "call",
"callType": "call",
"from": "0x9654cfbeda666dabc9bfde482cef5dadd9939a9f",
"to": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"input": "0xe275c88700000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000080000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000c000000000000000000000000000000000000000000000000000000000000000076b75616e74726500000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003061723a2f2f59733948532d655f6a6f346252505834462d733259535357386d437547304932476970344772555066413000000000000000000000000000000000",
"output": "0x000000000000000000000000000000000000000000000000000000000000ca69",
"traceAddress": [],
"value": "0x1450f",
"gas": "0x8f8d2",
"gasUsed": "0x8cd18"
}, {
"type": "call",
"callType": "staticcall",
"from": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"to": "0x7a250d5630b4cf539739df2c5dacb4c659f2488d",
"input": "0x1f00ca74000000000000000000000000000000000000000000000000000000000098968000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000002000000000000000000000000b4fbf271143f4fbf7b91a5ded31805e42b2208d6000000000000000000000000d87ba7a50b2e7e660f678a895e4b72e7cb4ccd9c",
"output": "0x",
"traceAddress": [0],
"value": "0x0",
"gas": "0x89048",
"gasUsed": "0x21ae"
}, {
"type": "call",
"callType": "staticcall",
"from": "0x7a250d5630b4cf539739df2c5dacb4c659f2488d",
"to": "0x00b64e468d2c705a0907f58505536a6c8c49ab26",
"input": "0x0902f1ac",
"output": "0x0000000000000000000000000000000000000000000000000009c4cf22ea73b0000000000000000000000000000000000000000000000000049976c7cea2ac080000000000000000000000000000000000000000000000000000000063774678",
"traceAddress": [0, 0],
"value": "0x0",
"gas": "0x85c22",
"gasUsed": "0x9c8"
}, {
"type": "call",
"callType": "call",
"from": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"to": "0x3c8d9f130970358b7e8cbc1dbd0a1eba6ebe368f",
"input": "0x",
"output": "0x",
"traceAddress": [1],
"value": "0x1450f",
"gas": "0x0",
"gasUsed": "0x0"
}, {
"type": "call",
"callType": "call",
"from": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"to": "0xb10ddc662bd561f0b26a8b555e15c71430a74faa",
"input": "0x5141453e0000000000000000000000009654cfbeda666dabc9bfde482cef5dadd9939a9f0000000000000000000000000000000000000000000000000000000000000080000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000c000000000000000000000000000000000000000000000000000000000000000076b75616e74726500000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003061723a2f2f59733948532d655f6a6f346252505834462d733259535357386d437547304932476970344772555066413000000000000000000000000000000000",
"output": "0x000000000000000000000000000000000000000000000000000000000000ca69",
"traceAddress": [2],
"value": "0x0",
"gas": "0x82666",
"gasUsed": "0x80fd7"
}, {
"type": "call",
"callType": "call",
"from": "0xb10ddc662bd561f0b26a8b555e15c71430a74faa",
"to": "0x15987a0417d14cc6f3554166bcb4a590f6891b18",
"input": "0x0c151fec0000000000000000000000009654cfbeda666dabc9bfde482cef5dadd9939a9f00000000000000000000000000000000000000000000000000000000000000a0000000000000000000000000000000000000000000000000000000000000ca69000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000076b75616e74726500000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003061723a2f2f59733948532d655f6a6f346252505834462d733259535357386d437547304932476970344772555066413000000000000000000000000000000000",
"output": "0x000000000000000000000000000000000000000000000000000000000000e338",
"traceAddress": [2, 0],
"value": "0x0",
"gas": "0x5f715",
"gasUsed": "0x5f715"
}, {
"type": "call",
"callType": "staticcall",
"from": "0x15987a0417d14cc6f3554166bcb4a590f6891b18",
"to": "0xb10ddc662bd561f0b26a8b555e15c71430a74faa",
"input": "0x6352211e000000000000000000000000000000000000000000000000000000000000ca69",
"output": "0x0000000000000000000000009654cfbeda666dabc9bfde482cef5dadd9939a9f",
"traceAddress": [2, 0, 0],
"value": "0x0",
"gas": "0x5b61d",
"gasUsed": "0x42b"
}]
}
```
</details>
We should implement a parser that would transform the response from callTracer to js tracer response. Here are the steps of transformation for each internal transaction:
- nestings should be flattened with the indication of nesting in `"traceAddress"` property
- add value "0x0", if it doesn't exist
- "type" should be moved to "callType" and the values should be in lowercase ("STATICCALL" -> "staticcall")
- add "\"type\": \"call\"", if this is a contract call | 1.0 | Implement fetching internal transactions from tracer: callTracer - Some forks of Geth remove JS tracer. For instance, avalanchego client did it, and here is the related issue https://github.com/blockscout/blockscout/issues/5729.
In this case, on request with JS tracer, the node responds with:
```
{"jsonrpc":"2.0","id":1,"error":{"code":-32000,"message":"tracer not found"}}
```
At the same time, `callTracer` option is still available:
```
curl -H "Content-type:application/json" --data '{"jsonrpc":"2.0","method":"debug_traceTransaction","params":["0x716d3cba6c5d892d6651cb9a29e90d55b23193b9490d0e35547ed9842922f648", {"tracer": "callTracer"}],"id":1}'
```
I suggest adding the environment variable `INDEXER_INTERNAL_TRANSACTIONS_TRACER_TYPE` with the default value `js`, which means using JS tracer. In the case of a disabled js tracer, the instance maintainer should set `call_tracer` in order to fetch internal transactions using `callTracer` tracer. This should be applied only for `ETHEREUM_JSONRPC_VARIANT=geth`.
<details>
<summary>An example of a response using `callTracer` tracer</summary>
```json
{
"jsonrpc": "2.0",
"id": 1,
"result": {
"type": "CALL",
"from": "0x278bb8e32ce12fd56f44facaaaf28101c6587186",
"to": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"value": "0x0",
"gas": "0x3cb6c4",
"gasUsed": "0x248cb",
"input": "0x1c00ab5b",
"output": "0x",
"calls": [{
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3bb30c",
"gasUsed": "0x5b88",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b6000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000132e746fb71a5f79430000000000000000000000000000000000000000000000000000000000d561900000000000000000000000000000000000000000000000000000000000d36a47",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x3ab21f",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc8f6c6a51050e22fdc1b44d550a52464c6675b12",
"gas": "0x3a95f7",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000006b04e9042e63abf3c4e5d0000000000000000000000000000000000000000000000000000004aa26ecdb70000000000000000000000000000000000000000000000000000000063769de2"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x3a79c5",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3b4e96",
"gasUsed": "0x26ae",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b6000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000000000000000d64f6300000000000000000000000000000000000000000000001325dbd6dc20aac1b70000000000000000000000000000000000000000000000000000000000d3f3e4",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x3a58eb",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc8f6c6a51050e22fdc1b44d550a52464c6675b12",
"gas": "0x3a4e08",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000006b04e9042e63abf3c4e5d0000000000000000000000000000000000000000000000000000004aa26ecdb70000000000000000000000000000000000000000000000000000000063769de2"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x3a432d",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3b1e26",
"gasUsed": "0x5b6a",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000818ec0a7fe18ff94269904fced6ae3dae6d6dc0b000000000000000000000000b44a9b6905af7c801311e8f4e76932ee959c663c000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000296323e4c25b2200000000000000000000000000000000000000000000000000000000000147fc0000000000000000000000000000000000000000000000000000000000d38ff7",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xedf8a513c0e1fc89491e03946f908432d589327b",
"gas": "0x3a1f8d",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000dd5ff4cf680c6b0b90000000000000000000000000000000000000000000000000000004732919c94000000000000000000000000000000000000000000000000000000006376944e"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x6e835ebf3a4c98f62b06f36823f6432a77a4291a",
"gas": "0x3a0365",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000d4c9e03ea249048100000000000000000000000000000000000000000000000000000000069b763860000000000000000000000000000000000000000000000000000000063769c62"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x9117372745df764d2f1feb6372b358de0f856b46",
"gas": "0x39e746",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000000000000a573b7840000000000000000000000000000000000000000000000000000006b0c14e1a50000000000000000000000000000000000000000000000000000000063765c61"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3ab9cc",
"gasUsed": "0x26cc",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000b44a9b6905af7c801311e8f4e76932ee959c663c000000000000000000000000818ec0a7fe18ff94269904fced6ae3dae6d6dc0b000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f80000000000000000000000000000000000000000000000000000000000001492c000000000000000000000000000000000000000000000000002948be8c580bb80000000000000000000000000000000000000000000000000000000000d3cb80",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x9117372745df764d2f1feb6372b358de0f856b46",
"gas": "0x39c661",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000000000000a573b7840000000000000000000000000000000000000000000000000000006b0c14e1a50000000000000000000000000000000000000000000000000000000063765c61"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x6e835ebf3a4c98f62b06f36823f6432a77a4291a",
"gas": "0x39bb73",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000d4c9e03ea249048100000000000000000000000000000000000000000000000000000000069b763860000000000000000000000000000000000000000000000000000000063769c62"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xedf8a513c0e1fc89491e03946f908432d589327b",
"gas": "0x39b098",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000000dd5ff4cf680c6b0b90000000000000000000000000000000000000000000000000000004732919c94000000000000000000000000000000000000000000000000000000006376944e"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3a893c",
"gasUsed": "0x49d6",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000efaeee334f0fd1712f9a8cc375f427d9cdd40d73000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000b9ab1b33740de20000000000000000000000000000000000000000000000000000000000d4fac30000000000000000000000000000000000000000000000000000000000d3046d",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc07a896d2e57d60a66f04d4ff893216e6432e18b",
"gas": "0x398d0a",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000446829ce9400000000000000000000000000000000000000000000003ba30cc3ce91e5f4c40000000000000000000000000000000000000000000000000000000063768629"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x669b2884e1e0941f97a44ecc29a8e2a39863a65e",
"gas": "0x3970ec",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000003e393149eadac9656600000000000000000000000000000000000000000000000000000047982230ee00000000000000000000000000000000000000000000000000000000637683d5"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x395e57",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3a362e",
"gasUsed": "0x26cc",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000fa9343c3897324496a05fc75abed6bac29f8a40f000000000000000000000000efaeee334f0fd1712f9a8cc375f427d9cdd40d73000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000000000000000d64f6300000000000000000000000000000000000000000000000000b9b130e49473330000000000000000000000000000000000000000000000000000000000d459ab",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x4f4d77f1f8b12b183c78c53d6dd9beac805e1b09",
"gas": "0x3944e5",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000b4748b9eb5000000000000000000000000000000000000000000000000000000b5956ae58d000000000000000000000000000000000000000000000000000000006376a049"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x669b2884e1e0941f97a44ecc29a8e2a39863a65e",
"gas": "0x393a01",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x00000000000000000000000000000000000000000000003e393149eadac9656600000000000000000000000000000000000000000000000000000047982230ee00000000000000000000000000000000000000000000000000000000637683d5"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xc07a896d2e57d60a66f04d4ff893216e6432e18b",
"gas": "0x392f13",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000000000000446829ce9400000000000000000000000000000000000000000000003ba30cc3ce91e5f4c40000000000000000000000000000000000000000000000000000000063768629"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x3a059b",
"gasUsed": "0x49f4",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b600000000000000000000000068b2dfc494362aae300f2c401019205d8960226b000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000132e746fb71a5f79430000000000000000000000000000000000000000000000005e42fe1fdc659dc80000000000000000000000000000000000000000000000000000000000d581c7",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x391501",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x75d65d46bb2cf345df7b7d005c0102a589a83034",
"gas": "0x390076",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000764251ac24516a7664d000000000000000000000000000000000000000000017fce4b065c9f5d9ddcde000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x3ceedd917d1f16f0404c03220056e96637b06f07",
"gas": "0x38e44e",
"gasUsed": "0x9c8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000004d41d5a00ba34c18d00a000000000000000000000000000000000000000000000000000000af8550620a000000000000000000000000000000000000000000000000000000006376a157"
}]
}, {
"type": "STATICCALL",
"from": "0x27fb33be927a17a783739e08a5752dab743c7f74",
"to": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"gas": "0x39b26d",
"gasUsed": "0x26ae",
"input": "0xd06ca61f0000000000000000000000000000000000000000000000000000000000d59f8000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000004000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d00000000000000000000000068b2dfc494362aae300f2c401019205d8960226b000000000000000000000000b58a9d5920af6ac1a9522b0b10f55df16686d1b6000000000000000000000000e3f5a90f9cb311505cd691a46596599aa1a0ad7d",
"output": "0x000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000040000000000000000000000000000000000000000000000000000000000d59f800000000000000000000000000000000000000000000000005dbe937c14416d3f000000000000000000000000000000000000000000000012f45f049954c7ddb20000000000000000000000000000000000000000000000000000000000d1d020",
"calls": [{
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x3ceedd917d1f16f0404c03220056e96637b06f07",
"gas": "0x38c31f",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000004d41d5a00ba34c18d00a000000000000000000000000000000000000000000000000000000af8550620a000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0x75d65d46bb2cf345df7b7d005c0102a589a83034",
"gas": "0x38b845",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000000764251ac24516a7664d000000000000000000000000000000000000000000017fce4b065c9f5d9ddcde000000000000000000000000000000000000000000000000000000006376a157"
}, {
"type": "STATICCALL",
"from": "0xa4196322aa900acc92cd5cd978ab47e77efa07eb",
"to": "0xf2fb82d81240e641a49076fe6879c128a485bb3f",
"gas": "0x38ad74",
"gasUsed": "0x1f8",
"input": "0x0902f1ac",
"output": "0x000000000000000000000000000000000000000000069575f2b4da04eb13c1a00000000000000000000000000000000000000000000000000000004919c030ab000000000000000000000000000000000000000000000000000000006376a157"
}]
}]
}
}
```
</details>
<details>
<summary>
An example of call with Blockscout's JS tracer
</summary>
```json
{
"jsonrpc": "2.0",
"id": 1,
"result": [{
"type": "call",
"callType": "call",
"from": "0x9654cfbeda666dabc9bfde482cef5dadd9939a9f",
"to": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"input": "0xe275c88700000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000080000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000c000000000000000000000000000000000000000000000000000000000000000076b75616e74726500000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003061723a2f2f59733948532d655f6a6f346252505834462d733259535357386d437547304932476970344772555066413000000000000000000000000000000000",
"output": "0x000000000000000000000000000000000000000000000000000000000000ca69",
"traceAddress": [],
"value": "0x1450f",
"gas": "0x8f8d2",
"gasUsed": "0x8cd18"
}, {
"type": "call",
"callType": "staticcall",
"from": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"to": "0x7a250d5630b4cf539739df2c5dacb4c659f2488d",
"input": "0x1f00ca74000000000000000000000000000000000000000000000000000000000098968000000000000000000000000000000000000000000000000000000000000000400000000000000000000000000000000000000000000000000000000000000002000000000000000000000000b4fbf271143f4fbf7b91a5ded31805e42b2208d6000000000000000000000000d87ba7a50b2e7e660f678a895e4b72e7cb4ccd9c",
"output": "0x",
"traceAddress": [0],
"value": "0x0",
"gas": "0x89048",
"gasUsed": "0x21ae"
}, {
"type": "call",
"callType": "staticcall",
"from": "0x7a250d5630b4cf539739df2c5dacb4c659f2488d",
"to": "0x00b64e468d2c705a0907f58505536a6c8c49ab26",
"input": "0x0902f1ac",
"output": "0x0000000000000000000000000000000000000000000000000009c4cf22ea73b0000000000000000000000000000000000000000000000000049976c7cea2ac080000000000000000000000000000000000000000000000000000000063774678",
"traceAddress": [0, 0],
"value": "0x0",
"gas": "0x85c22",
"gasUsed": "0x9c8"
}, {
"type": "call",
"callType": "call",
"from": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"to": "0x3c8d9f130970358b7e8cbc1dbd0a1eba6ebe368f",
"input": "0x",
"output": "0x",
"traceAddress": [1],
"value": "0x1450f",
"gas": "0x0",
"gasUsed": "0x0"
}, {
"type": "call",
"callType": "call",
"from": "0x4454d3892124ad4d859770660495461d1c5a37f3",
"to": "0xb10ddc662bd561f0b26a8b555e15c71430a74faa",
"input": "0x5141453e0000000000000000000000009654cfbeda666dabc9bfde482cef5dadd9939a9f0000000000000000000000000000000000000000000000000000000000000080000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000c000000000000000000000000000000000000000000000000000000000000000076b75616e74726500000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003061723a2f2f59733948532d655f6a6f346252505834462d733259535357386d437547304932476970344772555066413000000000000000000000000000000000",
"output": "0x000000000000000000000000000000000000000000000000000000000000ca69",
"traceAddress": [2],
"value": "0x0",
"gas": "0x82666",
"gasUsed": "0x80fd7"
}, {
"type": "call",
"callType": "call",
"from": "0xb10ddc662bd561f0b26a8b555e15c71430a74faa",
"to": "0x15987a0417d14cc6f3554166bcb4a590f6891b18",
"input": "0x0c151fec0000000000000000000000009654cfbeda666dabc9bfde482cef5dadd9939a9f00000000000000000000000000000000000000000000000000000000000000a0000000000000000000000000000000000000000000000000000000000000ca69000000000000000000000000000000000000000000000000000000000000000100000000000000000000000000000000000000000000000000000000000000e000000000000000000000000000000000000000000000000000000000000000076b75616e74726500000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000003061723a2f2f59733948532d655f6a6f346252505834462d733259535357386d437547304932476970344772555066413000000000000000000000000000000000",
"output": "0x000000000000000000000000000000000000000000000000000000000000e338",
"traceAddress": [2, 0],
"value": "0x0",
"gas": "0x5f715",
"gasUsed": "0x5f715"
}, {
"type": "call",
"callType": "staticcall",
"from": "0x15987a0417d14cc6f3554166bcb4a590f6891b18",
"to": "0xb10ddc662bd561f0b26a8b555e15c71430a74faa",
"input": "0x6352211e000000000000000000000000000000000000000000000000000000000000ca69",
"output": "0x0000000000000000000000009654cfbeda666dabc9bfde482cef5dadd9939a9f",
"traceAddress": [2, 0, 0],
"value": "0x0",
"gas": "0x5b61d",
"gasUsed": "0x42b"
}]
}
```
</details>
We should implement a parser that would transform the response from callTracer to js tracer response. Here are the steps of transformation for each internal transaction:
- nestings should be flattened with the indication of nesting in `"traceAddress"` property
- add value "0x0", if it doesn't exist
- "type" should be moved to "callType" and the values should be in lowercase ("STATICCALL" -> "staticcall")
- add "\"type\": \"call\"", if this is a contract call | priority | implement fetching internal transactions from tracer calltracer some forks of geth remove js tracer for instance avalanchego client did it and here is the related issue in this case on request with js tracer the node responds with jsonrpc id error code message tracer not found at the same time calltracer option is still available curl h content type application json data jsonrpc method debug tracetransaction params id i suggest adding the environment variable indexer internal transactions tracer type with the default value js which means using js tracer in the case of a disabled js tracer the instance maintainer should set call tracer in order to fetch internal transactions using calltracer tracer this should be applied only for ethereum jsonrpc variant geth an example of a response using calltracer tracer json jsonrpc id result type call from to value gas gasused input output calls type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output calls type staticcall from to gas gasused input output type staticcall from to gas gasused input output type staticcall from to gas gasused input output an example of call with blockscout s js tracer json jsonrpc id result type call calltype call from to input output traceaddress value gas gasused type call calltype staticcall from to input output traceaddress value gas gasused type call calltype staticcall from to input output traceaddress value gas gasused type call calltype call from to input output traceaddress value gas gasused type call calltype call from to input output traceaddress value gas gasused type call calltype call from to input output traceaddress value gas gasused type call calltype staticcall from to input output traceaddress value gas gasused we should implement a parser that would transform the response from calltracer to js tracer response here are the steps of transformation for each internal transaction nestings should be flattened with the indication of nesting in traceaddress property add value if it doesn t exist type should be moved to calltype and the values should be in lowercase staticcall staticcall add type call if this is a contract call | 1 |
762,275 | 26,713,319,844 | IssuesEvent | 2023-01-28 06:29:38 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Fetch changelogs from github.com via GraphQL | type:feature priority-3-medium status:in-progress core:changelogs | ### What would you like Renovate to be able to do?
I would like for changelog data from github.com to be fetched via GraphQL so they can be cached, thus preventing frequent rate limiting of my github.com API token.
### If you have any ideas on how this should be implemented, please tell us here.
`lib/workers/repository/update/pr/changelog/github/index.ts` is currently using the REST API to fetch tags and releases. Caching for those requests was removed in v32.194.1 via #17685. A number of changes have been implemented to Renovate since where many Github API requests now use GraphQL plus caching, such as the `github-tabs` and `github-releases` datasources via the newer `queryTags()` and `queryReleases()` methods.
I propose we update `lib/workers/repository/update/pr/changelog/github/index.ts` to use those same methods so we can benefit from that same caching.
### Is this a feature you are interested in implementing yourself?
Maybe | 1.0 | Fetch changelogs from github.com via GraphQL - ### What would you like Renovate to be able to do?
I would like for changelog data from github.com to be fetched via GraphQL so they can be cached, thus preventing frequent rate limiting of my github.com API token.
### If you have any ideas on how this should be implemented, please tell us here.
`lib/workers/repository/update/pr/changelog/github/index.ts` is currently using the REST API to fetch tags and releases. Caching for those requests was removed in v32.194.1 via #17685. A number of changes have been implemented to Renovate since where many Github API requests now use GraphQL plus caching, such as the `github-tabs` and `github-releases` datasources via the newer `queryTags()` and `queryReleases()` methods.
I propose we update `lib/workers/repository/update/pr/changelog/github/index.ts` to use those same methods so we can benefit from that same caching.
### Is this a feature you are interested in implementing yourself?
Maybe | priority | fetch changelogs from github com via graphql what would you like renovate to be able to do i would like for changelog data from github com to be fetched via graphql so they can be cached thus preventing frequent rate limiting of my github com api token if you have any ideas on how this should be implemented please tell us here lib workers repository update pr changelog github index ts is currently using the rest api to fetch tags and releases caching for those requests was removed in via a number of changes have been implemented to renovate since where many github api requests now use graphql plus caching such as the github tabs and github releases datasources via the newer querytags and queryreleases methods i propose we update lib workers repository update pr changelog github index ts to use those same methods so we can benefit from that same caching is this a feature you are interested in implementing yourself maybe | 1 |
475,044 | 13,686,260,960 | IssuesEvent | 2020-09-30 08:26:45 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] Disallow names with hyphen for content type fields | enhancement priority: medium | GraphQL disallows names with hyphens per spec. When a content has a field with hyphens, hyphen would get transformed to double underscore for the GraphQL queries.
We should also disallow hyphens to maintain consistency and avoid discordance/confusion in content retrieval and field usage.
| 1.0 | [studio-ui] Disallow names with hyphen for content type fields - GraphQL disallows names with hyphens per spec. When a content has a field with hyphens, hyphen would get transformed to double underscore for the GraphQL queries.
We should also disallow hyphens to maintain consistency and avoid discordance/confusion in content retrieval and field usage.
| priority | disallow names with hyphen for content type fields graphql disallows names with hyphens per spec when a content has a field with hyphens hyphen would get transformed to double underscore for the graphql queries we should also disallow hyphens to maintain consistency and avoid discordance confusion in content retrieval and field usage | 1 |
437,028 | 12,558,830,510 | IssuesEvent | 2020-06-07 17:05:02 | dompdf/dompdf | https://api.github.com/repos/dompdf/dompdf | closed | border-collapse: collapse on a non-flowed table causes its borders to render independently of the content | Priority-Medium bug css imported tables | _Original author: eclecticgeek (May 14, 2012 02:49:02)_
Layout: A table containing a row that has a small table in one cell and a large amount of content in another cell. The inner table is styled with `border-collapse: collapse;`.
Result: The borders of the inner table will always be positioned at the top of the cell. This occurs even if the inner table's height is less than the height of the containing cell and the vertical-align property is set to something other than top. The content of the inner table, however, is positioned correctly.
This bug was first reported in v0.6.0 beta 3. It is not present in v0.6.0 beta 2. I have not identified the specific revision in which it was introduced.
This report is based on the discussion here:
https://groups.google.com/d/topic/dompdf/2fQCjFXY6JM/discussion
_Original issue: http://code.google.com/p/dompdf/issues/detail?id=483_
| 1.0 | border-collapse: collapse on a non-flowed table causes its borders to render independently of the content - _Original author: eclecticgeek (May 14, 2012 02:49:02)_
Layout: A table containing a row that has a small table in one cell and a large amount of content in another cell. The inner table is styled with `border-collapse: collapse;`.
Result: The borders of the inner table will always be positioned at the top of the cell. This occurs even if the inner table's height is less than the height of the containing cell and the vertical-align property is set to something other than top. The content of the inner table, however, is positioned correctly.
This bug was first reported in v0.6.0 beta 3. It is not present in v0.6.0 beta 2. I have not identified the specific revision in which it was introduced.
This report is based on the discussion here:
https://groups.google.com/d/topic/dompdf/2fQCjFXY6JM/discussion
_Original issue: http://code.google.com/p/dompdf/issues/detail?id=483_
| priority | border collapse collapse on a non flowed table causes its borders to render independently of the content original author eclecticgeek may layout a table containing a row that has a small table in one cell and a large amount of content in another cell the inner table is styled with border collapse collapse result the borders of the inner table will always be positioned at the top of the cell this occurs even if the inner table s height is less than the height of the containing cell and the vertical align property is set to something other than top the content of the inner table however is positioned correctly this bug was first reported in beta it is not present in beta i have not identified the specific revision in which it was introduced this report is based on the discussion here original issue | 1 |
631,464 | 20,152,210,540 | IssuesEvent | 2022-02-09 13:29:47 | ita-social-projects/horondi_client_fe | https://api.github.com/repos/ita-social-projects/horondi_client_fe | closed | [Your Checkout Delivery page] delivery amount is not shown when user chose delivery by Nova poshta courier | bug UI priority: medium severity: minor | **Environment:**
Windows 10 Home edition, 20H2 browser Google Chrome Version 91.0.4472.77 (Official Build) (64-bit).
Reproducible: always.
Build found: last commit 5e32bd7"
**Preconditions:**
1. Go to https://horondi-front-staging.azurewebsites.net/
2. Register as user
3. Choose a product from main menu catalog (e.g.- Bag shopper)
4. Add the chosen product to the shopping cart
5. Click the button “Buy now”
**Steps to reproduce:**
1. Click on the cart label in the right top corner of the page
2. Choose delivery type Nova Poshta courier
3. Click ‘Go to checkout’ button on the Your Cart-Checkout page
4. Pay attention to the information about total price, delivery amount, total amount
**Actual result:**

**Expected result:**
1. When the user had chosen Nova post courier delivery the user sees the delivery amount, the amount of goods, and the total amount.
2.
User story and test case links E.g.: "User story #148
Labels to be added "Bug", Priority ("Medium"), Severity ("Major"), Type ("UI,).
| 1.0 | [Your Checkout Delivery page] delivery amount is not shown when user chose delivery by Nova poshta courier - **Environment:**
Windows 10 Home edition, 20H2 browser Google Chrome Version 91.0.4472.77 (Official Build) (64-bit).
Reproducible: always.
Build found: last commit 5e32bd7"
**Preconditions:**
1. Go to https://horondi-front-staging.azurewebsites.net/
2. Register as user
3. Choose a product from main menu catalog (e.g.- Bag shopper)
4. Add the chosen product to the shopping cart
5. Click the button “Buy now”
**Steps to reproduce:**
1. Click on the cart label in the right top corner of the page
2. Choose delivery type Nova Poshta courier
3. Click ‘Go to checkout’ button on the Your Cart-Checkout page
4. Pay attention to the information about total price, delivery amount, total amount
**Actual result:**

**Expected result:**
1. When the user had chosen Nova post courier delivery the user sees the delivery amount, the amount of goods, and the total amount.
2.
User story and test case links E.g.: "User story #148
Labels to be added "Bug", Priority ("Medium"), Severity ("Major"), Type ("UI,).
| priority | delivery amount is not shown when user chose delivery by nova poshta courier environment windows home edition browser google chrome version official build bit reproducible always build found last commit preconditions go to register as user choose a product from main menu catalog e g bag shopper add the chosen product to the shopping cart click the button “buy now” steps to reproduce click on the cart label in the right top corner of the page choose delivery type nova poshta courier click ‘go to checkout’ button on the your cart checkout page pay attention to the information about total price delivery amount total amount actual result expected result when the user had chosen nova post courier delivery the user sees the delivery amount the amount of goods and the total amount user story and test case links e g user story labels to be added bug priority medium severity major type ui | 1 |
635,911 | 20,513,633,274 | IssuesEvent | 2022-03-01 09:30:38 | zangGallery/frontend | https://api.github.com/repos/zangGallery/frontend | opened | Check if there are already pending transactions | Priority: Medium Type: New Feature | If a user executes a transaction and reloads the page, it stops showing info regarding the pending transactions. | 1.0 | Check if there are already pending transactions - If a user executes a transaction and reloads the page, it stops showing info regarding the pending transactions. | priority | check if there are already pending transactions if a user executes a transaction and reloads the page it stops showing info regarding the pending transactions | 1 |
401,481 | 11,790,760,143 | IssuesEvent | 2020-03-17 19:36:53 | opentargets/platform | https://api.github.com/repos/opentargets/platform | closed | Investigate why ACE2 has drug profile page but no evidence strings | Kind: Data Priority: Medium Topic: ChEMBL | For [ACE2 ](https://www.targetvalidation.org/target/ENSG00000130234), the tractability assessment includes data that there is a Phase IV compound that targets the gene.
Upon review of the new tractability file, the Phase IV assessment comes from [Cilazapril - CHEMBL515606](https://www.targetvalidation.org/summary?drug=CHEMBL515606). According to ChEMBL, the drug has been indicated for cardiovascular disease based on an ATC record.
@AsierGonzalez, can you please investigate this?
For reference, below you can find screenshots from the #dev Slack channel:
<img width="1393" alt="Screenshot 2020-03-17 at 15 53 10" src="https://user-images.githubusercontent.com/7490258/76874698-77db4380-6867-11ea-84ac-c451ad131bd4.png">
<img width="1394" alt="Screenshot 2020-03-17 at 15 53 29" src="https://user-images.githubusercontent.com/7490258/76874703-7ad63400-6867-11ea-9245-6e1f88f87a50.png"> | 1.0 | Investigate why ACE2 has drug profile page but no evidence strings - For [ACE2 ](https://www.targetvalidation.org/target/ENSG00000130234), the tractability assessment includes data that there is a Phase IV compound that targets the gene.
Upon review of the new tractability file, the Phase IV assessment comes from [Cilazapril - CHEMBL515606](https://www.targetvalidation.org/summary?drug=CHEMBL515606). According to ChEMBL, the drug has been indicated for cardiovascular disease based on an ATC record.
@AsierGonzalez, can you please investigate this?
For reference, below you can find screenshots from the #dev Slack channel:
<img width="1393" alt="Screenshot 2020-03-17 at 15 53 10" src="https://user-images.githubusercontent.com/7490258/76874698-77db4380-6867-11ea-84ac-c451ad131bd4.png">
<img width="1394" alt="Screenshot 2020-03-17 at 15 53 29" src="https://user-images.githubusercontent.com/7490258/76874703-7ad63400-6867-11ea-9245-6e1f88f87a50.png"> | priority | investigate why has drug profile page but no evidence strings for the tractability assessment includes data that there is a phase iv compound that targets the gene upon review of the new tractability file the phase iv assessment comes from according to chembl the drug has been indicated for cardiovascular disease based on an atc record asiergonzalez can you please investigate this for reference below you can find screenshots from the dev slack channel img width alt screenshot at src img width alt screenshot at src | 1 |
442,877 | 12,752,208,214 | IssuesEvent | 2020-06-27 15:15:53 | code4romania/mon-vot-android-kotlin | https://api.github.com/repos/code4romania/mon-vot-android-kotlin | closed | Sort forms list by the new order field | enhancement february-2020 good first issue help wanted medium-priority | forms should be sorted by order.
Backend implemented: https://github.com/code4romania/monitorizare-vot/issues/248 | 1.0 | Sort forms list by the new order field - forms should be sorted by order.
Backend implemented: https://github.com/code4romania/monitorizare-vot/issues/248 | priority | sort forms list by the new order field forms should be sorted by order backend implemented | 1 |
245,509 | 7,887,155,865 | IssuesEvent | 2018-06-27 17:31:36 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | 0.7.6.0 - no way to create divided land claim | Medium Priority | **Version:** 0.7.6.0 beta staging-4767aa88
So... Before - you need DEED in inventory, and you can claim multiple plot on far away from other to one deed.
Now, you cant.
(Yes, you can POSTmanage by estate desk, but this is not easy to know, and is not good to use on the fly) | 1.0 | 0.7.6.0 - no way to create divided land claim - **Version:** 0.7.6.0 beta staging-4767aa88
So... Before - you need DEED in inventory, and you can claim multiple plot on far away from other to one deed.
Now, you cant.
(Yes, you can POSTmanage by estate desk, but this is not easy to know, and is not good to use on the fly) | priority | no way to create divided land claim version beta staging so before you need deed in inventory and you can claim multiple plot on far away from other to one deed now you cant yes you can postmanage by estate desk but this is not easy to know and is not good to use on the fly | 1 |
465,209 | 13,358,754,556 | IssuesEvent | 2020-08-31 12:16:25 | ansible-collections/community.kubernetes | https://api.github.com/repos/ansible-collections/community.kubernetes | closed | k8s module src directive does not reference files directory when used in a role | needs_triage priority/medium type/bug | <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
The `k8s` module has a `src` directive for specifying the location of files to provide a path to a file containing a valid YAML definition of an object or objects to be created or updated. It appears that
this requires a full path name. When used in a role, this directive doesn't appear to use the roles/x/{files,templates,tasks}/ by default. It will use it if `"{{ role_path }}/files/my_file.yml` is used however.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```
ansible 2.7.8
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.5 (default, Jun 11 2019, 14:33:56) [GCC 4.8.5 20150623 (Red Hat 4.8.5-39)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```[blank]
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
RHEL 7.7
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Create secret for htpasswd
k8s:
kubeconfig: clusters/{{ cluster }}/ocp4/auth/kubeconfig
state: present
src: files/secret.yml
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Secret created based on the contents of `files/secret.yml`
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
Playbook fails with "Error accessing files/secret.yml. Does the file exist?"
<!--- Paste verbatim command output between quotes -->
```
ansible-playbook 2.7.8
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible-playbook
python version = 2.7.5 (default, Jun 11 2019, 14:33:56) [GCC 4.8.5 20150623 (Red Hat 4.8.5-39)]
Using /etc/ansible/ansible.cfg as config file
/root/ocp4/inventories/hosts.bal1 did not meet host_list requirements, check plugin documentation if this is unexpected
/root/ocp4/inventories/hosts.bal1 did not meet script requirements, check plugin documentation if this is unexpected
/root/ocp4/inventories/hosts.bal1 did not meet yaml requirements, check plugin documentation if this is unexpected
Parsed /root/ocp4/inventories/hosts.bal1 inventory source with ini plugin
PLAYBOOK: postconfig.yml ********************************************************************************************
1 plays in postconfig.yml
PLAY [localhost] ****************************************************************************************************
META: ran handlers
TASK [user-registry-backdoor : Create secret for htpasswd] ******************************************************
task path: /root/ocp4/roles/user-registry-backdoor/tasks/main.yml:1
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root
<127.0.0.1> EXEC /bin/sh -c 'echo ~root && sleep 0'
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932 `" && echo ansible-tmp-1585752693.84-12716652062932="` echo /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932 `" ) && sleep 0'
Using module file /usr/lib/python2.7/site-packages/ansible/modules/clustering/k8s/k8s.py
<127.0.0.1> PUT /root/.ansible/tmp/ansible-local-48900CduyEt/tmpc6jyoA TO /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/AnsiballZ_k8s.py
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/ /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/AnsiballZ_k8s.py && sleep 0'
<127.0.0.1> EXEC /bin/sh -c '/usr/bin/python2 /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/AnsiballZ_k8s.py && sleep 0'
<127.0.0.1> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/ > /dev/null 2>&1 && sleep 0'
fatal: [localhost]: FAILED! => {
"changed": false,
"invocation": {
"module_args": {
"api_key": null,
"cert_file": null,
"context": null,
"force": false,
"host": null,
"key_file": null,
"kubeconfig": "clusters/bal1/ocp4/auth/kubeconfig",
"merge_type": null,
"password": null,
"ssl_ca_cert": null,
"state": "present",
"username": null,
"verify_ssl": null
}
},
"msg": "Error accessing files/secret.yml. Does the file exist?"
}
to retry, use: --limit @/root/ocp4/postconfig.retry
PLAY RECAP **********************************************************************************************************
localhost : ok=0 changed=0 unreachable=0 failed=1
```
| 1.0 | k8s module src directive does not reference files directory when used in a role - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
The `k8s` module has a `src` directive for specifying the location of files to provide a path to a file containing a valid YAML definition of an object or objects to be created or updated. It appears that
this requires a full path name. When used in a role, this directive doesn't appear to use the roles/x/{files,templates,tasks}/ by default. It will use it if `"{{ role_path }}/files/my_file.yml` is used however.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```
ansible 2.7.8
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.5 (default, Jun 11 2019, 14:33:56) [GCC 4.8.5 20150623 (Red Hat 4.8.5-39)]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```[blank]
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
RHEL 7.7
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Create secret for htpasswd
k8s:
kubeconfig: clusters/{{ cluster }}/ocp4/auth/kubeconfig
state: present
src: files/secret.yml
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
Secret created based on the contents of `files/secret.yml`
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
Playbook fails with "Error accessing files/secret.yml. Does the file exist?"
<!--- Paste verbatim command output between quotes -->
```
ansible-playbook 2.7.8
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible-playbook
python version = 2.7.5 (default, Jun 11 2019, 14:33:56) [GCC 4.8.5 20150623 (Red Hat 4.8.5-39)]
Using /etc/ansible/ansible.cfg as config file
/root/ocp4/inventories/hosts.bal1 did not meet host_list requirements, check plugin documentation if this is unexpected
/root/ocp4/inventories/hosts.bal1 did not meet script requirements, check plugin documentation if this is unexpected
/root/ocp4/inventories/hosts.bal1 did not meet yaml requirements, check plugin documentation if this is unexpected
Parsed /root/ocp4/inventories/hosts.bal1 inventory source with ini plugin
PLAYBOOK: postconfig.yml ********************************************************************************************
1 plays in postconfig.yml
PLAY [localhost] ****************************************************************************************************
META: ran handlers
TASK [user-registry-backdoor : Create secret for htpasswd] ******************************************************
task path: /root/ocp4/roles/user-registry-backdoor/tasks/main.yml:1
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: root
<127.0.0.1> EXEC /bin/sh -c 'echo ~root && sleep 0'
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932 `" && echo ansible-tmp-1585752693.84-12716652062932="` echo /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932 `" ) && sleep 0'
Using module file /usr/lib/python2.7/site-packages/ansible/modules/clustering/k8s/k8s.py
<127.0.0.1> PUT /root/.ansible/tmp/ansible-local-48900CduyEt/tmpc6jyoA TO /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/AnsiballZ_k8s.py
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/ /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/AnsiballZ_k8s.py && sleep 0'
<127.0.0.1> EXEC /bin/sh -c '/usr/bin/python2 /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/AnsiballZ_k8s.py && sleep 0'
<127.0.0.1> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1585752693.84-12716652062932/ > /dev/null 2>&1 && sleep 0'
fatal: [localhost]: FAILED! => {
"changed": false,
"invocation": {
"module_args": {
"api_key": null,
"cert_file": null,
"context": null,
"force": false,
"host": null,
"key_file": null,
"kubeconfig": "clusters/bal1/ocp4/auth/kubeconfig",
"merge_type": null,
"password": null,
"ssl_ca_cert": null,
"state": "present",
"username": null,
"verify_ssl": null
}
},
"msg": "Error accessing files/secret.yml. Does the file exist?"
}
to retry, use: --limit @/root/ocp4/postconfig.retry
PLAY RECAP **********************************************************************************************************
localhost : ok=0 changed=0 unreachable=0 failed=1
```
| priority | module src directive does not reference files directory when used in a role summary the module has a src directive for specifying the location of files to provide a path to a file containing a valid yaml definition of an object or objects to be created or updated it appears that this requires a full path name when used in a role this directive doesn t appear to use the roles x files templates tasks by default it will use it if role path files my file yml is used however issue type bug report component name ansible version ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default jun configuration os environment rhel steps to reproduce name create secret for htpasswd kubeconfig clusters cluster auth kubeconfig state present src files secret yml expected results secret created based on the contents of files secret yml actual results playbook fails with error accessing files secret yml does the file exist ansible playbook config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible playbook python version default jun using etc ansible ansible cfg as config file root inventories hosts did not meet host list requirements check plugin documentation if this is unexpected root inventories hosts did not meet script requirements check plugin documentation if this is unexpected root inventories hosts did not meet yaml requirements check plugin documentation if this is unexpected parsed root inventories hosts inventory source with ini plugin playbook postconfig yml plays in postconfig yml play meta ran handlers task task path root roles user registry backdoor tasks main yml establish local connection for user root exec bin sh c echo root sleep exec bin sh c umask mkdir p echo root ansible tmp ansible tmp echo ansible tmp echo root ansible tmp ansible tmp sleep using module file usr lib site packages ansible modules clustering py put root ansible tmp ansible local to root ansible tmp ansible tmp ansiballz py exec bin sh c chmod u x root ansible tmp ansible tmp root ansible tmp ansible tmp ansiballz py sleep exec bin sh c usr bin root ansible tmp ansible tmp ansiballz py sleep exec bin sh c rm f r root ansible tmp ansible tmp dev null sleep fatal failed changed false invocation module args api key null cert file null context null force false host null key file null kubeconfig clusters auth kubeconfig merge type null password null ssl ca cert null state present username null verify ssl null msg error accessing files secret yml does the file exist to retry use limit root postconfig retry play recap localhost ok changed unreachable failed | 1 |
365,298 | 10,780,519,555 | IssuesEvent | 2019-11-04 13:10:18 | CESARBR/knot-setup-android | https://api.github.com/repos/CESARBR/knot-setup-android | opened | Develop connectivity with KNoT Cloud | priority: medium | The KNoT Cloud uses WebSockets as its means of communication. In order to GET and POST data from the Cloud, one should have a networking infrastructure to communicate with the server that takes cares of connection and error handling. | 1.0 | Develop connectivity with KNoT Cloud - The KNoT Cloud uses WebSockets as its means of communication. In order to GET and POST data from the Cloud, one should have a networking infrastructure to communicate with the server that takes cares of connection and error handling. | priority | develop connectivity with knot cloud the knot cloud uses websockets as its means of communication in order to get and post data from the cloud one should have a networking infrastructure to communicate with the server that takes cares of connection and error handling | 1 |
325,958 | 9,940,841,621 | IssuesEvent | 2019-07-03 10:04:05 | garden-io/garden | https://api.github.com/repos/garden-io/garden | closed | Issue with Rails example project | bug priority:medium | ## Bug
The Rails example project seems to be having problems post-v0.10.
See: https://medium.com/@jiricech94/love-this-c6a4adb282a4
### Current Behavior
```
Command failed with code 127: /Users/user/.garden/tools/kubectl/52606cbad4c7babc/kubectl — conte
xt=docker-for-desktop — namespace=vote-rails run run-frontend-1562062353574 — image=frontend:v-08424
01b63 — restart=Never — quiet — rm -i — overrides {“metadata”:{“annotations”:{“sidecar.istio.io/inje
ct”:”false”}}} — env=GARDEN_VERSION=v-0842401b63 — env=GARDEN_MODULE_POSTGRES__LOCAL-IMAGE-NAME=post
gres — env=GARDEN_MODULE_POSTGRES__DEPLOYMENT-IMAGE-NAME=postgres — rake db:migrate
pod vote-rails/run-frontend-1562062353574 terminated (ContainerCannotRun)
OCI runtime create failed: container_linux.go:344: starting container process caused “exec: \”rake d
b:migrate\”: executable file not found in $PATH”: unknown
```
### Expected behavior
The `db-migrate` task should run without problems.
### Reproducible example
See above.
### Your environment
User's environment was Docker Desktop (2.0.0.3) for OS X, presumably running Garden v0.10. | 1.0 | Issue with Rails example project - ## Bug
The Rails example project seems to be having problems post-v0.10.
See: https://medium.com/@jiricech94/love-this-c6a4adb282a4
### Current Behavior
```
Command failed with code 127: /Users/user/.garden/tools/kubectl/52606cbad4c7babc/kubectl — conte
xt=docker-for-desktop — namespace=vote-rails run run-frontend-1562062353574 — image=frontend:v-08424
01b63 — restart=Never — quiet — rm -i — overrides {“metadata”:{“annotations”:{“sidecar.istio.io/inje
ct”:”false”}}} — env=GARDEN_VERSION=v-0842401b63 — env=GARDEN_MODULE_POSTGRES__LOCAL-IMAGE-NAME=post
gres — env=GARDEN_MODULE_POSTGRES__DEPLOYMENT-IMAGE-NAME=postgres — rake db:migrate
pod vote-rails/run-frontend-1562062353574 terminated (ContainerCannotRun)
OCI runtime create failed: container_linux.go:344: starting container process caused “exec: \”rake d
b:migrate\”: executable file not found in $PATH”: unknown
```
### Expected behavior
The `db-migrate` task should run without problems.
### Reproducible example
See above.
### Your environment
User's environment was Docker Desktop (2.0.0.3) for OS X, presumably running Garden v0.10. | priority | issue with rails example project bug the rails example project seems to be having problems post see current behavior command failed with code users user garden tools kubectl kubectl — conte xt docker for desktop — namespace vote rails run run frontend — image frontend v — restart never — quiet — rm i — overrides “metadata” “annotations” “sidecar istio io inje ct” ”false” — env garden version v — env garden module postgres local image name post gres — env garden module postgres deployment image name postgres — rake db migrate pod vote rails run frontend terminated containercannotrun oci runtime create failed container linux go starting container process caused “exec ”rake d b migrate ” executable file not found in path” unknown expected behavior the db migrate task should run without problems reproducible example see above your environment user s environment was docker desktop for os x presumably running garden | 1 |
768,126 | 26,954,387,966 | IssuesEvent | 2023-02-08 13:59:53 | asastats/channel | https://api.github.com/repos/asastats/channel | opened | [Internal] Implement Octorand's Launchpad NFT markets | feature medium priority | https://launchpad.octorand.com/
Two collections/markets: Guardians and Tacos.
| 1.0 | [Internal] Implement Octorand's Launchpad NFT markets - https://launchpad.octorand.com/
Two collections/markets: Guardians and Tacos.
| priority | implement octorand s launchpad nft markets two collections markets guardians and tacos | 1 |
759,191 | 26,583,311,237 | IssuesEvent | 2023-01-22 18:24:27 | battlecode/galaxy | https://api.github.com/repos/battlecode/galaxy | closed | Create endpoint that returns an arbitrary team's timestamp-vs-elo-ranking | type: feature module: backend priority: p2 medium | For display in frontend's team performance charts.
Compile the data in backend. Return some mapping of timestamps-to-elo. The most straightforward one is, eg, for each scrimmage, return the scrimmage timestamp and the team's rating after. (Try to return just this data to cut down on size!) | 1.0 | Create endpoint that returns an arbitrary team's timestamp-vs-elo-ranking - For display in frontend's team performance charts.
Compile the data in backend. Return some mapping of timestamps-to-elo. The most straightforward one is, eg, for each scrimmage, return the scrimmage timestamp and the team's rating after. (Try to return just this data to cut down on size!) | priority | create endpoint that returns an arbitrary team s timestamp vs elo ranking for display in frontend s team performance charts compile the data in backend return some mapping of timestamps to elo the most straightforward one is eg for each scrimmage return the scrimmage timestamp and the team s rating after try to return just this data to cut down on size | 1 |
383,323 | 11,354,529,007 | IssuesEvent | 2020-01-24 17:50:25 | buttercup/buttercup-browser-extension | https://api.github.com/repos/buttercup/buttercup-browser-extension | closed | Selecting vault contents (entries and groups) very slow | Browser: Firefox Priority: Medium Status: Available Type: Bug | The UI is slow to react when clicking certain entries or groups in FF - v2.13.0. | 1.0 | Selecting vault contents (entries and groups) very slow - The UI is slow to react when clicking certain entries or groups in FF - v2.13.0. | priority | selecting vault contents entries and groups very slow the ui is slow to react when clicking certain entries or groups in ff | 1 |
399,661 | 11,758,658,719 | IssuesEvent | 2020-03-13 15:48:19 | garden-io/garden | https://api.github.com/repos/garden-io/garden | closed | Unhelpful error message when provider fails to resolve | bug cli priority:medium stale ux | ## Bug
### Current Behavior
When I run e.g. `garden deploy` and for some reason providers fail to resolve, I sometimes get very unhelpful log messages. Example:
```
demo-project (garden→helm3) g deploy
Deploy 🚀
providers → Getting status...
Failed resolving provider local-kubernetes. Here is the output:
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
Failed resolving one or more providers:
- local-kubernetes
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
Failed resolving one or more providers:
- local-kubernetes
See error.log for detailed error message
```
This both tells me very little and repeats the same message three times. I expect this is some recent regression, since @eysi09 recently made improvements in this area.
### Expected behavior
I'd expect some helpful text instead of just "Failed to resolve providers"
### Reproducible example
This essentially needs an internal error in the provider to come up. You could try adding a `throw` statement somewhere to trigger this, for example.
### Workaround
Just cat the error.log file, which still works fine.
### Your environment
Garden v0.10.16
| 1.0 | Unhelpful error message when provider fails to resolve - ## Bug
### Current Behavior
When I run e.g. `garden deploy` and for some reason providers fail to resolve, I sometimes get very unhelpful log messages. Example:
```
demo-project (garden→helm3) g deploy
Deploy 🚀
providers → Getting status...
Failed resolving provider local-kubernetes. Here is the output:
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
Failed resolving one or more providers:
- local-kubernetes
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
Failed resolving one or more providers:
- local-kubernetes
See error.log for detailed error message
```
This both tells me very little and repeats the same message three times. I expect this is some recent regression, since @eysi09 recently made improvements in this area.
### Expected behavior
I'd expect some helpful text instead of just "Failed to resolve providers"
### Reproducible example
This essentially needs an internal error in the provider to come up. You could try adding a `throw` statement somewhere to trigger this, for example.
### Workaround
Just cat the error.log file, which still works fine.
### Your environment
Garden v0.10.16
| priority | unhelpful error message when provider fails to resolve bug current behavior when i run e g garden deploy and for some reason providers fail to resolve i sometimes get very unhelpful log messages example demo project garden→ g deploy deploy 🚀 providers → getting status failed resolving provider local kubernetes here is the output ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ failed resolving one or more providers local kubernetes ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ failed resolving one or more providers local kubernetes see error log for detailed error message this both tells me very little and repeats the same message three times i expect this is some recent regression since recently made improvements in this area expected behavior i d expect some helpful text instead of just failed to resolve providers reproducible example this essentially needs an internal error in the provider to come up you could try adding a throw statement somewhere to trigger this for example workaround just cat the error log file which still works fine your environment garden | 1 |
43,478 | 2,889,803,515 | IssuesEvent | 2015-06-13 19:32:40 | damonkohler/android-scripting | https://api.github.com/repos/damonkohler/android-scripting | closed | Add Triggers for broadcast events | auto-migrated Priority-Medium Type-Enhancement | ```
Would like to tie a script directly to an OS level event such as 'incoming
call', where the script is only triggered (run) when the phone starts to
receive call, not constantly running or cron'd. The following hooks would be
excellent, though are not exhaustive.
onScreenAwake ( the moment after screen lock is turned on/off)
onIncomingPhoneCall ( returns number of the incoming call, current 'state'
(answered/unconnected etc)
onLocationChange (the location of the phone changed from it's previous
position. Returns the previous position and the new position
onCopy / onPaste Events.
onGSensorChange: Example when the phone is placed facedown, or faceup, this
event would fire. Optionally, would have a filter for 'when screen is on/off
filter to avoid receving events while the phone is in use (or not in use)
depending on the desired intent.
```
Original issue reported on code.google.com by `Gregory....@gmail.com` on 22 Aug 2010 at 1:38 | 1.0 | Add Triggers for broadcast events - ```
Would like to tie a script directly to an OS level event such as 'incoming
call', where the script is only triggered (run) when the phone starts to
receive call, not constantly running or cron'd. The following hooks would be
excellent, though are not exhaustive.
onScreenAwake ( the moment after screen lock is turned on/off)
onIncomingPhoneCall ( returns number of the incoming call, current 'state'
(answered/unconnected etc)
onLocationChange (the location of the phone changed from it's previous
position. Returns the previous position and the new position
onCopy / onPaste Events.
onGSensorChange: Example when the phone is placed facedown, or faceup, this
event would fire. Optionally, would have a filter for 'when screen is on/off
filter to avoid receving events while the phone is in use (or not in use)
depending on the desired intent.
```
Original issue reported on code.google.com by `Gregory....@gmail.com` on 22 Aug 2010 at 1:38 | priority | add triggers for broadcast events would like to tie a script directly to an os level event such as incoming call where the script is only triggered run when the phone starts to receive call not constantly running or cron d the following hooks would be excellent though are not exhaustive onscreenawake the moment after screen lock is turned on off onincomingphonecall returns number of the incoming call current state answered unconnected etc onlocationchange the location of the phone changed from it s previous position returns the previous position and the new position oncopy onpaste events ongsensorchange example when the phone is placed facedown or faceup this event would fire optionally would have a filter for when screen is on off filter to avoid receving events while the phone is in use or not in use depending on the desired intent original issue reported on code google com by gregory gmail com on aug at | 1 |
388,575 | 11,489,468,822 | IssuesEvent | 2020-02-11 15:33:31 | unitymakesus/tcf | https://api.github.com/repos/unitymakesus/tcf | closed | Nonconsecutive heading used (h1 -> h3) | Priority: Medium Type: a11y | This document contains an `<h3>` tag directly following an `<h1>`. In order to maintain a consistent outline of the page for assistive technologies, reduce the gap in the heading level by upgrading this tag to an `<h2>`.
Relevant Code:
```
<h3><a href="https://trianglecf.beta.unitymakes.us/award/achievement-scholarship-for-refugees-and-immigrants-of-wake-county/">Achievement Scholarship for Refugees and Immigrants of Wake County</a></h3>
```
Found on:
https://trianglecf.beta.unitymakes.us/students/scholarships-awards/ | 1.0 | Nonconsecutive heading used (h1 -> h3) - This document contains an `<h3>` tag directly following an `<h1>`. In order to maintain a consistent outline of the page for assistive technologies, reduce the gap in the heading level by upgrading this tag to an `<h2>`.
Relevant Code:
```
<h3><a href="https://trianglecf.beta.unitymakes.us/award/achievement-scholarship-for-refugees-and-immigrants-of-wake-county/">Achievement Scholarship for Refugees and Immigrants of Wake County</a></h3>
```
Found on:
https://trianglecf.beta.unitymakes.us/students/scholarships-awards/ | priority | nonconsecutive heading used this document contains an tag directly following an in order to maintain a consistent outline of the page for assistive technologies reduce the gap in the heading level by upgrading this tag to an relevant code found on | 1 |
549,039 | 16,084,386,987 | IssuesEvent | 2021-04-26 09:28:16 | AbsaOSS/enceladus | https://api.github.com/repos/AbsaOSS/enceladus | closed | Initial Spark 3 upgrade efforts | Conformance Standardization Streaming feature priority: medium | ## Background
Enceladus should migrate to use Spark 3
## Feature
The basic changes needed for migration to Spark 3.
More complex changes / problems will have separate issues.
| 1.0 | Initial Spark 3 upgrade efforts - ## Background
Enceladus should migrate to use Spark 3
## Feature
The basic changes needed for migration to Spark 3.
More complex changes / problems will have separate issues.
| priority | initial spark upgrade efforts background enceladus should migrate to use spark feature the basic changes needed for migration to spark more complex changes problems will have separate issues | 1 |
552,036 | 16,193,330,888 | IssuesEvent | 2021-05-04 11:37:53 | space-wizards/space-station-14 | https://api.github.com/repos/space-wizards/space-station-14 | closed | Vapor throws exceptions in space | Difficulty: 2 - Medium Priority: 3-low Size: 2 - Small Type: Bug | <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] -->
## Description
<!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.-->
Vapor in space (e.g. using a fire extinguisher) causes errors. | 1.0 | Vapor throws exceptions in space - <!-- To automatically tag this issue, add the uppercase label(s) surrounded by brackets below, for example: [LABEL] -->
## Description
<!-- Explain your issue in detail, including the steps to reproduce it if applicable. Issues without proper explanation are liable to be closed by maintainers.-->
Vapor in space (e.g. using a fire extinguisher) causes errors. | priority | vapor throws exceptions in space description vapor in space e g using a fire extinguisher causes errors | 1 |
714,564 | 24,566,472,862 | IssuesEvent | 2022-10-13 03:51:12 | AY2223S1-CS2113-T17-1/tp | https://api.github.com/repos/AY2223S1-CS2113-T17-1/tp | closed | As an AOM, I can delete passenger details | type.Story priority.Medium | So that I am able to delete passenger details manually.
Due Date: 10th of Oct (Monday) | 1.0 | As an AOM, I can delete passenger details - So that I am able to delete passenger details manually.
Due Date: 10th of Oct (Monday) | priority | as an aom i can delete passenger details so that i am able to delete passenger details manually due date of oct monday | 1 |
285,773 | 8,774,373,145 | IssuesEvent | 2018-12-18 19:37:32 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | [CVEDBSyncTask] dependency-check.sh --updateonly ended with 248 | SEV3-medium env/prod-preview priority/P4 status/open-for-dev team/analytics team/analytics/core type/bug | I've spotted the following on **prod-preview** couple times, so it's probably not a one-time issue:
{"task_id": "680a9129-890d-4b67-b9ed-04e1b2f64f33", "node_args": {"job_id": "cveCheckJob", "only_already_scanned": true, "ignore_modification_time": false}, "parent": {}, "queue": "STAGE_ingestion_CVEDBSyncTask_v0", "task_name": "CVEDBSyncTask", "flow_name": "cveCheckFlow", "dispatcher_id": "57ce1748-1d8a-4f40-8133-0686d58c97f2", "event": "TASK_START"}
| 2018-02-11 19:27:11,190 [WARNING] f8a_worker.utils: command ['/opt/dependency-check/bin/dependency-check.sh', '--updateonly', '--data', '/tmp/tmp152kkh6q'] ended with 248
| [INFO] Checking for updates
| [INFO] starting getUpdatesNeeded() ...
| [INFO] NVD CVE requires several updates; this could take a couple of minutes.
| [INFO] Download Started for NVD CVE - 2002
| [INFO] Download Started for NVD CVE - 2003
| [INFO] Download Started for NVD CVE - 2004
| [INFO] Download Started for NVD CVE - 2005
| [INFO] Download Started for NVD CVE - 2006
| [INFO] Download Started for NVD CVE - 2007
| [INFO] Download Complete for NVD CVE - 2003 (2475 ms)
| [INFO] Download Started for NVD CVE - 2008
| [INFO] Processing Started for NVD CVE - 2003
| [INFO] Download Complete for NVD CVE - 2005 (3803 ms)
| [INFO] Download Started for NVD CVE - 2009
| [INFO] Processing Started for NVD CVE - 2005
| [INFO] Download Complete for NVD CVE - 2004 (4683 ms)
| [INFO] Download Started for NVD CVE - 2010
| [INFO] Processing Started for NVD CVE - 2004
| [INFO] Download Complete for NVD CVE - 2002 (5286 ms)
| [INFO] Download Started for NVD CVE - 2011
| [INFO] Processing Started for NVD CVE - 2002
| [INFO] Download Complete for NVD CVE - 2007 (9501 ms)
| [INFO] Download Complete for NVD CVE - 2006 (10006 ms)
| [INFO] Download Started for NVD CVE - 2012
| [INFO] Download Started for NVD CVE - 2013
| [INFO] Download Complete for NVD CVE - 2008 (9200 ms)
| [INFO] Download Started for NVD CVE - 2014
| [INFO] Download Complete for NVD CVE - 2009 (10914 ms)
| [INFO] Download Started for NVD CVE - 2015
| [INFO] Download Complete for NVD CVE - 2010 (12292 ms)
| [INFO] Download Started for NVD CVE - 2016
| [WARN] An exception occurred downloading NVD CVE - 2013
| Some CVEs may not be reported.
| [INFO] Download Started for NVD CVE - 2017
| [WARN] An exception occurred downloading NVD CVE - 2012
| Some CVEs may not be reported.
| [INFO] Download Started for NVD CVE - 2018
| [WARN] An exception occurred downloading NVD CVE - 2014
| Some CVEs may not be reported.
| [INFO] Download Started for NVD CVE - Modified
| [WARN] An exception occurred downloading NVD CVE - 2011
| Some CVEs may not be reported.
| [WARN] An exception occurred downloading NVD CVE - 2015
| Some CVEs may not be reported.
| [WARN] Download Failed for NVD CVE - 2017
| Some CVEs may not be reported.
| [INFO] If you are behind a proxy you may need to configure dependency-check to use the proxy.
| [INFO] Processing Started for NVD CVE - 2007
| [INFO] Processing Started for NVD CVE - 2006
| [INFO] Processing Started for NVD CVE - 2008
| [INFO] Processing Started for NVD CVE - 2009
| [WARN] An exception occurred downloading NVD CVE - 2018
| Some CVEs may not be reported.
| [WARN] An exception occurred downloading NVD CVE - 2016
| Some CVEs may not be reported.
| [ERROR] The download was interrupted; unable to complete the update
| [INFO] Download Complete for NVD CVE - Modified (2801 ms)
| [WARN] An exception occurred downloading NVD CVE - Modified
| Some CVEs may not be reported.
EDIT: on dev cluster the process is simply killed (unknown reasons) without any errors | 1.0 | [CVEDBSyncTask] dependency-check.sh --updateonly ended with 248 - I've spotted the following on **prod-preview** couple times, so it's probably not a one-time issue:
{"task_id": "680a9129-890d-4b67-b9ed-04e1b2f64f33", "node_args": {"job_id": "cveCheckJob", "only_already_scanned": true, "ignore_modification_time": false}, "parent": {}, "queue": "STAGE_ingestion_CVEDBSyncTask_v0", "task_name": "CVEDBSyncTask", "flow_name": "cveCheckFlow", "dispatcher_id": "57ce1748-1d8a-4f40-8133-0686d58c97f2", "event": "TASK_START"}
| 2018-02-11 19:27:11,190 [WARNING] f8a_worker.utils: command ['/opt/dependency-check/bin/dependency-check.sh', '--updateonly', '--data', '/tmp/tmp152kkh6q'] ended with 248
| [INFO] Checking for updates
| [INFO] starting getUpdatesNeeded() ...
| [INFO] NVD CVE requires several updates; this could take a couple of minutes.
| [INFO] Download Started for NVD CVE - 2002
| [INFO] Download Started for NVD CVE - 2003
| [INFO] Download Started for NVD CVE - 2004
| [INFO] Download Started for NVD CVE - 2005
| [INFO] Download Started for NVD CVE - 2006
| [INFO] Download Started for NVD CVE - 2007
| [INFO] Download Complete for NVD CVE - 2003 (2475 ms)
| [INFO] Download Started for NVD CVE - 2008
| [INFO] Processing Started for NVD CVE - 2003
| [INFO] Download Complete for NVD CVE - 2005 (3803 ms)
| [INFO] Download Started for NVD CVE - 2009
| [INFO] Processing Started for NVD CVE - 2005
| [INFO] Download Complete for NVD CVE - 2004 (4683 ms)
| [INFO] Download Started for NVD CVE - 2010
| [INFO] Processing Started for NVD CVE - 2004
| [INFO] Download Complete for NVD CVE - 2002 (5286 ms)
| [INFO] Download Started for NVD CVE - 2011
| [INFO] Processing Started for NVD CVE - 2002
| [INFO] Download Complete for NVD CVE - 2007 (9501 ms)
| [INFO] Download Complete for NVD CVE - 2006 (10006 ms)
| [INFO] Download Started for NVD CVE - 2012
| [INFO] Download Started for NVD CVE - 2013
| [INFO] Download Complete for NVD CVE - 2008 (9200 ms)
| [INFO] Download Started for NVD CVE - 2014
| [INFO] Download Complete for NVD CVE - 2009 (10914 ms)
| [INFO] Download Started for NVD CVE - 2015
| [INFO] Download Complete for NVD CVE - 2010 (12292 ms)
| [INFO] Download Started for NVD CVE - 2016
| [WARN] An exception occurred downloading NVD CVE - 2013
| Some CVEs may not be reported.
| [INFO] Download Started for NVD CVE - 2017
| [WARN] An exception occurred downloading NVD CVE - 2012
| Some CVEs may not be reported.
| [INFO] Download Started for NVD CVE - 2018
| [WARN] An exception occurred downloading NVD CVE - 2014
| Some CVEs may not be reported.
| [INFO] Download Started for NVD CVE - Modified
| [WARN] An exception occurred downloading NVD CVE - 2011
| Some CVEs may not be reported.
| [WARN] An exception occurred downloading NVD CVE - 2015
| Some CVEs may not be reported.
| [WARN] Download Failed for NVD CVE - 2017
| Some CVEs may not be reported.
| [INFO] If you are behind a proxy you may need to configure dependency-check to use the proxy.
| [INFO] Processing Started for NVD CVE - 2007
| [INFO] Processing Started for NVD CVE - 2006
| [INFO] Processing Started for NVD CVE - 2008
| [INFO] Processing Started for NVD CVE - 2009
| [WARN] An exception occurred downloading NVD CVE - 2018
| Some CVEs may not be reported.
| [WARN] An exception occurred downloading NVD CVE - 2016
| Some CVEs may not be reported.
| [ERROR] The download was interrupted; unable to complete the update
| [INFO] Download Complete for NVD CVE - Modified (2801 ms)
| [WARN] An exception occurred downloading NVD CVE - Modified
| Some CVEs may not be reported.
EDIT: on dev cluster the process is simply killed (unknown reasons) without any errors | priority | dependency check sh updateonly ended with i ve spotted the following on prod preview couple times so it s probably not a one time issue task id node args job id cvecheckjob only already scanned true ignore modification time false parent queue stage ingestion cvedbsynctask task name cvedbsynctask flow name cvecheckflow dispatcher id event task start worker utils command ended with checking for updates starting getupdatesneeded nvd cve requires several updates this could take a couple of minutes download started for nvd cve download started for nvd cve download started for nvd cve download started for nvd cve download started for nvd cve download started for nvd cve download complete for nvd cve ms download started for nvd cve processing started for nvd cve download complete for nvd cve ms download started for nvd cve processing started for nvd cve download complete for nvd cve ms download started for nvd cve processing started for nvd cve download complete for nvd cve ms download started for nvd cve processing started for nvd cve download complete for nvd cve ms download complete for nvd cve ms download started for nvd cve download started for nvd cve download complete for nvd cve ms download started for nvd cve download complete for nvd cve ms download started for nvd cve download complete for nvd cve ms download started for nvd cve an exception occurred downloading nvd cve some cves may not be reported download started for nvd cve an exception occurred downloading nvd cve some cves may not be reported download started for nvd cve an exception occurred downloading nvd cve some cves may not be reported download started for nvd cve modified an exception occurred downloading nvd cve some cves may not be reported an exception occurred downloading nvd cve some cves may not be reported download failed for nvd cve some cves may not be reported if you are behind a proxy you may need to configure dependency check to use the proxy processing started for nvd cve processing started for nvd cve processing started for nvd cve processing started for nvd cve an exception occurred downloading nvd cve some cves may not be reported an exception occurred downloading nvd cve some cves may not be reported the download was interrupted unable to complete the update download complete for nvd cve modified ms an exception occurred downloading nvd cve modified some cves may not be reported edit on dev cluster the process is simply killed unknown reasons without any errors | 1 |
595,844 | 18,075,984,696 | IssuesEvent | 2021-09-21 09:56:15 | vanjarosoftware/Vanjaro.Platform | https://api.github.com/repos/vanjarosoftware/Vanjaro.Platform | closed | Missing Next and Previous buttons in Pagination | Enhancement Release: Minor Priority: Medium Area: Backend | Missing Next and Previous buttons in Pagination | 1.0 | Missing Next and Previous buttons in Pagination - Missing Next and Previous buttons in Pagination | priority | missing next and previous buttons in pagination missing next and previous buttons in pagination | 1 |
575,219 | 17,024,703,348 | IssuesEvent | 2021-07-03 08:32:01 | Materials-Consortia/optimade-python-tools | https://api.github.com/repos/Materials-Consortia/optimade-python-tools | closed | Add timeout parameter to validator | priority/medium validator | Currently the validator has no timeout on requests. This should be user-controlled and given a generous default value (~5 mins). | 1.0 | Add timeout parameter to validator - Currently the validator has no timeout on requests. This should be user-controlled and given a generous default value (~5 mins). | priority | add timeout parameter to validator currently the validator has no timeout on requests this should be user controlled and given a generous default value mins | 1 |
471,044 | 13,552,256,667 | IssuesEvent | 2020-09-17 12:21:25 | silentium-labs/merlin-gql | https://api.github.com/repos/silentium-labs/merlin-gql | opened | Refactor code generator | Priority: Medium Status: In Progress Type: Enhancement | Currently, the generator has a single method for 5 different types of files:
- Entities
- Resolvers
- InputTypes
- FilterTypes
- SortTypes
We should split this into 5 separate and more maintainable methods that can be chained or called individually when needed.
I suggest to do the same with the handlebars helpers, split them into common helpers, and specific helpers according to what's needed for each generation. | 1.0 | Refactor code generator - Currently, the generator has a single method for 5 different types of files:
- Entities
- Resolvers
- InputTypes
- FilterTypes
- SortTypes
We should split this into 5 separate and more maintainable methods that can be chained or called individually when needed.
I suggest to do the same with the handlebars helpers, split them into common helpers, and specific helpers according to what's needed for each generation. | priority | refactor code generator currently the generator has a single method for different types of files entities resolvers inputtypes filtertypes sorttypes we should split this into separate and more maintainable methods that can be chained or called individually when needed i suggest to do the same with the handlebars helpers split them into common helpers and specific helpers according to what s needed for each generation | 1 |
758,718 | 26,566,372,837 | IssuesEvent | 2023-01-20 20:40:22 | ooni/probe | https://api.github.com/repos/ooni/probe | closed | android: fix broken tests | bug priority/medium platform/android interrupt | These logs show the tests that fail:
```
org.openobservatory.ooniprobe.test.TestAsyncTaskTest > runTest_withError FAILED
org.mockito.exceptions.verification.WantedButNotInvoked at TestAsyncTaskTest.java:188
org.openobservatory.ooniprobe.test.TestAsyncTaskTest > runTest_withProgress FAILED
org.mockito.exceptions.verification.WantedButNotInvoked at TestAsyncTaskTest.java:165
```
See https://github.com/ooni/probe-android/actions/runs/3647279332/jobs/6159319329 | 1.0 | android: fix broken tests - These logs show the tests that fail:
```
org.openobservatory.ooniprobe.test.TestAsyncTaskTest > runTest_withError FAILED
org.mockito.exceptions.verification.WantedButNotInvoked at TestAsyncTaskTest.java:188
org.openobservatory.ooniprobe.test.TestAsyncTaskTest > runTest_withProgress FAILED
org.mockito.exceptions.verification.WantedButNotInvoked at TestAsyncTaskTest.java:165
```
See https://github.com/ooni/probe-android/actions/runs/3647279332/jobs/6159319329 | priority | android fix broken tests these logs show the tests that fail org openobservatory ooniprobe test testasynctasktest runtest witherror failed org mockito exceptions verification wantedbutnotinvoked at testasynctasktest java org openobservatory ooniprobe test testasynctasktest runtest withprogress failed org mockito exceptions verification wantedbutnotinvoked at testasynctasktest java see | 1 |
320,916 | 9,791,106,375 | IssuesEvent | 2019-06-10 14:20:32 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | Monolane Improvement - Multilane | priority: medium team: automotive type: feature request | Improve monolane to support segments with multiple lanes of constant lateral offset from a reference. The only thing we care about is representing a highway on & off ramp single lane merging onto (off of) a trivial multi-lane highway (without any lane drops, etc.). | 1.0 | Monolane Improvement - Multilane - Improve monolane to support segments with multiple lanes of constant lateral offset from a reference. The only thing we care about is representing a highway on & off ramp single lane merging onto (off of) a trivial multi-lane highway (without any lane drops, etc.). | priority | monolane improvement multilane improve monolane to support segments with multiple lanes of constant lateral offset from a reference the only thing we care about is representing a highway on off ramp single lane merging onto off of a trivial multi lane highway without any lane drops etc | 1 |
649,564 | 21,315,820,113 | IssuesEvent | 2022-04-16 08:52:32 | dnnsoftware/Dnn.Platform | https://api.github.com/repos/dnnsoftware/Dnn.Platform | closed | Edit page rights to all users | Type: Enhancement stale Effort: Medium Priority: Medium Status: Ready for Development | 99% of the time an admin user would **not** want to give edit rights to all users on a page (or module)
It is a feature and a responsibility of the user but a mistake is made too easy now.
## Description of solution
Warning when setting the checkbox for all users on edit.
## Description of alternatives considered
Warning in the security analyzer so the admin that checks this can contact the one responsible if needed or take action.
| 1.0 | Edit page rights to all users - 99% of the time an admin user would **not** want to give edit rights to all users on a page (or module)
It is a feature and a responsibility of the user but a mistake is made too easy now.
## Description of solution
Warning when setting the checkbox for all users on edit.
## Description of alternatives considered
Warning in the security analyzer so the admin that checks this can contact the one responsible if needed or take action.
| priority | edit page rights to all users of the time an admin user would not want to give edit rights to all users on a page or module it is a feature and a responsibility of the user but a mistake is made too easy now description of solution warning when setting the checkbox for all users on edit description of alternatives considered warning in the security analyzer so the admin that checks this can contact the one responsible if needed or take action | 1 |
206,501 | 7,112,853,861 | IssuesEvent | 2018-01-17 18:24:45 | vladgh/docker_base_images | https://api.github.com/repos/vladgh/docker_base_images | closed | [backup] Avoid writing temporary files to disk | Priority: Medium Type: Enhancement | Backup script should avoid writing anything to disk, and use pipes instead (tar | gpg | s3). | 1.0 | [backup] Avoid writing temporary files to disk - Backup script should avoid writing anything to disk, and use pipes instead (tar | gpg | s3). | priority | avoid writing temporary files to disk backup script should avoid writing anything to disk and use pipes instead tar gpg | 1 |
618,475 | 19,471,439,591 | IssuesEvent | 2021-12-24 02:18:19 | foolip/mdn-bcd-collector | https://api.github.com/repos/foolip/mdn-bcd-collector | closed | Custom tests (alternative names, odd test cases, etc.) | Priority 3 (Medium) Component: Custom | In `build.js`, we should implement a dictionary (whether hard-coded or a JSON file) that includes code for custom tests on different features. This way, we can test for things such as alternative names, or APIs accessible through other APIs (like `ANGLE_instanced_arrays`).
- [x] Custom tests for APIs and its members (#341)
- [x] Custom tests for CSS (#347)
- [ ] Multiple test variations (may eliminate need for the two tasks below)
- [ ] Browser-specific alt. tests (for alternative names)?
- [ ] Known alt. names? | 1.0 | Custom tests (alternative names, odd test cases, etc.) - In `build.js`, we should implement a dictionary (whether hard-coded or a JSON file) that includes code for custom tests on different features. This way, we can test for things such as alternative names, or APIs accessible through other APIs (like `ANGLE_instanced_arrays`).
- [x] Custom tests for APIs and its members (#341)
- [x] Custom tests for CSS (#347)
- [ ] Multiple test variations (may eliminate need for the two tasks below)
- [ ] Browser-specific alt. tests (for alternative names)?
- [ ] Known alt. names? | priority | custom tests alternative names odd test cases etc in build js we should implement a dictionary whether hard coded or a json file that includes code for custom tests on different features this way we can test for things such as alternative names or apis accessible through other apis like angle instanced arrays custom tests for apis and its members custom tests for css multiple test variations may eliminate need for the two tasks below browser specific alt tests for alternative names known alt names | 1 |
378,188 | 11,198,349,229 | IssuesEvent | 2020-01-03 15:43:37 | dhenry-KCI/FredCo-Post-Go-Live- | https://api.github.com/repos/dhenry-KCI/FredCo-Post-Go-Live- | opened | STOPNOT Log Functionality - Need to Add Correspondence to this | Medium Priority | I spoke with Christian to confirm whether or not the STOPNOT log prevents notifications AND correspondence from sending to applicants. He does not believe the code behind this log stops correspondence.
We need to add the functionality to this log to stop correspondence from sending as well.

| 1.0 | STOPNOT Log Functionality - Need to Add Correspondence to this - I spoke with Christian to confirm whether or not the STOPNOT log prevents notifications AND correspondence from sending to applicants. He does not believe the code behind this log stops correspondence.
We need to add the functionality to this log to stop correspondence from sending as well.

| priority | stopnot log functionality need to add correspondence to this i spoke with christian to confirm whether or not the stopnot log prevents notifications and correspondence from sending to applicants he does not believe the code behind this log stops correspondence we need to add the functionality to this log to stop correspondence from sending as well | 1 |
540,511 | 15,812,832,139 | IssuesEvent | 2021-04-05 06:30:40 | reconness/reconness-frontend | https://api.github.com/repos/reconness/reconness-frontend | closed | Missing Remove buttons in Add Subdomain popup | bug priority: medium severity: minor | ### Expected scenario
**Clicking on Remove button (red round button), will remove all the row from the popup.**

### Current scenario

| 1.0 | Missing Remove buttons in Add Subdomain popup - ### Expected scenario
**Clicking on Remove button (red round button), will remove all the row from the popup.**

### Current scenario

| priority | missing remove buttons in add subdomain popup expected scenario clicking on remove button red round button will remove all the row from the popup current scenario | 1 |
489,042 | 14,100,336,186 | IssuesEvent | 2020-11-06 03:53:25 | hydroshare/hydroshare | https://api.github.com/repos/hydroshare/hydroshare | closed | Discover page: sorting of facets is counter-productive | Discover Medium Priority | The discovery sorting of facets by number of instances is very confusing; alphabetical order recommended. | 1.0 | Discover page: sorting of facets is counter-productive - The discovery sorting of facets by number of instances is very confusing; alphabetical order recommended. | priority | discover page sorting of facets is counter productive the discovery sorting of facets by number of instances is very confusing alphabetical order recommended | 1 |
813,721 | 30,469,112,108 | IssuesEvent | 2023-07-17 12:32:33 | asastats/channel | https://api.github.com/repos/asastats/channel | closed | [Internal] Old ALGOxNFT listing is not visible | bug medium priority | Old as in "more than 18 months old" as they changed their contracts around that time. | 1.0 | [Internal] Old ALGOxNFT listing is not visible - Old as in "more than 18 months old" as they changed their contracts around that time. | priority | old algoxnft listing is not visible old as in more than months old as they changed their contracts around that time | 1 |
576,822 | 17,095,630,418 | IssuesEvent | 2021-07-09 01:49:19 | visit-dav/visit | https://api.github.com/repos/visit-dav/visit | closed | build_visit cannot build cgns (3.2RC) | bug impact medium likelihood medium priority reviewed | ### Describe the bug
Using build_visit from 3.2RC, the build of cgns fails. From the build_visit log file:
```
Building CGNS (~2 minutes)
prepare_build_dir: CGNS-4.1.0/src CGNS-4.1.0.tar.gz
Unzipping/Untarring CGNS-4.1.0.tar.gz . . .
Patching CGNS 4.1.0
patching file CGNS-4.1.0/src/configure
patching file CGNS-4.1.0/src/Makefile.in
Configuring CGNS . . .
checking CGNS version... 4.10
checking if gcc is requested... no
checking for gcc... gcc
checking whether the C compiler works... yes
checking for C compiler default output file name... a.out
checking for suffix of executables...
checking whether we are cross compiling... configure: error: in `/usr/WS1/kbonnell/BuildVisIt/320_gcc6/CGNS-4.1.0/src':
configure: error: cannot run C compiled programs.
If you meant to cross compile, use `--host'.
See `config.log' for more details
```
The config.log from CGNS:
[config.log.txt](https://github.com/visit-dav/visit/files/6749921/config.log.txt)
Could be related to changes that were made for getting CGNS to build on MacOS.
### Desktop
- OS and version: linux
- VisIt Version: 3.2RC
| 1.0 | build_visit cannot build cgns (3.2RC) - ### Describe the bug
Using build_visit from 3.2RC, the build of cgns fails. From the build_visit log file:
```
Building CGNS (~2 minutes)
prepare_build_dir: CGNS-4.1.0/src CGNS-4.1.0.tar.gz
Unzipping/Untarring CGNS-4.1.0.tar.gz . . .
Patching CGNS 4.1.0
patching file CGNS-4.1.0/src/configure
patching file CGNS-4.1.0/src/Makefile.in
Configuring CGNS . . .
checking CGNS version... 4.10
checking if gcc is requested... no
checking for gcc... gcc
checking whether the C compiler works... yes
checking for C compiler default output file name... a.out
checking for suffix of executables...
checking whether we are cross compiling... configure: error: in `/usr/WS1/kbonnell/BuildVisIt/320_gcc6/CGNS-4.1.0/src':
configure: error: cannot run C compiled programs.
If you meant to cross compile, use `--host'.
See `config.log' for more details
```
The config.log from CGNS:
[config.log.txt](https://github.com/visit-dav/visit/files/6749921/config.log.txt)
Could be related to changes that were made for getting CGNS to build on MacOS.
### Desktop
- OS and version: linux
- VisIt Version: 3.2RC
| priority | build visit cannot build cgns describe the bug using build visit from the build of cgns fails from the build visit log file building cgns minutes prepare build dir cgns src cgns tar gz unzipping untarring cgns tar gz patching cgns patching file cgns src configure patching file cgns src makefile in configuring cgns checking cgns version checking if gcc is requested no checking for gcc gcc checking whether the c compiler works yes checking for c compiler default output file name a out checking for suffix of executables checking whether we are cross compiling configure error in usr kbonnell buildvisit cgns src configure error cannot run c compiled programs if you meant to cross compile use host see config log for more details the config log from cgns could be related to changes that were made for getting cgns to build on macos desktop os and version linux visit version | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.