Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
1,048
2,507,024,719
IssuesEvent
2015-01-12 15:39:29
G-Node/GCA-Web
https://api.github.com/repos/G-Node/GCA-Web
closed
Editor does not work in IE
bug high priority
The editor does not update the abstract properly in internet explorer (version 11 and others).
1.0
Editor does not work in IE - The editor does not update the abstract properly in internet explorer (version 11 and others).
priority
editor does not work in ie the editor does not update the abstract properly in internet explorer version and others
1
239,601
7,799,873,036
IssuesEvent
2018-06-09 01:30:38
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0005720: 'my contacts' favorite does not work correctly
Addressbook Bug Mantis high priority
**Reported by pschuele on 15 Feb 2012 14:41** **Version:** Milan (2012-03) Beta 4 'my contacts' favorite does not work correctly it looks like the server does not send the correct node info, the users displayname is shown in the filter instead of the node/path ...
1.0
0005720: 'my contacts' favorite does not work correctly - **Reported by pschuele on 15 Feb 2012 14:41** **Version:** Milan (2012-03) Beta 4 'my contacts' favorite does not work correctly it looks like the server does not send the correct node info, the users displayname is shown in the filter instead of the node/path ...
priority
my contacts favorite does not work correctly reported by pschuele on feb version milan beta my contacts favorite does not work correctly it looks like the server does not send the correct node info the users displayname is shown in the filter instead of the node path
1
390,186
11,527,787,139
IssuesEvent
2020-02-16 00:23:26
tempor1s/oracli
https://api.github.com/repos/tempor1s/oracli
closed
Require user auth to be pulled from header
enhancement help wanted high priority
Currently, you need to send the token over the body, but I would prefer if we can move this over to the token being sent in the header as it is a much cleaner implementation of authentication :)
1.0
Require user auth to be pulled from header - Currently, you need to send the token over the body, but I would prefer if we can move this over to the token being sent in the header as it is a much cleaner implementation of authentication :)
priority
require user auth to be pulled from header currently you need to send the token over the body but i would prefer if we can move this over to the token being sent in the header as it is a much cleaner implementation of authentication
1
712,312
24,490,137,051
IssuesEvent
2022-10-09 23:48:10
python/mypy
https://api.github.com/repos/python/mypy
closed
Give an error if there is a variable annotation within a function but no signature
feature priority-0-high topic-usability good-first-issue
Mypy should perhaps give an error about a variable annotation in an otherwise unanotated function. Example: ```py def f(): a: int = 'x' # Maybe this annotation should be an error? ``` The rationale is that mypy will ignore the type annotation since the function is still considered unannotated, but this is confusing because there is an annotation *within* the function so it can appear to be annotated. If `--check-untyped-defs` is being used this error shouldn't be generated. Originally reported in #3945.
1.0
Give an error if there is a variable annotation within a function but no signature - Mypy should perhaps give an error about a variable annotation in an otherwise unanotated function. Example: ```py def f(): a: int = 'x' # Maybe this annotation should be an error? ``` The rationale is that mypy will ignore the type annotation since the function is still considered unannotated, but this is confusing because there is an annotation *within* the function so it can appear to be annotated. If `--check-untyped-defs` is being used this error shouldn't be generated. Originally reported in #3945.
priority
give an error if there is a variable annotation within a function but no signature mypy should perhaps give an error about a variable annotation in an otherwise unanotated function example py def f a int x maybe this annotation should be an error the rationale is that mypy will ignore the type annotation since the function is still considered unannotated but this is confusing because there is an annotation within the function so it can appear to be annotated if check untyped defs is being used this error shouldn t be generated originally reported in
1
716,371
24,630,274,496
IssuesEvent
2022-10-17 00:57:10
adisve/tumble-for-kronox
https://api.github.com/repos/adisve/tumble-for-kronox
closed
German and French translation
enhancement High Priority
Update to the translations with the new strings that were added.
1.0
German and French translation - Update to the translations with the new strings that were added.
priority
german and french translation update to the translations with the new strings that were added
1
260,961
8,221,636,317
IssuesEvent
2018-09-06 03:06:55
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] items are not deleted in stage on delete
CI bug priority: high
### Expected behavior Deletes should be pushed to both live and staging ### Actual behavior In 3.0.x delete is an immediate deploy to live. When staging is enabled, the delete takes place on live but not staging. ### Steps to reproduce the problem * Create an a page * Publish it to staging * Publish it to live * Delete the item * Not that it still exists in staging ### Log/stack trace (use https://gist.github.com) N/A ### Specs #### Version Studio Version Number: 3.0.16-SNAPSHOT-4c989c Build Number: 4c989c7cb50201155c988637a6db453257b7a1cc Build Date/Time: 08-02-2018 14:57:36 -0400 #### OS Any #### Browser Any
1.0
[studio] items are not deleted in stage on delete - ### Expected behavior Deletes should be pushed to both live and staging ### Actual behavior In 3.0.x delete is an immediate deploy to live. When staging is enabled, the delete takes place on live but not staging. ### Steps to reproduce the problem * Create an a page * Publish it to staging * Publish it to live * Delete the item * Not that it still exists in staging ### Log/stack trace (use https://gist.github.com) N/A ### Specs #### Version Studio Version Number: 3.0.16-SNAPSHOT-4c989c Build Number: 4c989c7cb50201155c988637a6db453257b7a1cc Build Date/Time: 08-02-2018 14:57:36 -0400 #### OS Any #### Browser Any
priority
items are not deleted in stage on delete expected behavior deletes should be pushed to both live and staging actual behavior in x delete is an immediate deploy to live when staging is enabled the delete takes place on live but not staging steps to reproduce the problem create an a page publish it to staging publish it to live delete the item not that it still exists in staging log stack trace use n a specs version studio version number snapshot build number build date time os any browser any
1
290,459
8,895,463,430
IssuesEvent
2019-01-16 08:45:24
telstra/open-kilda
https://api.github.com/repos/telstra/open-kilda
closed
New API: Reroute all flows which go through a particular ISL
area/api area/arch feature priority/2-high
Initiate a reroute for all flows on this ISL. API: PATCH `links/flows/reroute?<params>` Related: #1548 (Maintenance mode for ISL)
1.0
New API: Reroute all flows which go through a particular ISL - Initiate a reroute for all flows on this ISL. API: PATCH `links/flows/reroute?<params>` Related: #1548 (Maintenance mode for ISL)
priority
new api reroute all flows which go through a particular isl initiate a reroute for all flows on this isl api patch links flows reroute related maintenance mode for isl
1
369,920
10,919,947,713
IssuesEvent
2019-11-21 20:08:49
flextype/flextype
https://api.github.com/repos/flextype/flextype
closed
Flextype Core: Forms - add ability to hide fieldsets from entries type select
priority: high type: feature
We should have ability to hide fieldsets from entries type select, this will be useful for nested fieldsets. Possible we should add new property `hide` - by default is `false` Logic: if **hide** property is **true** then hide fieldsets from entries type select. if **hide** property is **false** then show fieldsets from entries type select. if **hide** property is **is not exists** then show fieldsets from entries type select.
1.0
Flextype Core: Forms - add ability to hide fieldsets from entries type select - We should have ability to hide fieldsets from entries type select, this will be useful for nested fieldsets. Possible we should add new property `hide` - by default is `false` Logic: if **hide** property is **true** then hide fieldsets from entries type select. if **hide** property is **false** then show fieldsets from entries type select. if **hide** property is **is not exists** then show fieldsets from entries type select.
priority
flextype core forms add ability to hide fieldsets from entries type select we should have ability to hide fieldsets from entries type select this will be useful for nested fieldsets possible we should add new property hide by default is false logic if hide property is true then hide fieldsets from entries type select if hide property is false then show fieldsets from entries type select if hide property is is not exists then show fieldsets from entries type select
1
139,443
5,375,479,836
IssuesEvent
2017-02-23 05:02:26
ArctosDB/documentation-wiki
https://api.github.com/repos/ArctosDB/documentation-wiki
opened
remove hyphens from anchor links when creating from <h3>subtitles
bug Priority: High question
Here's the problem: in the current site the anchors were not necessarily the subheader/subtitle phrase: https://arctosdb.org/documentation/agent/#namesearch BUT the subheader is "Searching Agents" Opt#1 I can go through and update the subheaders to match but in some cases like this one is leaves wtih some awkward phrasing. Opt#2? Anyway to create anchors not from <h3>tags? In any case we need to get rid of hyphens when constructing the anchors from subtitles (readability pfft!) This blocks the launch if we dont have a solid fix
1.0
remove hyphens from anchor links when creating from <h3>subtitles - Here's the problem: in the current site the anchors were not necessarily the subheader/subtitle phrase: https://arctosdb.org/documentation/agent/#namesearch BUT the subheader is "Searching Agents" Opt#1 I can go through and update the subheaders to match but in some cases like this one is leaves wtih some awkward phrasing. Opt#2? Anyway to create anchors not from <h3>tags? In any case we need to get rid of hyphens when constructing the anchors from subtitles (readability pfft!) This blocks the launch if we dont have a solid fix
priority
remove hyphens from anchor links when creating from subtitles here s the problem in the current site the anchors were not necessarily the subheader subtitle phrase but the subheader is searching agents opt i can go through and update the subheaders to match but in some cases like this one is leaves wtih some awkward phrasing opt anyway to create anchors not from tags in any case we need to get rid of hyphens when constructing the anchors from subtitles readability pfft this blocks the launch if we dont have a solid fix
1
396,193
11,705,177,872
IssuesEvent
2020-03-07 14:22:28
localstack/localstack
https://api.github.com/repos/localstack/localstack
closed
Missing required parameter in input: "FunctionName"
bug needs-triaging priority-high
While creating a lambda using CFN, my command fails with the following message: ``` 2019-11-08 15:52:29,358:API: Error on request: Traceback (most recent call last): File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 304, in run_wsgi execute(self.server.app) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 292, in execute application_iter = app(environ, start_response) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/server.py", line 132, in __call__ return backend_app(environ, start_response) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__ return self.wsgi_app(environ, start_response) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app response = self.handle_exception(e) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function return cors_after_request(app.make_response(f(*args, **kwargs))) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1741, in handle_exception reraise(exc_type, exc_value, tb) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise raise value File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app response = self.full_dispatch_request() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request rv = self.handle_user_exception(e) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function return cors_after_request(app.make_response(f(*args, **kwargs))) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception reraise(exc_type, exc_value, tb) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise raise value File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request rv = self.dispatch_request() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 140, in __call__ result = self.callback(request, request.url, {}) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 168, in dispatch return cls()._dispatch(*args, **kwargs) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 259, in _dispatch return self.call_action() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 340, in call_action response = method() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 264, in _wrapper response = f(*args, **kwargs) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/responses.py", line 108, in create_change_set change_set_type=update_or_create, File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 452, in create_change_set cross_stack_resources=self.exports File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 295, in __init__ create_change_set=True, File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 187, in __init__ self.resource_map = self._create_resource_map() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 195, in _create_resource_map resource_map.create() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 482, in create if isinstance(self[resource], ec2_models.TaggedEC2Resource): File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 415, in __getitem__ resource_logical_id, resource_json, self, self._region_name) File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 173, in parse_and_create_resource return _parse_and_create_resource(logical_id, resource_json, resources_map, region_name) File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 260, in _parse_and_create_resource result = deploy_func(logical_id, resource_wrapped, stack_name=stack_name) File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 671, in deploy_resource result = deploy_resource_via_sdk_function(resource_id, resources, resource_type, func, stack_name) File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 741, in deploy_resource_via_sdk_function raise e File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 738, in deploy_resource_via_sdk_function result = function(**params) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 357, in _api_call return self._make_api_call(operation_name, kwargs) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 634, in _make_api_call api_params, operation_model, context=request_context) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 682, in _convert_to_request_dict api_params, operation_model) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/validate.py", line 297, in serialize_to_request raise ParamValidationError(report=report.generate_report()) botocore.exceptions.ParamValidationError: Parameter validation failed: Missing required parameter in input: "FunctionName" ``` Considering the fact that the `FunctionName` property is not mandatory from CFN perspective (it's even worse, since having it or not has significant implications), is it possible to create a name on the fly?
1.0
Missing required parameter in input: "FunctionName" - While creating a lambda using CFN, my command fails with the following message: ``` 2019-11-08 15:52:29,358:API: Error on request: Traceback (most recent call last): File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 304, in run_wsgi execute(self.server.app) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 292, in execute application_iter = app(environ, start_response) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/server.py", line 132, in __call__ return backend_app(environ, start_response) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__ return self.wsgi_app(environ, start_response) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app response = self.handle_exception(e) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function return cors_after_request(app.make_response(f(*args, **kwargs))) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1741, in handle_exception reraise(exc_type, exc_value, tb) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise raise value File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app response = self.full_dispatch_request() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request rv = self.handle_user_exception(e) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function return cors_after_request(app.make_response(f(*args, **kwargs))) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception reraise(exc_type, exc_value, tb) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise raise value File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request rv = self.dispatch_request() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 140, in __call__ result = self.callback(request, request.url, {}) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 168, in dispatch return cls()._dispatch(*args, **kwargs) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 259, in _dispatch return self.call_action() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 340, in call_action response = method() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 264, in _wrapper response = f(*args, **kwargs) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/responses.py", line 108, in create_change_set change_set_type=update_or_create, File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 452, in create_change_set cross_stack_resources=self.exports File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 295, in __init__ create_change_set=True, File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 187, in __init__ self.resource_map = self._create_resource_map() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 195, in _create_resource_map resource_map.create() File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 482, in create if isinstance(self[resource], ec2_models.TaggedEC2Resource): File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 415, in __getitem__ resource_logical_id, resource_json, self, self._region_name) File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 173, in parse_and_create_resource return _parse_and_create_resource(logical_id, resource_json, resources_map, region_name) File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 260, in _parse_and_create_resource result = deploy_func(logical_id, resource_wrapped, stack_name=stack_name) File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 671, in deploy_resource result = deploy_resource_via_sdk_function(resource_id, resources, resource_type, func, stack_name) File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 741, in deploy_resource_via_sdk_function raise e File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 738, in deploy_resource_via_sdk_function result = function(**params) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 357, in _api_call return self._make_api_call(operation_name, kwargs) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 634, in _make_api_call api_params, operation_model, context=request_context) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 682, in _convert_to_request_dict api_params, operation_model) File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/validate.py", line 297, in serialize_to_request raise ParamValidationError(report=report.generate_report()) botocore.exceptions.ParamValidationError: Parameter validation failed: Missing required parameter in input: "FunctionName" ``` Considering the fact that the `FunctionName` property is not mandatory from CFN perspective (it's even worse, since having it or not has significant implications), is it possible to create a name on the fly?
priority
missing required parameter in input functionname while creating a lambda using cfn my command fails with the following message api error on request traceback most recent call last file opt code localstack venv lib site packages werkzeug serving py line in run wsgi execute self server app file opt code localstack venv lib site packages werkzeug serving py line in execute application iter app environ start response file opt code localstack venv lib site packages moto server py line in call return backend app environ start response file opt code localstack venv lib site packages flask app py line in call return self wsgi app environ start response file opt code localstack venv lib site packages flask app py line in wsgi app response self handle exception e file opt code localstack venv lib site packages flask cors extension py line in wrapped function return cors after request app make response f args kwargs file opt code localstack venv lib site packages flask app py line in handle exception reraise exc type exc value tb file opt code localstack venv lib site packages flask compat py line in reraise raise value file opt code localstack venv lib site packages flask app py line in wsgi app response self full dispatch request file opt code localstack venv lib site packages flask app py line in full dispatch request rv self handle user exception e file opt code localstack venv lib site packages flask cors extension py line in wrapped function return cors after request app make response f args kwargs file opt code localstack venv lib site packages flask app py line in handle user exception reraise exc type exc value tb file opt code localstack venv lib site packages flask compat py line in reraise raise value file opt code localstack venv lib site packages flask app py line in full dispatch request rv self dispatch request file opt code localstack venv lib site packages flask app py line in dispatch request return self view functions req view args file opt code localstack venv lib site packages moto core utils py line in call result self callback request request url file opt code localstack venv lib site packages moto core responses py line in dispatch return cls dispatch args kwargs file opt code localstack venv lib site packages moto core responses py line in dispatch return self call action file opt code localstack venv lib site packages moto core responses py line in call action response method file opt code localstack venv lib site packages moto core utils py line in wrapper response f args kwargs file opt code localstack venv lib site packages moto cloudformation responses py line in create change set change set type update or create file opt code localstack venv lib site packages moto cloudformation models py line in create change set cross stack resources self exports file opt code localstack venv lib site packages moto cloudformation models py line in init create change set true file opt code localstack venv lib site packages moto cloudformation models py line in init self resource map self create resource map file opt code localstack venv lib site packages moto cloudformation models py line in create resource map resource map create file opt code localstack venv lib site packages moto cloudformation parsing py line in create if isinstance self models file opt code localstack venv lib site packages moto cloudformation parsing py line in getitem resource logical id resource json self self region name file opt code localstack localstack services cloudformation cloudformation starter py line in parse and create resource return parse and create resource logical id resource json resources map region name file opt code localstack localstack services cloudformation cloudformation starter py line in parse and create resource result deploy func logical id resource wrapped stack name stack name file opt code localstack localstack utils cloudformation template deployer py line in deploy resource result deploy resource via sdk function resource id resources resource type func stack name file opt code localstack localstack utils cloudformation template deployer py line in deploy resource via sdk function raise e file opt code localstack localstack utils cloudformation template deployer py line in deploy resource via sdk function result function params file opt code localstack venv lib site packages botocore client py line in api call return self make api call operation name kwargs file opt code localstack venv lib site packages botocore client py line in make api call api params operation model context request context file opt code localstack venv lib site packages botocore client py line in convert to request dict api params operation model file opt code localstack venv lib site packages botocore validate py line in serialize to request raise paramvalidationerror report report generate report botocore exceptions paramvalidationerror parameter validation failed missing required parameter in input functionname considering the fact that the functionname property is not mandatory from cfn perspective it s even worse since having it or not has significant implications is it possible to create a name on the fly
1
319,617
9,747,282,420
IssuesEvent
2019-06-03 14:05:46
dojot/dojot
https://api.github.com/repos/dojot/dojot
closed
[DataBroker, device-manager, history] Real-time does not work on actuate attributes
Priority:High Status:In Progress Team:Backend Type:Bug
**Steps to reproduce the problem:** _test 1_: 1. create a flow with attribute of actuation 2. activate the flow ![Screenshot at 2019-05-06 08-39-35](https://user-images.githubusercontent.com/37310063/57232692-832cd400-6ff3-11e9-81ca-2f76293450e9.png) 3. view attribute detail ![Screenshot at 2019-05-06 08-40-04](https://user-images.githubusercontent.com/37310063/57232693-83c56a80-6ff3-11e9-9022-e3be07ec1630.png) 4. After F5 ![Screenshot at 2019-05-06 08-40-34](https://user-images.githubusercontent.com/37310063/57232723-9770d100-6ff3-11e9-8a2a-32d35e125a75.png) _test 2_: 1. actuate on the device ![Screenshot at 2019-05-06 11-16-11](https://user-images.githubusercontent.com/37310063/57232002-1c5aeb00-6ff2-11e9-9fea-31fced307b1f.png) 2. view attribute detail ![Screenshot at 2019-05-06 11-16-17](https://user-images.githubusercontent.com/37310063/57232003-1c5aeb00-6ff2-11e9-8e8b-a44f0d4120ad.png) 3. publish data from another attribute ![Screenshot at 2019-05-06 11-17-26](https://user-images.githubusercontent.com/37310063/57232187-7956a100-6ff2-11e9-9e40-63c7b3710a7c.png) ![Screenshot at 2019-05-06 11-17-31](https://user-images.githubusercontent.com/37310063/57232189-7956a100-6ff2-11e9-9f5d-0fc58c75cb33.png) 4. After F5 ![Screenshot at 2019-05-06 11-17-55](https://user-images.githubusercontent.com/37310063/57232138-5d52ff80-6ff2-11e9-92dc-6970945f7f87.png) **Affected Version**: 61.1-20190423
1.0
[DataBroker, device-manager, history] Real-time does not work on actuate attributes - **Steps to reproduce the problem:** _test 1_: 1. create a flow with attribute of actuation 2. activate the flow ![Screenshot at 2019-05-06 08-39-35](https://user-images.githubusercontent.com/37310063/57232692-832cd400-6ff3-11e9-81ca-2f76293450e9.png) 3. view attribute detail ![Screenshot at 2019-05-06 08-40-04](https://user-images.githubusercontent.com/37310063/57232693-83c56a80-6ff3-11e9-9022-e3be07ec1630.png) 4. After F5 ![Screenshot at 2019-05-06 08-40-34](https://user-images.githubusercontent.com/37310063/57232723-9770d100-6ff3-11e9-8a2a-32d35e125a75.png) _test 2_: 1. actuate on the device ![Screenshot at 2019-05-06 11-16-11](https://user-images.githubusercontent.com/37310063/57232002-1c5aeb00-6ff2-11e9-9fea-31fced307b1f.png) 2. view attribute detail ![Screenshot at 2019-05-06 11-16-17](https://user-images.githubusercontent.com/37310063/57232003-1c5aeb00-6ff2-11e9-8e8b-a44f0d4120ad.png) 3. publish data from another attribute ![Screenshot at 2019-05-06 11-17-26](https://user-images.githubusercontent.com/37310063/57232187-7956a100-6ff2-11e9-9e40-63c7b3710a7c.png) ![Screenshot at 2019-05-06 11-17-31](https://user-images.githubusercontent.com/37310063/57232189-7956a100-6ff2-11e9-9f5d-0fc58c75cb33.png) 4. After F5 ![Screenshot at 2019-05-06 11-17-55](https://user-images.githubusercontent.com/37310063/57232138-5d52ff80-6ff2-11e9-92dc-6970945f7f87.png) **Affected Version**: 61.1-20190423
priority
real time does not work on actuate attributes steps to reproduce the problem test create a flow with attribute of actuation activate the flow view attribute detail after test actuate on the device view attribute detail publish data from another attribute after affected version
1
530,665
15,435,525,670
IssuesEvent
2021-03-07 09:15:45
VelvetThePanda/Silk
https://api.github.com/repos/VelvetThePanda/Silk
closed
Daily command for newcomers doesn't save balance correctly
Bugged Priority: HIGH
**Describe the bug** If you're new, and you run the daily command, it will say you've collected $500, but in reality it saves 0 dollars. However, if you have an account, it saves correctly. **To Reproduce** - Be new - Run daily command - Run cash command - Observe **Expected behavior** An account is created, and $500 is deposited. **Actual behavior** An account is created, but $0 is saved.
1.0
Daily command for newcomers doesn't save balance correctly - **Describe the bug** If you're new, and you run the daily command, it will say you've collected $500, but in reality it saves 0 dollars. However, if you have an account, it saves correctly. **To Reproduce** - Be new - Run daily command - Run cash command - Observe **Expected behavior** An account is created, and $500 is deposited. **Actual behavior** An account is created, but $0 is saved.
priority
daily command for newcomers doesn t save balance correctly describe the bug if you re new and you run the daily command it will say you ve collected but in reality it saves dollars however if you have an account it saves correctly to reproduce be new run daily command run cash command observe expected behavior an account is created and is deposited actual behavior an account is created but is saved
1
580,146
17,210,729,901
IssuesEvent
2021-07-19 03:42:53
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
reopened
Allow negative learning rates
enhancement high priority module: optimizer triage review triaged
## 🚀 Feature Currently, optimizers throw an assertion error when negative learning rates are supplied at construction. This proposal suggests removing this restriction. ## Motivation Since maximization is equivalent to minimizing a negative loss function, negative learning rates are useful and make sense. For example, in GANs, the generator and discriminator can be trained adversarially by giving the discriminator a negative learning rate. This avoids having two backward passes, which improves both computational efficiency and conceptual clarity. ## Pitch As learning rates are typically parameterized by constants, providing a negative rate by accident is highly unlikely. I argue that making this mistake is much less likely than wanting a negative learning rate, and so the defensive assertion is better removed. ## Alternatives It is currently possible to set a negative LR through an ugly loop through the optimizer's `param_group`. An alternative would be to provide a cleaner way to do so, i.e., a `maximize=True` flag on optimizer construction. cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @vincentqb @iramazanli
1.0
Allow negative learning rates - ## 🚀 Feature Currently, optimizers throw an assertion error when negative learning rates are supplied at construction. This proposal suggests removing this restriction. ## Motivation Since maximization is equivalent to minimizing a negative loss function, negative learning rates are useful and make sense. For example, in GANs, the generator and discriminator can be trained adversarially by giving the discriminator a negative learning rate. This avoids having two backward passes, which improves both computational efficiency and conceptual clarity. ## Pitch As learning rates are typically parameterized by constants, providing a negative rate by accident is highly unlikely. I argue that making this mistake is much less likely than wanting a negative learning rate, and so the defensive assertion is better removed. ## Alternatives It is currently possible to set a negative LR through an ugly loop through the optimizer's `param_group`. An alternative would be to provide a cleaner way to do so, i.e., a `maximize=True` flag on optimizer construction. cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @vincentqb @iramazanli
priority
allow negative learning rates 🚀 feature currently optimizers throw an assertion error when negative learning rates are supplied at construction this proposal suggests removing this restriction motivation since maximization is equivalent to minimizing a negative loss function negative learning rates are useful and make sense for example in gans the generator and discriminator can be trained adversarially by giving the discriminator a negative learning rate this avoids having two backward passes which improves both computational efficiency and conceptual clarity pitch as learning rates are typically parameterized by constants providing a negative rate by accident is highly unlikely i argue that making this mistake is much less likely than wanting a negative learning rate and so the defensive assertion is better removed alternatives it is currently possible to set a negative lr through an ugly loop through the optimizer s param group an alternative would be to provide a cleaner way to do so i e a maximize true flag on optimizer construction cc ezyang gchanan bdhirsh jbschlosser vincentqb iramazanli
1
760,111
26,629,051,264
IssuesEvent
2023-01-24 16:28:26
valantic/vue-template
https://api.github.com/repos/valantic/vue-template
closed
Update api documentation
enhancement high priority vue-3
All attributes on the api helper are documented as required. But most of them are optional (e.g. `[notificationOptions]`). Update documentation and wrap optional attributes in square brackets.
1.0
Update api documentation - All attributes on the api helper are documented as required. But most of them are optional (e.g. `[notificationOptions]`). Update documentation and wrap optional attributes in square brackets.
priority
update api documentation all attributes on the api helper are documented as required but most of them are optional e g update documentation and wrap optional attributes in square brackets
1
423,538
12,298,203,675
IssuesEvent
2020-05-11 10:06:46
MyDataTaiwan/mylog14
https://api.github.com/repos/MyDataTaiwan/mylog14
reopened
[Todo] Style adjustment for daily-overview page
enhancement priority-high
# Issue description 1. Center the Main Header title 2. Fix that the lower part of the countdown animation overlap with the dividing line
1.0
[Todo] Style adjustment for daily-overview page - # Issue description 1. Center the Main Header title 2. Fix that the lower part of the countdown animation overlap with the dividing line
priority
style adjustment for daily overview page issue description center the main header title fix that the lower part of the countdown animation overlap with the dividing line
1
178,636
6,613,080,570
IssuesEvent
2017-09-20 07:51:23
OpenWebslides/OpenWebslides
https://api.github.com/repos/OpenWebslides/OpenWebslides
opened
Docker race condition
bug high priority operations
When restarting Docker/rebooting server, the NGINX container cannot reach the Docker-defined `app` host.
1.0
Docker race condition - When restarting Docker/rebooting server, the NGINX container cannot reach the Docker-defined `app` host.
priority
docker race condition when restarting docker rebooting server the nginx container cannot reach the docker defined app host
1
556,187
16,477,192,564
IssuesEvent
2021-05-24 07:14:51
MathiasReker/Delfinen
https://api.github.com/repos/MathiasReker/Delfinen
closed
View Member info
feature request high priority required
Feature Request Display Member infomation nicely in terminal. - [x] The user must choose which member he wants to view a. The User searches on ID or name * If the user searches on ID, and it exists, the user is directed to the "member view" * If there is more than one match, the matches are displayed, and the user must choose by picking a corresponding number - [x] The system displays member information (Name, Age, etc.) - [x] The system displays different options * Update Member info * Anonymize the member
1.0
View Member info - Feature Request Display Member infomation nicely in terminal. - [x] The user must choose which member he wants to view a. The User searches on ID or name * If the user searches on ID, and it exists, the user is directed to the "member view" * If there is more than one match, the matches are displayed, and the user must choose by picking a corresponding number - [x] The system displays member information (Name, Age, etc.) - [x] The system displays different options * Update Member info * Anonymize the member
priority
view member info feature request display member infomation nicely in terminal the user must choose which member he wants to view a the user searches on id or name if the user searches on id and it exists the user is directed to the member view if there is more than one match the matches are displayed and the user must choose by picking a corresponding number the system displays member information name age etc the system displays different options update member info anonymize the member
1
291,505
8,926,061,704
IssuesEvent
2019-01-22 02:16:03
SalvatoreTosti/spicy-bingo
https://api.github.com/repos/SalvatoreTosti/spicy-bingo
opened
Update validation on user-inputted name fields
high priority
User inputted fields should be checked to ensure they conform to the back-end specifications of the information. Ex. space checking on new room names.
1.0
Update validation on user-inputted name fields - User inputted fields should be checked to ensure they conform to the back-end specifications of the information. Ex. space checking on new room names.
priority
update validation on user inputted name fields user inputted fields should be checked to ensure they conform to the back end specifications of the information ex space checking on new room names
1
348,771
10,452,589,496
IssuesEvent
2019-09-19 14:57:06
ansible/galaxy
https://api.github.com/repos/ansible/galaxy
opened
Security Updates
area/backend priority/high status/new type/bug
[ ] Update packages/dependencies to fix failing security audit [ ] Address the concerns identified by @chrismeyersfsu
1.0
Security Updates - [ ] Update packages/dependencies to fix failing security audit [ ] Address the concerns identified by @chrismeyersfsu
priority
security updates update packages dependencies to fix failing security audit address the concerns identified by chrismeyersfsu
1
458,086
13,168,399,022
IssuesEvent
2020-08-11 12:03:46
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
closed
custom project role permission denied
area/iam kind/bug kind/need-to-verify priority/high
**Describe the Bug** A custom project role with `Application Workloads View / Application Workloads Management` configured HPA but denied ![Screen Shot 2020-08-11 at 4 04 47 PM](https://user-images.githubusercontent.com/28859385/89873156-95c98c00-dbec-11ea-8f52-d63746f300f1.png) ![Screen Shot 2020-08-11 at 4 06 26 PM](https://user-images.githubusercontent.com/28859385/89873175-9eba5d80-dbec-11ea-99ec-1c6840713bb9.png) **Versions Used** KubeSphere: 3.0.0 (2020-08-11)
1.0
custom project role permission denied - **Describe the Bug** A custom project role with `Application Workloads View / Application Workloads Management` configured HPA but denied ![Screen Shot 2020-08-11 at 4 04 47 PM](https://user-images.githubusercontent.com/28859385/89873156-95c98c00-dbec-11ea-8f52-d63746f300f1.png) ![Screen Shot 2020-08-11 at 4 06 26 PM](https://user-images.githubusercontent.com/28859385/89873175-9eba5d80-dbec-11ea-99ec-1c6840713bb9.png) **Versions Used** KubeSphere: 3.0.0 (2020-08-11)
priority
custom project role permission denied describe the bug a custom project role with application workloads view application workloads management configured hpa but denied versions used kubesphere
1
425,412
12,339,932,104
IssuesEvent
2020-05-14 19:00:13
canonical-web-and-design/jaas-dashboard
https://api.github.com/repos/canonical-web-and-design/jaas-dashboard
closed
Empty table warning is partly in header
Bug 🐛 Models Listing Priority: High
![Screen Shot 2020-05-13 at 2 04 17 PM](https://user-images.githubusercontent.com/532033/81859748-f2040d80-9522-11ea-9a99-83e0cc99a7e5.png) To reproduce, run the dashboard with no models.
1.0
Empty table warning is partly in header - ![Screen Shot 2020-05-13 at 2 04 17 PM](https://user-images.githubusercontent.com/532033/81859748-f2040d80-9522-11ea-9a99-83e0cc99a7e5.png) To reproduce, run the dashboard with no models.
priority
empty table warning is partly in header to reproduce run the dashboard with no models
1
631,396
20,151,422,599
IssuesEvent
2022-02-09 12:47:28
Blosc/caterva
https://api.github.com/repos/Blosc/caterva
closed
Implement a resize functionality
enhancement high priority
This would allow to extend/shrink an array in different dimensions. I suggest a new function with a signature similar to this: ``` /** * @brief Resize a caterva array * * Changes the shape of the caterva array by growing or shrinking one or more dimensions. * * @param ctx The caterva context to be used. * @param array The caterva array. * @param new_dims New dimensions of the array. * * @return An error code */ int caterva_resize(caterva_ctx_t *ctx, caterva_array_t *array, int *new_dims); ```
1.0
Implement a resize functionality - This would allow to extend/shrink an array in different dimensions. I suggest a new function with a signature similar to this: ``` /** * @brief Resize a caterva array * * Changes the shape of the caterva array by growing or shrinking one or more dimensions. * * @param ctx The caterva context to be used. * @param array The caterva array. * @param new_dims New dimensions of the array. * * @return An error code */ int caterva_resize(caterva_ctx_t *ctx, caterva_array_t *array, int *new_dims); ```
priority
implement a resize functionality this would allow to extend shrink an array in different dimensions i suggest a new function with a signature similar to this brief resize a caterva array changes the shape of the caterva array by growing or shrinking one or more dimensions param ctx the caterva context to be used param array the caterva array param new dims new dimensions of the array return an error code int caterva resize caterva ctx t ctx caterva array t array int new dims
1
659,192
21,919,302,097
IssuesEvent
2022-05-22 10:20:07
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Snapshot releases (Mac and Linux versions) on MacOS are not properly loading
bug priority: High packaging
When opening the Mac OS (.dmg) snapshot release, it doesn't load OpenRefine any further than the logo, as seen in this screenshot: ![image](https://user-images.githubusercontent.com/9881645/162605130-f818968a-eed4-44df-9499-273de95bfdf2.png) When I try with the Linux download, that one doesn't even open a new browser tab at all. ### Expected Behavior OpenRefine should fully load. ### Context Terminal log when trying to start the MacOS .dmg (pretty minimal): ``` Sandras-MacBook-Air:~ fokky$ /Applications/OpenRefine\ snapshot\ 20220408.app/Contents/MacOS/JavaAppLauncher ; exit; 08:15:29.649 [ refine_server] Starting Server bound to '127.0.0.1:3333' (0ms) 08:15:29.717 [ refine_server] Initializing context: '/' from '/Applications/OpenRefine snapshot 20220408.app/Contents/Resources/webapp' (68ms) 08:15:31.931 [ refine] Starting OpenRefine 3.6-SNAPSHOT [0ba3b92]... (2214ms) 08:15:31.937 [ refine] initializing FileProjectManager with dir (6ms) 08:15:31.937 [ refine] /Users/fokky/Library/Application Support/OpenRefine (0ms) ``` When I try to open the Linux version of the snapshot release, that one doesn't even start up a new browser tab. Terminal log for that one is: ``` miniMac:~ fokky$ /Users/fokky/Desktop/openrefine-3.6-SNAPSHOT\ 3/refine ; exit; Using refine.ini for configuration cat: refine.ini: No such file or directory ------------------------------------------------------------------------------------------------- You have 812M of free memory. Your current configuration is set to use 1024M of memory. OpenRefine can run better when given more memory. Read our FAQ on how to allocate more memory here: https://docs.openrefine.org/manual/installing#increasing-memory-allocation ------------------------------------------------------------------------------------------------- ls: server/target/lib: No such file or directory \nCould not find Maven locally, starting download for Maven ... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 8469k 100 8469k 0 0 10.7M 0 --:--:-- --:--:-- --:--:-- 10.7M /Users/fokky/Desktop/openrefine-3.6-SNAPSHOT 3/refine: line 294: cd: main/webapp: No such file or directory [INFO] Scanning for projects... [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 0.300 s [INFO] Finished at: 2022-04-10T09:26:43+02:00 [INFO] ------------------------------------------------------------------------ [ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException [INFO] Scanning for projects... [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 0.426 s [INFO] Finished at: 2022-04-10T09:26:47+02:00 [INFO] ------------------------------------------------------------------------ [ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException Error: Could not find or load main class com.google.refine.Refine Caused by: java.lang.ClassNotFoundException: com.google.refine.Refine logout Saving session... ...copying shared history... ...saving history...truncating history files... ...completed. [Process completed] ``` ### Versions<!-- (please complete the following information)--> - Operating System: MacOS 10.13.6, MacOS 12.2.1 - Browser Version: Chrome - JRE or JDK Version: n/a? - OpenRefine: 3.6 snapshot release ### Files used I just tried with the April 8 and April 10 snapshot releases, but the problem has been around for longer (several weeks?). I heard from a Linux user (via the Structured Data on Commons Telegram group) that they have the same issue of OpenRefine snapshot releases not loading for them.
1.0
Snapshot releases (Mac and Linux versions) on MacOS are not properly loading - When opening the Mac OS (.dmg) snapshot release, it doesn't load OpenRefine any further than the logo, as seen in this screenshot: ![image](https://user-images.githubusercontent.com/9881645/162605130-f818968a-eed4-44df-9499-273de95bfdf2.png) When I try with the Linux download, that one doesn't even open a new browser tab at all. ### Expected Behavior OpenRefine should fully load. ### Context Terminal log when trying to start the MacOS .dmg (pretty minimal): ``` Sandras-MacBook-Air:~ fokky$ /Applications/OpenRefine\ snapshot\ 20220408.app/Contents/MacOS/JavaAppLauncher ; exit; 08:15:29.649 [ refine_server] Starting Server bound to '127.0.0.1:3333' (0ms) 08:15:29.717 [ refine_server] Initializing context: '/' from '/Applications/OpenRefine snapshot 20220408.app/Contents/Resources/webapp' (68ms) 08:15:31.931 [ refine] Starting OpenRefine 3.6-SNAPSHOT [0ba3b92]... (2214ms) 08:15:31.937 [ refine] initializing FileProjectManager with dir (6ms) 08:15:31.937 [ refine] /Users/fokky/Library/Application Support/OpenRefine (0ms) ``` When I try to open the Linux version of the snapshot release, that one doesn't even start up a new browser tab. Terminal log for that one is: ``` miniMac:~ fokky$ /Users/fokky/Desktop/openrefine-3.6-SNAPSHOT\ 3/refine ; exit; Using refine.ini for configuration cat: refine.ini: No such file or directory ------------------------------------------------------------------------------------------------- You have 812M of free memory. Your current configuration is set to use 1024M of memory. OpenRefine can run better when given more memory. Read our FAQ on how to allocate more memory here: https://docs.openrefine.org/manual/installing#increasing-memory-allocation ------------------------------------------------------------------------------------------------- ls: server/target/lib: No such file or directory \nCould not find Maven locally, starting download for Maven ... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 8469k 100 8469k 0 0 10.7M 0 --:--:-- --:--:-- --:--:-- 10.7M /Users/fokky/Desktop/openrefine-3.6-SNAPSHOT 3/refine: line 294: cd: main/webapp: No such file or directory [INFO] Scanning for projects... [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 0.300 s [INFO] Finished at: 2022-04-10T09:26:43+02:00 [INFO] ------------------------------------------------------------------------ [ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException [INFO] Scanning for projects... [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 0.426 s [INFO] Finished at: 2022-04-10T09:26:47+02:00 [INFO] ------------------------------------------------------------------------ [ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException Error: Could not find or load main class com.google.refine.Refine Caused by: java.lang.ClassNotFoundException: com.google.refine.Refine logout Saving session... ...copying shared history... ...saving history...truncating history files... ...completed. [Process completed] ``` ### Versions<!-- (please complete the following information)--> - Operating System: MacOS 10.13.6, MacOS 12.2.1 - Browser Version: Chrome - JRE or JDK Version: n/a? - OpenRefine: 3.6 snapshot release ### Files used I just tried with the April 8 and April 10 snapshot releases, but the problem has been around for longer (several weeks?). I heard from a Linux user (via the Structured Data on Commons Telegram group) that they have the same issue of OpenRefine snapshot releases not loading for them.
priority
snapshot releases mac and linux versions on macos are not properly loading when opening the mac os dmg snapshot release it doesn t load openrefine any further than the logo as seen in this screenshot when i try with the linux download that one doesn t even open a new browser tab at all expected behavior openrefine should fully load context terminal log when trying to start the macos dmg pretty minimal sandras macbook air fokky applications openrefine snapshot app contents macos javaapplauncher exit starting server bound to initializing context from applications openrefine snapshot app contents resources webapp starting openrefine snapshot initializing fileprojectmanager with dir users fokky library application support openrefine when i try to open the linux version of the snapshot release that one doesn t even start up a new browser tab terminal log for that one is minimac fokky users fokky desktop openrefine snapshot refine exit using refine ini for configuration cat refine ini no such file or directory you have of free memory your current configuration is set to use of memory openrefine can run better when given more memory read our faq on how to allocate more memory here ls server target lib no such file or directory ncould not find maven locally starting download for maven total received xferd average speed time time time current dload upload total spent left speed users fokky desktop openrefine snapshot refine line cd main webapp no such file or directory scanning for projects build failure total time s finished at the goal you specified requires a project to execute but there is no pom in this directory users fokky desktop please verify you invoked maven from the correct directory to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles scanning for projects build failure total time s finished at the goal you specified requires a project to execute but there is no pom in this directory users fokky desktop please verify you invoked maven from the correct directory to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles error could not find or load main class com google refine refine caused by java lang classnotfoundexception com google refine refine logout saving session copying shared history saving history truncating history files completed versions operating system macos macos browser version chrome jre or jdk version n a openrefine snapshot release files used i just tried with the april and april snapshot releases but the problem has been around for longer several weeks i heard from a linux user via the structured data on commons telegram group that they have the same issue of openrefine snapshot releases not loading for them
1
510,546
14,792,652,967
IssuesEvent
2021-01-12 15:00:06
staxrip/staxrip
https://api.github.com/repos/staxrip/staxrip
closed
Staxrip will try to append CHUNKS even if chunks = 1
added/fixed/done bug priority high
**BUG:** When there are files called file_chunk2.hevc and file_chunk3.hevc (...) in the temp folder, Staxrip will attempt to append them when muxing, even if chunk encoding is not activated (i.e chunks =1). These files could happen to be there due to a previous encoding. **Expected behaviour** If chunks = 1 then those files should be ignored, and muxing should not append them. Muxing should work SMARTLY. If chunks = 1 , look for file_out.hevc, ignore any other If chunks = 2 look for file_out.hevc and file_chunk2.hevc, ignore any other If chunks = 3 look for file_out.hevc and file_chunk2.hevc and file_chunk3.hevc, ignore any other Here is the complete log for your pleasure: it shows that NO second chunk has been encoded, but muxer has appended a chunk from a previous processing. ``` ------------------------- System Environment ------------------------- StaxRip : 2.1.7.1 Windows : Windows 10 Home 2004 Language : English (United States) CPU : Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz GPU : Intel(R) HD Graphics 530, NVIDIA GeForce GTX 960M Resolution : 1920 x 1080 DPI : 96 ----------------------- Media Info Source File ----------------------- D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4 General Complete name : D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4 Format : MPEG-4 Format profile : Base Media / Version 2 Codec ID : mp42 (isom/mp42) File size : 4.88 MiB Duration : 1 min 18 s Overall bit rate mode : Variable Overall bit rate : 522 kb/s Encoded date : UTC 2017-12-17 15:18:14 Tagged date : UTC 2017-12-17 15:18:14 gsst : 0 gstd : 78460 Video ID : 1 Format : AVC Format/Info : Advanced Video Codec Format profile : High@L3.1 Format settings : CABAC / 1 Ref Frames Format, CABAC : Yes Format, Reference frames : 1 frame Format, GOP : M=1, N=60 Codec ID : avc1 Codec ID/Info : Advanced Video Coding Duration : 1 min 18 s Bit rate : 327 kb/s Maximum bit rate : 471 kb/s Width : 1 280 pixels Height : 720 pixels Display aspect ratio : 16:9 Frame rate mode : Constant Frame rate : 25.000 FPS Color space : YUV Chroma subsampling : 4:2:0 Bit depth : 8 bits Scan type : Progressive Bits/(Pixel*Frame) : 0.014 Stream size : 3.06 MiB (63%) Tagged date : UTC 2017-12-17 15:18:15 Color range : Limited Color primaries : BT.709 Transfer characteristics : BT.709 Matrix coefficients : BT.709 Codec configuration box : avcC Audio ID : 2 Format : AAC LC Format/Info : Advanced Audio Codec Low Complexity Codec ID : mp4a-40-2 Duration : 1 min 18 s Bit rate mode : Variable Bit rate : 192 kb/s Maximum bit rate : 201 kb/s Channel(s) : 2 channels Channel layout : L R Sampling rate : 44.1 kHz Frame rate : 43.066 FPS (1024 SPF) Compression mode : Lossy Stream size : 1.79 MiB (37%) Title : IsoMedia File Produced by Google, 5-11-2011 Encoded date : UTC 2017-12-17 15:18:15 Tagged date : UTC 2017-12-17 15:18:15 ----------------------------- Demux audio ----------------------------- MP4Box 1.1.0-DEV-rev390-g4228658a9-x64-gcc10.2.0 Patman86 "C:\Program Files\StaxRip\Apps\Support\MP4Box\MP4Box.exe" -single 2 -out "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4" Start: 11:12:51 End: 11:12:51 Duration: 00:00:00 General Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a Format : MPEG-4 Format profile : Base Media Codec ID : isom (isom) File size : 1.81 MiB Duration : 1 min 18 s Overall bit rate mode : Variable Overall bit rate : 194 kb/s Encoded date : UTC 2021-01-11 09:12:51 Tagged date : UTC 2021-01-11 09:12:51 Audio ID : 2 Format : AAC LC Format/Info : Advanced Audio Codec Low Complexity Format profile : AAC@L2 Codec ID : mp4a-40-2 Duration : 1 min 18 s Bit rate mode : Variable Bit rate : 192 kb/s Nominal bit rate : 728 b/s Maximum bit rate : 201 kb/s Channel(s) : 2 channels Channel layout : L R Sampling rate : 44.1 kHz Frame rate : 43.066 FPS (1024 SPF) Compression mode : Lossy Stream size : 1.79 MiB (99%) Encoded date : UTC 2021-01-11 09:12:51 Tagged date : UTC 2021-01-11 09:12:51 ---------------------- Indexing using ffmsindex ---------------------- "C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffmsindex.exe" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex" Writing index... done. Start: 11:12:52 End: 11:12:52 Duration: 00:00:00 ---------------------------- Configuration ---------------------------- Template : ZK x265 480p BT709 Base Video Encoder Profile : x265 Container/Muxer Profile : MKV (mkvmerge) --------------------------- AviSynth Script --------------------------- AddAutoloadDir("C:\Program Files\StaxRip\Apps\FrameServer\AviSynth\plugins") AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\AviSynth") AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\Dual") LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffms2.dll") LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\AVS\JPSDR\Plugins_JPSDR.dll") FFVideoSource("D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4", cachefile="C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex") #AssumeFPS(25) LanczosResizeMT(856, 480, prefetch=4) Trim(100, 145) + Trim(1375, 1420) ------------------------- Source Script Info ------------------------- Width : 1280 Height : 720 Frames : 1960 Time : 01:18.400 Framerate : 25 (25/1) Format : YUV420P8 ------------------------- Target Script Info ------------------------- Width : 856 Height : 480 Frames : 92 Time : 00:03.680 Framerate : 25 (25/1) Format : YUV420P8 --------------------------- Video encoding --------------------------- x265 M-3.4+35-772bb4c84-x64-gcc10.2.0 Patman86 "C:\Program Files\StaxRip\Apps\Encoders\x265\x265.exe" --crf 24 --preset slow --output-depth 10 --output "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps.avs" avs+ [info]: AviSynth+ 3.6.2 (r3341, master, x86_64) avs+ [info]: Video colorspace: YUV420 (YV12) avs+ [info]: Video depth: 8 avs+ [info]: Video resolution: 856x480 avs+ [info]: Video framerate: 25/1 avs+ [info]: Video framecount: 92 avs+ [info]: 856x480 fps 25/1 i420p8 frames 0 - 91 of 92 raw [info]: output file: C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc x265 [info]: HEVC encoder version x265M 3.4+35-772bb4c84 x265 [info]: build info [Windows][MSVC 1928][64 bit] 10bit x265 [info]: using cpu capabilities: MMX2 SSE2Fast LZCNT SSSE3 SSE4.2 AVX FMA3 BMI2 AVX2 x265 [info]: Main 10 profile, Level-3 (Main tier) x265 [info]: Thread pool created using 8 threads x265 [info]: Slices : 1 x265 [info]: frame threads / pool features : 3 / wpp(8 rows) x265 [warning]: Source height < 720p; disabling lookahead-slices x265 [info]: Coding QT: max CU size, min CU size : 64 / 8 x265 [info]: Residual QT: max TU size, max depth : 32 / 1 inter / 1 intra x265 [info]: ME / range / subpel / merge : star / 57 / 3 / 3 x265 [info]: Keyframe min / max / scenecut / bias : 25 / 250 / 40 / 5.00 x265 [info]: Lookahead / bframes / badapt : 25 / 4 / 2 x265 [info]: b-pyramid / weightp / weightb : 1 / 1 / 0 x265 [info]: References / ref-limit cu / depth : 4 / on / on x265 [info]: AQ: mode / str / qg-size / cu-tree : 2 / 1.0 / 32 / 1 x265 [info]: Rate Control / qCompress : CRF-24.0 / 0.60 x265 [info]: tools: rect limit-modes rd=4 psy-rd=2.00 rdoq=2 psy-rdoq=1.00 x265 [info]: tools: rskip mode=1 signhide tmvp strong-intra-smoothing deblock x265 [info]: tools: sao x265 [info]: frame I: 1, Avg QP:22.44 kb/s: 1123.60 x265 [info]: frame P: 20, Avg QP:27.63 kb/s: 119.84 x265 [info]: frame B: 71, Avg QP:31.24 kb/s: 58.88 x265 [info]: Weighted P-Frames: Y:0.0% UV:0.0% x265 [info]: consecutive B-frames: 9.5% 4.8% 4.8% 0.0% 81.0% encoded 92 frames in 1.98s (46.46 fps), 83.71 kb/s, Avg QP:30.36 Start: 23:13:42 End: 23:13:45 Duration: 00:00:02 General Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc Format : HEVC Format/Info : High Efficiency Video Coding File size : 40.3 KiB Duration : 640 ms Overall bit rate : 516 kb/s Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit Video Format : HEVC Format/Info : High Efficiency Video Coding Format profile : Main 10@L3@Main Duration : 640 ms Width : 856 pixels Height : 480 pixels Display aspect ratio : 16:9 Frame rate : 25.000 FPS Color space : YUV Chroma subsampling : 4:2:0 Bit depth : 10 bits Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit ---------------------------- Muxing to MKV ---------------------------- mkvmerge 52 "C:\Program Files\StaxRip\Apps\Support\MKVToolNix\mkvmerge.exe" -o "R:\Timecode sample - 25fps.mkv" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" + "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc" --ui-language en --title "" mkvmerge v52.0.0 ('Secret For The Mad') 64-bit 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc': Using the demultiplexer for the format 'HEVC/H.265'. 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc': Using the demultiplexer for the format 'HEVC/H.265'. 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'. 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'. No append mapping was given for the file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc'). A default mapping of 1:0:0:0 will be used instead. Please keep that in mind if mkvmerge aborts with an error message regarding invalid '--append-to' options. The file 'R:\Timecode sample - 25fps.mkv' has been opened for writing. Appending track 0 from file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc') to track 0 from file no. 0 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc'). The cue entries (the index) are being written... Multiplexing took 0 seconds. Start: 23:13:45 End: 23:13:45 Duration: 00:00:00 General Complete name : R:\Timecode sample - 25fps.mkv Format : Matroska Format version : Version 4 File size : 73.1 KiB Duration : 5 s 520 ms Overall bit rate : 108 kb/s Encoded date : UTC 2021-01-11 21:13:45 Writing application : mkvmerge v52.0.0 ('Secret For The Mad') 64-bit Writing library : libebml v1.4.1 + libmatroska v1.6.2 Video ID : 1 Format : HEVC Format/Info : High Efficiency Video Coding Format profile : Main 10@L3@Main Codec ID : V_MPEGH/ISO/HEVC Duration : 5 s 520 ms Bit rate : 95.0 kb/s Width : 856 pixels Height : 480 pixels Display aspect ratio : 16:9 Frame rate mode : Constant Frame rate : 25.000 FPS Color space : YUV Chroma subsampling : 4:2:0 Bit depth : 10 bits Bits/(Pixel*Frame) : 0.009 Stream size : 64.0 KiB (88%) Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit Default : Yes Forced : No ---------------------------- Job Complete ---------------------------- Start: 23:13:42 End: 23:13:45 Duration: 00:00:03 ```
1.0
Staxrip will try to append CHUNKS even if chunks = 1 - **BUG:** When there are files called file_chunk2.hevc and file_chunk3.hevc (...) in the temp folder, Staxrip will attempt to append them when muxing, even if chunk encoding is not activated (i.e chunks =1). These files could happen to be there due to a previous encoding. **Expected behaviour** If chunks = 1 then those files should be ignored, and muxing should not append them. Muxing should work SMARTLY. If chunks = 1 , look for file_out.hevc, ignore any other If chunks = 2 look for file_out.hevc and file_chunk2.hevc, ignore any other If chunks = 3 look for file_out.hevc and file_chunk2.hevc and file_chunk3.hevc, ignore any other Here is the complete log for your pleasure: it shows that NO second chunk has been encoded, but muxer has appended a chunk from a previous processing. ``` ------------------------- System Environment ------------------------- StaxRip : 2.1.7.1 Windows : Windows 10 Home 2004 Language : English (United States) CPU : Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz GPU : Intel(R) HD Graphics 530, NVIDIA GeForce GTX 960M Resolution : 1920 x 1080 DPI : 96 ----------------------- Media Info Source File ----------------------- D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4 General Complete name : D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4 Format : MPEG-4 Format profile : Base Media / Version 2 Codec ID : mp42 (isom/mp42) File size : 4.88 MiB Duration : 1 min 18 s Overall bit rate mode : Variable Overall bit rate : 522 kb/s Encoded date : UTC 2017-12-17 15:18:14 Tagged date : UTC 2017-12-17 15:18:14 gsst : 0 gstd : 78460 Video ID : 1 Format : AVC Format/Info : Advanced Video Codec Format profile : High@L3.1 Format settings : CABAC / 1 Ref Frames Format, CABAC : Yes Format, Reference frames : 1 frame Format, GOP : M=1, N=60 Codec ID : avc1 Codec ID/Info : Advanced Video Coding Duration : 1 min 18 s Bit rate : 327 kb/s Maximum bit rate : 471 kb/s Width : 1 280 pixels Height : 720 pixels Display aspect ratio : 16:9 Frame rate mode : Constant Frame rate : 25.000 FPS Color space : YUV Chroma subsampling : 4:2:0 Bit depth : 8 bits Scan type : Progressive Bits/(Pixel*Frame) : 0.014 Stream size : 3.06 MiB (63%) Tagged date : UTC 2017-12-17 15:18:15 Color range : Limited Color primaries : BT.709 Transfer characteristics : BT.709 Matrix coefficients : BT.709 Codec configuration box : avcC Audio ID : 2 Format : AAC LC Format/Info : Advanced Audio Codec Low Complexity Codec ID : mp4a-40-2 Duration : 1 min 18 s Bit rate mode : Variable Bit rate : 192 kb/s Maximum bit rate : 201 kb/s Channel(s) : 2 channels Channel layout : L R Sampling rate : 44.1 kHz Frame rate : 43.066 FPS (1024 SPF) Compression mode : Lossy Stream size : 1.79 MiB (37%) Title : IsoMedia File Produced by Google, 5-11-2011 Encoded date : UTC 2017-12-17 15:18:15 Tagged date : UTC 2017-12-17 15:18:15 ----------------------------- Demux audio ----------------------------- MP4Box 1.1.0-DEV-rev390-g4228658a9-x64-gcc10.2.0 Patman86 "C:\Program Files\StaxRip\Apps\Support\MP4Box\MP4Box.exe" -single 2 -out "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4" Start: 11:12:51 End: 11:12:51 Duration: 00:00:00 General Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a Format : MPEG-4 Format profile : Base Media Codec ID : isom (isom) File size : 1.81 MiB Duration : 1 min 18 s Overall bit rate mode : Variable Overall bit rate : 194 kb/s Encoded date : UTC 2021-01-11 09:12:51 Tagged date : UTC 2021-01-11 09:12:51 Audio ID : 2 Format : AAC LC Format/Info : Advanced Audio Codec Low Complexity Format profile : AAC@L2 Codec ID : mp4a-40-2 Duration : 1 min 18 s Bit rate mode : Variable Bit rate : 192 kb/s Nominal bit rate : 728 b/s Maximum bit rate : 201 kb/s Channel(s) : 2 channels Channel layout : L R Sampling rate : 44.1 kHz Frame rate : 43.066 FPS (1024 SPF) Compression mode : Lossy Stream size : 1.79 MiB (99%) Encoded date : UTC 2021-01-11 09:12:51 Tagged date : UTC 2021-01-11 09:12:51 ---------------------- Indexing using ffmsindex ---------------------- "C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffmsindex.exe" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex" Writing index... done. Start: 11:12:52 End: 11:12:52 Duration: 00:00:00 ---------------------------- Configuration ---------------------------- Template : ZK x265 480p BT709 Base Video Encoder Profile : x265 Container/Muxer Profile : MKV (mkvmerge) --------------------------- AviSynth Script --------------------------- AddAutoloadDir("C:\Program Files\StaxRip\Apps\FrameServer\AviSynth\plugins") AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\AviSynth") AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\Dual") LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffms2.dll") LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\AVS\JPSDR\Plugins_JPSDR.dll") FFVideoSource("D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4", cachefile="C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex") #AssumeFPS(25) LanczosResizeMT(856, 480, prefetch=4) Trim(100, 145) + Trim(1375, 1420) ------------------------- Source Script Info ------------------------- Width : 1280 Height : 720 Frames : 1960 Time : 01:18.400 Framerate : 25 (25/1) Format : YUV420P8 ------------------------- Target Script Info ------------------------- Width : 856 Height : 480 Frames : 92 Time : 00:03.680 Framerate : 25 (25/1) Format : YUV420P8 --------------------------- Video encoding --------------------------- x265 M-3.4+35-772bb4c84-x64-gcc10.2.0 Patman86 "C:\Program Files\StaxRip\Apps\Encoders\x265\x265.exe" --crf 24 --preset slow --output-depth 10 --output "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps.avs" avs+ [info]: AviSynth+ 3.6.2 (r3341, master, x86_64) avs+ [info]: Video colorspace: YUV420 (YV12) avs+ [info]: Video depth: 8 avs+ [info]: Video resolution: 856x480 avs+ [info]: Video framerate: 25/1 avs+ [info]: Video framecount: 92 avs+ [info]: 856x480 fps 25/1 i420p8 frames 0 - 91 of 92 raw [info]: output file: C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc x265 [info]: HEVC encoder version x265M 3.4+35-772bb4c84 x265 [info]: build info [Windows][MSVC 1928][64 bit] 10bit x265 [info]: using cpu capabilities: MMX2 SSE2Fast LZCNT SSSE3 SSE4.2 AVX FMA3 BMI2 AVX2 x265 [info]: Main 10 profile, Level-3 (Main tier) x265 [info]: Thread pool created using 8 threads x265 [info]: Slices : 1 x265 [info]: frame threads / pool features : 3 / wpp(8 rows) x265 [warning]: Source height < 720p; disabling lookahead-slices x265 [info]: Coding QT: max CU size, min CU size : 64 / 8 x265 [info]: Residual QT: max TU size, max depth : 32 / 1 inter / 1 intra x265 [info]: ME / range / subpel / merge : star / 57 / 3 / 3 x265 [info]: Keyframe min / max / scenecut / bias : 25 / 250 / 40 / 5.00 x265 [info]: Lookahead / bframes / badapt : 25 / 4 / 2 x265 [info]: b-pyramid / weightp / weightb : 1 / 1 / 0 x265 [info]: References / ref-limit cu / depth : 4 / on / on x265 [info]: AQ: mode / str / qg-size / cu-tree : 2 / 1.0 / 32 / 1 x265 [info]: Rate Control / qCompress : CRF-24.0 / 0.60 x265 [info]: tools: rect limit-modes rd=4 psy-rd=2.00 rdoq=2 psy-rdoq=1.00 x265 [info]: tools: rskip mode=1 signhide tmvp strong-intra-smoothing deblock x265 [info]: tools: sao x265 [info]: frame I: 1, Avg QP:22.44 kb/s: 1123.60 x265 [info]: frame P: 20, Avg QP:27.63 kb/s: 119.84 x265 [info]: frame B: 71, Avg QP:31.24 kb/s: 58.88 x265 [info]: Weighted P-Frames: Y:0.0% UV:0.0% x265 [info]: consecutive B-frames: 9.5% 4.8% 4.8% 0.0% 81.0% encoded 92 frames in 1.98s (46.46 fps), 83.71 kb/s, Avg QP:30.36 Start: 23:13:42 End: 23:13:45 Duration: 00:00:02 General Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc Format : HEVC Format/Info : High Efficiency Video Coding File size : 40.3 KiB Duration : 640 ms Overall bit rate : 516 kb/s Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit Video Format : HEVC Format/Info : High Efficiency Video Coding Format profile : Main 10@L3@Main Duration : 640 ms Width : 856 pixels Height : 480 pixels Display aspect ratio : 16:9 Frame rate : 25.000 FPS Color space : YUV Chroma subsampling : 4:2:0 Bit depth : 10 bits Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit ---------------------------- Muxing to MKV ---------------------------- mkvmerge 52 "C:\Program Files\StaxRip\Apps\Support\MKVToolNix\mkvmerge.exe" -o "R:\Timecode sample - 25fps.mkv" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" + "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc" --ui-language en --title "" mkvmerge v52.0.0 ('Secret For The Mad') 64-bit 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc': Using the demultiplexer for the format 'HEVC/H.265'. 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc': Using the demultiplexer for the format 'HEVC/H.265'. 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'. 'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'. No append mapping was given for the file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc'). A default mapping of 1:0:0:0 will be used instead. Please keep that in mind if mkvmerge aborts with an error message regarding invalid '--append-to' options. The file 'R:\Timecode sample - 25fps.mkv' has been opened for writing. Appending track 0 from file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc') to track 0 from file no. 0 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc'). The cue entries (the index) are being written... Multiplexing took 0 seconds. Start: 23:13:45 End: 23:13:45 Duration: 00:00:00 General Complete name : R:\Timecode sample - 25fps.mkv Format : Matroska Format version : Version 4 File size : 73.1 KiB Duration : 5 s 520 ms Overall bit rate : 108 kb/s Encoded date : UTC 2021-01-11 21:13:45 Writing application : mkvmerge v52.0.0 ('Secret For The Mad') 64-bit Writing library : libebml v1.4.1 + libmatroska v1.6.2 Video ID : 1 Format : HEVC Format/Info : High Efficiency Video Coding Format profile : Main 10@L3@Main Codec ID : V_MPEGH/ISO/HEVC Duration : 5 s 520 ms Bit rate : 95.0 kb/s Width : 856 pixels Height : 480 pixels Display aspect ratio : 16:9 Frame rate mode : Constant Frame rate : 25.000 FPS Color space : YUV Chroma subsampling : 4:2:0 Bit depth : 10 bits Bits/(Pixel*Frame) : 0.009 Stream size : 64.0 KiB (88%) Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit Default : Yes Forced : No ---------------------------- Job Complete ---------------------------- Start: 23:13:42 End: 23:13:45 Duration: 00:00:03 ```
priority
staxrip will try to append chunks even if chunks bug when there are files called file hevc and file hevc in the temp folder staxrip will attempt to append them when muxing even if chunk encoding is not activated i e chunks these files could happen to be there due to a previous encoding expected behaviour if chunks then those files should be ignored and muxing should not append them muxing should work smartly if chunks look for file out hevc ignore any other if chunks look for file out hevc and file hevc ignore any other if chunks look for file out hevc and file hevc and file hevc ignore any other here is the complete log for your pleasure it shows that no second chunk has been encoded but muxer has appended a chunk from a previous processing system environment staxrip windows windows home language english united states cpu intel r core tm cpu gpu intel r hd graphics nvidia geforce gtx resolution x dpi media info source file d downloads idm downloads timecode sample general complete name d downloads idm downloads timecode sample format mpeg format profile base media version codec id isom file size mib duration min s overall bit rate mode variable overall bit rate kb s encoded date utc tagged date utc gsst gstd video id format avc format info advanced video codec format profile high format settings cabac ref frames format cabac yes format reference frames frame format gop m n codec id codec id info advanced video coding duration min s bit rate kb s maximum bit rate kb s width pixels height pixels display aspect ratio frame rate mode constant frame rate fps color space yuv chroma subsampling bit depth bits scan type progressive bits pixel frame stream size mib tagged date utc color range limited color primaries bt transfer characteristics bt matrix coefficients bt codec configuration box avcc audio id format aac lc format info advanced audio codec low complexity codec id duration min s bit rate mode variable bit rate kb s maximum bit rate kb s channel s channels channel layout l r sampling rate khz frame rate fps spf compression mode lossy stream size mib title isomedia file produced by google encoded date utc tagged date utc demux audio dev c program files staxrip apps support exe single out c temp staxrip timecode sample temp isomedia file produced by google d downloads idm downloads timecode sample start end duration general complete name c temp staxrip timecode sample temp isomedia file produced by google format mpeg format profile base media codec id isom isom file size mib duration min s overall bit rate mode variable overall bit rate kb s encoded date utc tagged date utc audio id format aac lc format info advanced audio codec low complexity format profile aac codec id duration min s bit rate mode variable bit rate kb s nominal bit rate b s maximum bit rate kb s channel s channels channel layout l r sampling rate khz frame rate fps spf compression mode lossy stream size mib encoded date utc tagged date utc indexing using ffmsindex c program files staxrip apps plugins dual ffmsindex exe d downloads idm downloads timecode sample c temp staxrip timecode sample temp temp ffindex writing index done start end duration configuration template zk base video encoder profile container muxer profile mkv mkvmerge avisynth script addautoloaddir c program files staxrip apps frameserver avisynth plugins addautoloaddir c program files staxrip settings plugins avisynth addautoloaddir c program files staxrip settings plugins dual loadplugin c program files staxrip apps plugins dual dll loadplugin c program files staxrip apps plugins avs jpsdr plugins jpsdr dll ffvideosource d downloads idm downloads timecode sample cachefile c temp staxrip timecode sample temp temp ffindex assumefps lanczosresizemt prefetch trim trim source script info width height frames time framerate format target script info width height frames time framerate format video encoding m c program files staxrip apps encoders exe crf preset slow output depth output c temp staxrip timecode sample temp timecode sample out hevc c temp staxrip timecode sample temp timecode sample avs avs avisynth master avs video colorspace avs video depth avs video resolution avs video framerate avs video framecount avs fps frames of raw output file c temp staxrip timecode sample temp timecode sample out hevc hevc encoder version build info using cpu capabilities lzcnt avx main profile level main tier thread pool created using threads slices frame threads pool features wpp rows source height disabling lookahead slices coding qt max cu size min cu size residual qt max tu size max depth inter intra me range subpel merge star keyframe min max scenecut bias lookahead bframes badapt b pyramid weightp weightb references ref limit cu depth on on aq mode str qg size cu tree rate control qcompress crf tools rect limit modes rd psy rd rdoq psy rdoq tools rskip mode signhide tmvp strong intra smoothing deblock tools sao frame i avg qp kb s frame p avg qp kb s frame b avg qp kb s weighted p frames y uv consecutive b frames encoded frames in fps kb s avg qp start end duration general complete name c temp staxrip timecode sample temp timecode sample out hevc format hevc format info high efficiency video coding file size kib duration ms overall bit rate kb s writing library video format hevc format info high efficiency video coding format profile main main duration ms width pixels height pixels display aspect ratio frame rate fps color space yuv chroma subsampling bit depth bits writing library muxing to mkv mkvmerge c program files staxrip apps support mkvtoolnix mkvmerge exe o r timecode sample mkv c temp staxrip timecode sample temp timecode sample out hevc c temp staxrip timecode sample temp timecode sample out hevc ui language en title mkvmerge secret for the mad bit c temp staxrip timecode sample temp timecode sample out hevc using the demultiplexer for the format hevc h c temp staxrip timecode sample temp timecode sample out hevc using the demultiplexer for the format hevc h c temp staxrip timecode sample temp timecode sample out hevc track using the output module for the format hevc h unframed c temp staxrip timecode sample temp timecode sample out hevc track using the output module for the format hevc h unframed no append mapping was given for the file no c temp staxrip timecode sample temp timecode sample out hevc a default mapping of will be used instead please keep that in mind if mkvmerge aborts with an error message regarding invalid append to options the file r timecode sample mkv has been opened for writing appending track from file no c temp staxrip timecode sample temp timecode sample out hevc to track from file no c temp staxrip timecode sample temp timecode sample out hevc the cue entries the index are being written multiplexing took seconds start end duration general complete name r timecode sample mkv format matroska format version version file size kib duration s ms overall bit rate kb s encoded date utc writing application mkvmerge secret for the mad bit writing library libebml libmatroska video id format hevc format info high efficiency video coding format profile main main codec id v mpegh iso hevc duration s ms bit rate kb s width pixels height pixels display aspect ratio frame rate mode constant frame rate fps color space yuv chroma subsampling bit depth bits bits pixel frame stream size kib writing library default yes forced no job complete start end duration
1
208,680
7,157,243,900
IssuesEvent
2018-01-26 19:11:32
unitystation/unitystation
https://api.github.com/repos/unitystation/unitystation
reopened
Clients cannot walk on tiles where explosions took place
High Priority bug
# Preliminary information ## Category BUG ## Bug: Release or Development? 0.2.2 beta 4 steam # Report ## Current Behavior You cannot step on tiles (you lerp back) explosives were located on (prior to their explosion) ## Expected/Wanted/Requested Behavior Want to walk free ## Steps to Reproduce Please enter the steps to reproduce the bug or behaviour: 1. Explode a fuel tank from safe distance 2. Try to step on the tile where the tank was ![2018-01-20 15_48_31](https://user-images.githubusercontent.com/10403536/35183646-9d64e28e-fdfa-11e7-9f9d-2320eca07835.gif) ## OS used: OSX 10.9 ## System specifications MBP mid2010 ## Network beta server 60ms
1.0
Clients cannot walk on tiles where explosions took place - # Preliminary information ## Category BUG ## Bug: Release or Development? 0.2.2 beta 4 steam # Report ## Current Behavior You cannot step on tiles (you lerp back) explosives were located on (prior to their explosion) ## Expected/Wanted/Requested Behavior Want to walk free ## Steps to Reproduce Please enter the steps to reproduce the bug or behaviour: 1. Explode a fuel tank from safe distance 2. Try to step on the tile where the tank was ![2018-01-20 15_48_31](https://user-images.githubusercontent.com/10403536/35183646-9d64e28e-fdfa-11e7-9f9d-2320eca07835.gif) ## OS used: OSX 10.9 ## System specifications MBP mid2010 ## Network beta server 60ms
priority
clients cannot walk on tiles where explosions took place preliminary information category bug bug release or development beta steam report current behavior you cannot step on tiles you lerp back explosives were located on prior to their explosion expected wanted requested behavior want to walk free steps to reproduce please enter the steps to reproduce the bug or behaviour explode a fuel tank from safe distance try to step on the tile where the tank was os used osx system specifications mbp network beta server
1
201,800
7,036,735,001
IssuesEvent
2017-12-28 10:32:11
akvo/akvo-rsr
https://api.github.com/repos/akvo/akvo-rsr
closed
Hide field labels from the qualitative indicator update for IUCN
Priority: High Type: Results framework
IUCN uses qualitative indicators for regular narrative reporting. They would like to avoid potential confusion for their Users by removing field names referring to a qualitative indictor. Also, the 'Actual' field label will be changed to 'Narrative Reporting'. ![iucn labaling slide](https://user-images.githubusercontent.com/26227538/33184248-f013b070-d07b-11e7-82f4-7eb5a2cf63ec.png)
1.0
Hide field labels from the qualitative indicator update for IUCN - IUCN uses qualitative indicators for regular narrative reporting. They would like to avoid potential confusion for their Users by removing field names referring to a qualitative indictor. Also, the 'Actual' field label will be changed to 'Narrative Reporting'. ![iucn labaling slide](https://user-images.githubusercontent.com/26227538/33184248-f013b070-d07b-11e7-82f4-7eb5a2cf63ec.png)
priority
hide field labels from the qualitative indicator update for iucn iucn uses qualitative indicators for regular narrative reporting they would like to avoid potential confusion for their users by removing field names referring to a qualitative indictor also the actual field label will be changed to narrative reporting
1
647,601
21,112,162,355
IssuesEvent
2022-04-05 03:47:54
ethnt/orchard
https://api.github.com/repos/ethnt/orchard
opened
HTPC VM not properly utilizing passthrough GPU
bug machine: htpc high priority
``` # lspci -nnk | grep -iA2 VG 00:10.0 VGA compatible controller [0300]: Intel Corporation TigerLake-LP GT2 [Iris Xe Graphics] [8086:9a49] (rev 01) Subsystem: Intel Corporation Device [8086:3004] Kernel driver in use: i915 ``` ``` # inxi -G Graphics: Message: No device data found. Display: server: No display server data found. Headless machine? tty: 198x39 Message: Advanced graphics data unavailable in console for root. ``` ``` # lshw -C display *-display physical id: 10 bus info: pci@0000:00:10.0 version: 01 width: 64 bits clock: 33MHz capabilities: pciexpress msi pm bus_master cap_list rom configuration: driver=i915 latency=0 resources: iomemory:80-7f iomemory:80-7f irq:40 memory:810000000-810ffffff memory:800000000-80fffffff ioport:f000(size=64) memory:c0000-dffff ```
1.0
HTPC VM not properly utilizing passthrough GPU - ``` # lspci -nnk | grep -iA2 VG 00:10.0 VGA compatible controller [0300]: Intel Corporation TigerLake-LP GT2 [Iris Xe Graphics] [8086:9a49] (rev 01) Subsystem: Intel Corporation Device [8086:3004] Kernel driver in use: i915 ``` ``` # inxi -G Graphics: Message: No device data found. Display: server: No display server data found. Headless machine? tty: 198x39 Message: Advanced graphics data unavailable in console for root. ``` ``` # lshw -C display *-display physical id: 10 bus info: pci@0000:00:10.0 version: 01 width: 64 bits clock: 33MHz capabilities: pciexpress msi pm bus_master cap_list rom configuration: driver=i915 latency=0 resources: iomemory:80-7f iomemory:80-7f irq:40 memory:810000000-810ffffff memory:800000000-80fffffff ioport:f000(size=64) memory:c0000-dffff ```
priority
htpc vm not properly utilizing passthrough gpu lspci nnk grep vg vga compatible controller intel corporation tigerlake lp rev subsystem intel corporation device kernel driver in use inxi g graphics message no device data found display server no display server data found headless machine tty message advanced graphics data unavailable in console for root lshw c display display physical id bus info pci version width bits clock capabilities pciexpress msi pm bus master cap list rom configuration driver latency resources iomemory iomemory irq memory memory ioport size memory dffff
1
111,216
4,466,884,802
IssuesEvent
2016-08-25 01:05:03
AtlasOfLivingAustralia/spatial-portal
https://api.github.com/repos/AtlasOfLivingAustralia/spatial-portal
closed
Add select box/generate selected layer for contextual layers with multi-class legends
enhancement priority-high
Currently, it is difficult to select and identify multiple polygons from one or more contextual layers that have multi-class legends. It seems consistent to offer the parallel to facet class selection and layer generation to contextual layers.
1.0
Add select box/generate selected layer for contextual layers with multi-class legends - Currently, it is difficult to select and identify multiple polygons from one or more contextual layers that have multi-class legends. It seems consistent to offer the parallel to facet class selection and layer generation to contextual layers.
priority
add select box generate selected layer for contextual layers with multi class legends currently it is difficult to select and identify multiple polygons from one or more contextual layers that have multi class legends it seems consistent to offer the parallel to facet class selection and layer generation to contextual layers
1
517,108
14,994,738,290
IssuesEvent
2021-01-29 13:22:04
lumian/grcentral
https://api.github.com/repos/lumian/grcentral
closed
Изменение логики отдачи прошивок
Priority: High Type: Planned improvement
По результатам общения с технической поддержкой Grandstream и нескольких неудачных обновлений было принято решение об изменении логики отдачи прошивок. Суть изменения в следующем: * Устройства всегда должны получать статус HTTP 200 и актуальный файл прошивки. Устройство само решает, будет оно обновляться или нет. * Статус HTTP 404 должен отдаваться только в случае каких-либо ошибок или не правильных настроек сервера.
1.0
Изменение логики отдачи прошивок - По результатам общения с технической поддержкой Grandstream и нескольких неудачных обновлений было принято решение об изменении логики отдачи прошивок. Суть изменения в следующем: * Устройства всегда должны получать статус HTTP 200 и актуальный файл прошивки. Устройство само решает, будет оно обновляться или нет. * Статус HTTP 404 должен отдаваться только в случае каких-либо ошибок или не правильных настроек сервера.
priority
изменение логики отдачи прошивок по результатам общения с технической поддержкой grandstream и нескольких неудачных обновлений было принято решение об изменении логики отдачи прошивок суть изменения в следующем устройства всегда должны получать статус http и актуальный файл прошивки устройство само решает будет оно обновляться или нет статус http должен отдаваться только в случае каких либо ошибок или не правильных настроек сервера
1
242,716
7,845,789,188
IssuesEvent
2018-06-19 13:54:00
nanoframework/Home
https://api.github.com/repos/nanoframework/Home
closed
MDP is not generating a valid PE when project has skeleton project enabled
Area: Visual Studio extension Priority: High Status: FIXED Type: Bug
The generated PE seems to be missing it's CRC32 so the resolving assemblies step (at boot time) is failing with a "missing" referenced assembly.
1.0
MDP is not generating a valid PE when project has skeleton project enabled - The generated PE seems to be missing it's CRC32 so the resolving assemblies step (at boot time) is failing with a "missing" referenced assembly.
priority
mdp is not generating a valid pe when project has skeleton project enabled the generated pe seems to be missing it s so the resolving assemblies step at boot time is failing with a missing referenced assembly
1
219,679
7,344,935,487
IssuesEvent
2018-03-07 16:00:53
MyCryptoHQ/MyCrypto
https://api.github.com/repos/MyCryptoHQ/MyCrypto
opened
Update Translations
blocking stable launch enhancement high priority
### Description of the Issue With the release of the beta, we've added and update a lot of copy throughout the site. To provide translations, all the new 'raw' text needs to be added to the translation files.
1.0
Update Translations - ### Description of the Issue With the release of the beta, we've added and update a lot of copy throughout the site. To provide translations, all the new 'raw' text needs to be added to the translation files.
priority
update translations description of the issue with the release of the beta we ve added and update a lot of copy throughout the site to provide translations all the new raw text needs to be added to the translation files
1
710,803
24,435,280,399
IssuesEvent
2022-10-06 11:01:40
HiAvatar/backend
https://api.github.com/repos/HiAvatar/backend
closed
신규 프로젝트를 생성할 때 아바타에 대한 default 값이 변경됨
Type: Enhancement Priority: High
### Description 기존에는 avatarName, avatarType, bgName에 대해 '미정', '미정', '배경0'으로 default 값으로 지정됐지만, FE 요청에 따라 'avatar1', '미정', '미정'으로 변경한다. <br> ### Todo List - [x] Embeddable 아바타 클래스 수정하기 <br> ### e.t.c (Optional) (참조 및 링크 첨부하기)
1.0
신규 프로젝트를 생성할 때 아바타에 대한 default 값이 변경됨 - ### Description 기존에는 avatarName, avatarType, bgName에 대해 '미정', '미정', '배경0'으로 default 값으로 지정됐지만, FE 요청에 따라 'avatar1', '미정', '미정'으로 변경한다. <br> ### Todo List - [x] Embeddable 아바타 클래스 수정하기 <br> ### e.t.c (Optional) (참조 및 링크 첨부하기)
priority
신규 프로젝트를 생성할 때 아바타에 대한 default 값이 변경됨 description 기존에는 avatarname avatartype bgname에 대해 미정 미정 으로 default 값으로 지정됐지만 fe 요청에 따라 미정 미정 으로 변경한다 todo list embeddable 아바타 클래스 수정하기 e t c optional 참조 및 링크 첨부하기
1
32,029
2,743,072,485
IssuesEvent
2015-04-21 19:44:37
afollestad/cabinet-issue-tracker
https://api.github.com/repos/afollestad/cabinet-issue-tracker
closed
Fast scroller crashes from Fabric.io
bug high priority
https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5534e10f5141dcfd8f847269 ```groovy java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2334) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2392) at android.app.ActivityThread.access$900(ActivityThread.java:169) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1280) at android.os.Handler.dispatchMessage(Handler.java:102) at android.os.Looper.loop(Looper.java:146) at android.app.ActivityThread.main(ActivityThread.java:5487) at java.lang.reflect.Method.invokeNative(Method.java) at java.lang.reflect.Method.invoke(Method.java:515) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1283) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1099) at dalvik.system.NativeStart.main(NativeStart.java) Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.view.LayoutInflater.createView(LayoutInflater.java:626) at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:702) ``` https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568 ```groovy java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087) at android.app.ActivityThread.access$600(ActivityThread.java:133) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:146) at android.app.ActivityThread.main(ActivityThread.java:4777) at java.lang.reflect.Method.invokeNative(Method.java) at java.lang.reflect.Method.invoke(Method.java:511) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556) at dalvik.system.NativeStart.main(NativeStart.java) Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.view.LayoutInflater.createView(LayoutInflater.java:613) at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687) ``` https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568 ```groovy java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087) at android.app.ActivityThread.access$600(ActivityThread.java:133) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:146) at android.app.ActivityThread.main(ActivityThread.java:4777) at java.lang.reflect.Method.invokeNative(Method.java) at java.lang.reflect.Method.invoke(Method.java:511) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556) at dalvik.system.NativeStart.main(NativeStart.java) Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.view.LayoutInflater.createView(LayoutInflater.java:613) at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687) ```
1.0
Fast scroller crashes from Fabric.io - https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5534e10f5141dcfd8f847269 ```groovy java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2334) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2392) at android.app.ActivityThread.access$900(ActivityThread.java:169) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1280) at android.os.Handler.dispatchMessage(Handler.java:102) at android.os.Looper.loop(Looper.java:146) at android.app.ActivityThread.main(ActivityThread.java:5487) at java.lang.reflect.Method.invokeNative(Method.java) at java.lang.reflect.Method.invoke(Method.java:515) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1283) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1099) at dalvik.system.NativeStart.main(NativeStart.java) Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.view.LayoutInflater.createView(LayoutInflater.java:626) at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:702) ``` https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568 ```groovy java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087) at android.app.ActivityThread.access$600(ActivityThread.java:133) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:146) at android.app.ActivityThread.main(ActivityThread.java:4777) at java.lang.reflect.Method.invokeNative(Method.java) at java.lang.reflect.Method.invoke(Method.java:511) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556) at dalvik.system.NativeStart.main(NativeStart.java) Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.view.LayoutInflater.createView(LayoutInflater.java:613) at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687) ``` https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568 ```groovy java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062) at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087) at android.app.ActivityThread.access$600(ActivityThread.java:133) at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198) at android.os.Handler.dispatchMessage(Handler.java:99) at android.os.Looper.loop(Looper.java:146) at android.app.ActivityThread.main(ActivityThread.java:4777) at java.lang.reflect.Method.invokeNative(Method.java) at java.lang.reflect.Method.invoke(Method.java:511) at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556) at dalvik.system.NativeStart.main(NativeStart.java) Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller at android.view.LayoutInflater.createView(LayoutInflater.java:613) at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687) ```
priority
fast scroller crashes from fabric io groovy java lang runtimeexception unable to start activity componentinfo com afollestad cabinet com afollestad cabinet ui mainactivity android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative method java at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main nativestart java caused by android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android view layoutinflater createview layoutinflater java at android view layoutinflater createviewfromtag layoutinflater java groovy java lang runtimeexception unable to start activity componentinfo com afollestad cabinet com afollestad cabinet ui mainactivity android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative method java at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main nativestart java caused by android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android view layoutinflater createview layoutinflater java at android view layoutinflater createviewfromtag layoutinflater java groovy java lang runtimeexception unable to start activity componentinfo com afollestad cabinet com afollestad cabinet ui mainactivity android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative method java at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main nativestart java caused by android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android view layoutinflater createview layoutinflater java at android view layoutinflater createviewfromtag layoutinflater java
1
464,733
13,339,134,609
IssuesEvent
2020-08-28 12:21:57
ruby-concurrency/concurrent-ruby
https://api.github.com/repos/ruby-concurrency/concurrent-ruby
reopened
Segmentation fault with ruby-2.6.3
bug high-priority
When updating an app to run with ruby 2.6.3 (from 2.4.4) I get the following error: ``` /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87: [BUG] Segmentation fault at 0x000000002222002d ruby 2.6.3p62 (2019-04-16 revision 67580) [x86_64-linux] -- Control frame information ----------------------------------------------- c:0002 p:0006 s:0007 e:000005 BLOCK /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var. [FINISH] c:0001 p:---- s:0003 e:000002 (none) [FINISH] -- Ruby level backtrace information ---------------------------------------- /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87:in `block (2 levels) in threadlocal_finalizer' -- Machine register context ------------------------------------------------ RIP: 0x00007fde9b71bf49 RBP: 0x0000000000021deb RSP: 0x00007fde97cb77e8 RAX: 0x000000000b275620 RBX: 0x0000000001719630 RCX: 0x000000002222002d RDX: 0x0000000000000000 RDI: 0x000000000b275620 RSI: 0x000000000a667f70 R8: 0x0000000001718d50 R9: 0x0000000001719618 R10: 0x0000000000000008 R11: 0x000000000a568038 R12: 0x000000000b2750f0 R13: 0x000000000a567ff8 R14: 0x0000000001719640 R15: 0x000000000a667f70 EFL: 0x0000000000010206 -- C level backtrace information ------------------------------------------- ``` I tried using `Concurrent.use_stdlib_logger(Logger::DEBUG)` to get more information but nothing else is printed to the console. How can I start debugging this problem? ``` * Operating system: linux * Ruby implementation: Ruby 2.6.3 * `concurrent-ruby` version: 1.1.5 * `concurrent-ruby-ext` installed: no * `concurrent-ruby-edge` used: no ```
1.0
Segmentation fault with ruby-2.6.3 - When updating an app to run with ruby 2.6.3 (from 2.4.4) I get the following error: ``` /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87: [BUG] Segmentation fault at 0x000000002222002d ruby 2.6.3p62 (2019-04-16 revision 67580) [x86_64-linux] -- Control frame information ----------------------------------------------- c:0002 p:0006 s:0007 e:000005 BLOCK /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var. [FINISH] c:0001 p:---- s:0003 e:000002 (none) [FINISH] -- Ruby level backtrace information ---------------------------------------- /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87:in `block (2 levels) in threadlocal_finalizer' -- Machine register context ------------------------------------------------ RIP: 0x00007fde9b71bf49 RBP: 0x0000000000021deb RSP: 0x00007fde97cb77e8 RAX: 0x000000000b275620 RBX: 0x0000000001719630 RCX: 0x000000002222002d RDX: 0x0000000000000000 RDI: 0x000000000b275620 RSI: 0x000000000a667f70 R8: 0x0000000001718d50 R9: 0x0000000001719618 R10: 0x0000000000000008 R11: 0x000000000a568038 R12: 0x000000000b2750f0 R13: 0x000000000a567ff8 R14: 0x0000000001719640 R15: 0x000000000a667f70 EFL: 0x0000000000010206 -- C level backtrace information ------------------------------------------- ``` I tried using `Concurrent.use_stdlib_logger(Logger::DEBUG)` to get more information but nothing else is printed to the console. How can I start debugging this problem? ``` * Operating system: linux * Ruby implementation: Ruby 2.6.3 * `concurrent-ruby` version: 1.1.5 * `concurrent-ruby-ext` installed: no * `concurrent-ruby-edge` used: no ```
priority
segmentation fault with ruby when updating an app to run with ruby from i get the following error home user rvm gems ruby project gems concurrent ruby lib concurrent atomic ruby thread local var rb segmentation fault at ruby revision control frame information c p s e block home user rvm gems ruby project gems concurrent ruby lib concurrent atomic ruby thread local var c p s e none ruby level backtrace information home user rvm gems ruby project gems concurrent ruby lib concurrent atomic ruby thread local var rb in block levels in threadlocal finalizer machine register context rip rbp rsp rax rbx rcx rdx rdi rsi efl c level backtrace information i tried using concurrent use stdlib logger logger debug to get more information but nothing else is printed to the console how can i start debugging this problem operating system linux ruby implementation ruby concurrent ruby version concurrent ruby ext installed no concurrent ruby edge used no
1
78,518
3,510,761,726
IssuesEvent
2016-01-09 18:58:50
Co0sh/BetonQuest
https://api.github.com/repos/Co0sh/BetonQuest
closed
More journal options
Feature High Priority
In the _config.yml_ file there should be these additional options: * configurable amount of characters per page * one entry per page option * reversed order of entries option * hide date/time in journal entries
1.0
More journal options - In the _config.yml_ file there should be these additional options: * configurable amount of characters per page * one entry per page option * reversed order of entries option * hide date/time in journal entries
priority
more journal options in the config yml file there should be these additional options configurable amount of characters per page one entry per page option reversed order of entries option hide date time in journal entries
1
526,738
15,300,311,458
IssuesEvent
2021-02-24 12:08:29
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Feature Gate
Priority/High Severity/Major feature identity-core improvement task
### Story - Why do we need the Feature Gate - Half-baked features might propagate to staging or prod environments. - Even if the feature development is completed, we might want to keep it disabled for time due to various reasons such as pricing. - We need to enable/disable some features based on the tenant. ### Requirements - Ability to turn on/off features on each environment separately. - Via config file - Via REST API - Identified features that will be controlled by Feature Gate. - https://docs.google.com/spreadsheets/d/14q4qD_O14Lvd78WZeyKy48tkHz0shzMwVWaOxdWwos8/edit#gid=0 ### Implementation - BE - API level - UI - Levels of block in UI - Hide from UI - Disable from UI (Greyed icons) - Need to show an indication for the feature status. - Experimental - New feature Beta feature
1.0
Feature Gate - ### Story - Why do we need the Feature Gate - Half-baked features might propagate to staging or prod environments. - Even if the feature development is completed, we might want to keep it disabled for time due to various reasons such as pricing. - We need to enable/disable some features based on the tenant. ### Requirements - Ability to turn on/off features on each environment separately. - Via config file - Via REST API - Identified features that will be controlled by Feature Gate. - https://docs.google.com/spreadsheets/d/14q4qD_O14Lvd78WZeyKy48tkHz0shzMwVWaOxdWwos8/edit#gid=0 ### Implementation - BE - API level - UI - Levels of block in UI - Hide from UI - Disable from UI (Greyed icons) - Need to show an indication for the feature status. - Experimental - New feature Beta feature
priority
feature gate story why do we need the feature gate half baked features might propagate to staging or prod environments even if the feature development is completed we might want to keep it disabled for time due to various reasons such as pricing we need to enable disable some features based on the tenant requirements ability to turn on off features on each environment separately via config file via rest api identified features that will be controlled by feature gate implementation be api level ui levels of block in ui hide from ui disable from ui greyed icons need to show an indication for the feature status experimental new feature beta feature
1
435,013
12,530,592,424
IssuesEvent
2020-06-04 13:19:35
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
opened
Elasticsearch exporter ignores errors on flush
Impact: Integration Scope: broker Severity: High Status: Needs Priority Type: Bug
**Describe the bug** If a failure occurs while exporting records to Elasticsearch then the failure is not reported and no retry happens. As a result, records may not be exported. **To Reproduce** Not sure how to trigger a failure expect using mocks. Found while fixing #4640. **Expected behavior** The failure is reported to the log and the exporter try again. **Log/Stacktrace** Nope. **Environment:** - OS: [e.g. Linux] - Zeebe Version: 0.24.0-alpha2 - Configuration: elasticsearch-exporter
1.0
Elasticsearch exporter ignores errors on flush - **Describe the bug** If a failure occurs while exporting records to Elasticsearch then the failure is not reported and no retry happens. As a result, records may not be exported. **To Reproduce** Not sure how to trigger a failure expect using mocks. Found while fixing #4640. **Expected behavior** The failure is reported to the log and the exporter try again. **Log/Stacktrace** Nope. **Environment:** - OS: [e.g. Linux] - Zeebe Version: 0.24.0-alpha2 - Configuration: elasticsearch-exporter
priority
elasticsearch exporter ignores errors on flush describe the bug if a failure occurs while exporting records to elasticsearch then the failure is not reported and no retry happens as a result records may not be exported to reproduce not sure how to trigger a failure expect using mocks found while fixing expected behavior the failure is reported to the log and the exporter try again log stacktrace nope environment os zeebe version configuration elasticsearch exporter
1
723,642
24,904,602,731
IssuesEvent
2022-10-29 04:32:50
AY2223S1-CS2103T-F12-2/tp
https://api.github.com/repos/AY2223S1-CS2103T-F12-2/tp
closed
Possible to save task with illegal values
priority.High
Entering the command : "addTask n/homework d/coding assignment pr/high c/backend dl/2022-12-12 pe/johnd@example.com" will cause the AddressBook to save the task with illegal vales. This results in the AddressBook starting with empty data the next time it is run
1.0
Possible to save task with illegal values - Entering the command : "addTask n/homework d/coding assignment pr/high c/backend dl/2022-12-12 pe/johnd@example.com" will cause the AddressBook to save the task with illegal vales. This results in the AddressBook starting with empty data the next time it is run
priority
possible to save task with illegal values entering the command addtask n homework d coding assignment pr high c backend dl pe johnd example com will cause the addressbook to save the task with illegal vales this results in the addressbook starting with empty data the next time it is run
1
99,001
4,044,937,408
IssuesEvent
2016-05-21 17:10:02
CPLamb/RouteTracker
https://api.github.com/repos/CPLamb/RouteTracker
opened
Rework Home screen to add final elements & fix constraints
Priority - High
1. As discussed a million times we need to get the constraints fixed. Test for largest iPhone on stimulator. 2. Add paper logo to this page above the filename field 3. Make the NewpaperTaxi logo properly fixed. It disappears for some stupid reason 4. Not sure what else 5. ---
1.0
Rework Home screen to add final elements & fix constraints - 1. As discussed a million times we need to get the constraints fixed. Test for largest iPhone on stimulator. 2. Add paper logo to this page above the filename field 3. Make the NewpaperTaxi logo properly fixed. It disappears for some stupid reason 4. Not sure what else 5. ---
priority
rework home screen to add final elements fix constraints as discussed a million times we need to get the constraints fixed test for largest iphone on stimulator add paper logo to this page above the filename field make the newpapertaxi logo properly fixed it disappears for some stupid reason not sure what else
1
784,348
27,567,428,888
IssuesEvent
2023-03-08 05:49:54
curiouslearning/FeedTheMonsterJS
https://api.github.com/repos/curiouslearning/FeedTheMonsterJS
closed
Create manifest file for all apps in container app
High Priority
File should ideally contain these infos:- 1. App icon 2. Title 3. Url **Acceptance criteria**: Manifest file should ideally contain the enough information to render app icons on container app. Clicking on icon should take us to specified link.
1.0
Create manifest file for all apps in container app - File should ideally contain these infos:- 1. App icon 2. Title 3. Url **Acceptance criteria**: Manifest file should ideally contain the enough information to render app icons on container app. Clicking on icon should take us to specified link.
priority
create manifest file for all apps in container app file should ideally contain these infos app icon title url acceptance criteria manifest file should ideally contain the enough information to render app icons on container app clicking on icon should take us to specified link
1
720,059
24,777,003,526
IssuesEvent
2022-10-23 21:11:16
RoboJackets/urc-drone
https://api.github.com/repos/RoboJackets/urc-drone
closed
Gazebo Sim Drone
area ➤ misc priority ➤ high level ➤ hard
Create a gazebo model with dynamics for the drone. This is very necessary for testing our flight, land, and control procedures.
1.0
Gazebo Sim Drone - Create a gazebo model with dynamics for the drone. This is very necessary for testing our flight, land, and control procedures.
priority
gazebo sim drone create a gazebo model with dynamics for the drone this is very necessary for testing our flight land and control procedures
1
601,292
18,396,894,095
IssuesEvent
2021-10-12 12:25:36
bounswe/2021SpringGroup4
https://api.github.com/repos/bounswe/2021SpringGroup4
closed
Creating wiki page for Fatih Akgöz
wiki individual Priority: High
1. Personal wiki page will be created and filled. 2. Weekly Effort wiki page will be created and filled.
1.0
Creating wiki page for Fatih Akgöz - 1. Personal wiki page will be created and filled. 2. Weekly Effort wiki page will be created and filled.
priority
creating wiki page for fatih akgöz personal wiki page will be created and filled weekly effort wiki page will be created and filled
1
213,288
7,247,673,561
IssuesEvent
2018-02-15 04:50:58
yanis333/SOEN341_Website
https://api.github.com/repos/yanis333/SOEN341_Website
closed
#56 Upvote & DownVote : add the functionality & the permission to Accepte or Remove Reply
High value Priority 2 Risk1 feature sprint 2
This is one of the feature for issue #56 [sp]=3 Need to add the Upvote and DownVote for the reply
1.0
#56 Upvote & DownVote : add the functionality & the permission to Accepte or Remove Reply - This is one of the feature for issue #56 [sp]=3 Need to add the Upvote and DownVote for the reply
priority
upvote downvote add the functionality the permission to accepte or remove reply this is one of the feature for issue need to add the upvote and downvote for the reply
1
210,660
7,192,110,948
IssuesEvent
2018-02-03 00:21:59
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: Webserver fails to start for solo game in 6.4.2
High Priority
**Version:** 0.6.4.2 alpha **Steps to Reproduce:** Start any solo game with a pre existing game save **Expected behavior:** webserver should start so that a webpage will generate **Actual behavior:** webserver fails to initialize. This error in console: Failed to start Webserver: Exception has been thrown by the target of an invocation.
1.0
USER ISSUE: Webserver fails to start for solo game in 6.4.2 - **Version:** 0.6.4.2 alpha **Steps to Reproduce:** Start any solo game with a pre existing game save **Expected behavior:** webserver should start so that a webpage will generate **Actual behavior:** webserver fails to initialize. This error in console: Failed to start Webserver: Exception has been thrown by the target of an invocation.
priority
user issue webserver fails to start for solo game in version alpha steps to reproduce start any solo game with a pre existing game save expected behavior webserver should start so that a webpage will generate actual behavior webserver fails to initialize this error in console failed to start webserver exception has been thrown by the target of an invocation
1
802,196
28,780,886,697
IssuesEvent
2023-05-02 00:27:08
agency-of-learning/PairApp
https://api.github.com/repos/agency-of-learning/PairApp
opened
Create dev Procfile
good first issue high priority solo
We should add a `Procfile.dev` that will help with setting up our local host. Params we want are: 1. server 2. worker
1.0
Create dev Procfile - We should add a `Procfile.dev` that will help with setting up our local host. Params we want are: 1. server 2. worker
priority
create dev procfile we should add a procfile dev that will help with setting up our local host params we want are server worker
1
181,259
6,657,791,431
IssuesEvent
2017-09-30 10:38:27
status-im/status-react
https://api.github.com/repos/status-im/status-react
closed
Second send transaction can't be sent as send button does nothing and fee value is not shown [feature/upgrade-status-go-g0270590]
blocked bug high-priority
### Description *Type*: Bug *Summary*: User can create 1 send transaction in 1-1 chat. If user does not confirm this transaction and returns to 1-1 chat in order to make 2nd send transaction then it's stuck on Fee step: 1. There is no fee value shown 2. If tap on send button wile input filed contains `/send username 0.2` then nothing happens. Expected: Unsigned screen is shown with 2 transactions. ![img_0579](https://user-images.githubusercontent.com/6584810/30162725-9eded380-93de-11e7-856a-a29b99efa254.jpg) #### Expected behavior can send second time send transaction #### Actual behavior can't send 2nd transaction as tap on send does nothing ### Reproduction from 00:12 https://app.testfairy.com/projects/4803590-status/builds/6511547/sessions/5/?accessToken=1xw9BL-p5dEIONbItbxAqqE99xg - Open Status - Open 1-1 chat - tap on /send, select username and type some sum that can be send from your wallet (enough balance), e.g `/send username 0.1` - on unsigned `transaction screen tap on x in top left corner to close the screen - in 1-1 chat tap on /send, select username and type some other sum to send. As a result Fee will be shown. Tap on send button ### Additional Information * Status version: PR `feature/upgrade-status-go-g0270590` 0.9.11d60 (1429.188) * Operating System: Real device iPhone 6s, iOS 10.3.3
1.0
Second send transaction can't be sent as send button does nothing and fee value is not shown [feature/upgrade-status-go-g0270590] - ### Description *Type*: Bug *Summary*: User can create 1 send transaction in 1-1 chat. If user does not confirm this transaction and returns to 1-1 chat in order to make 2nd send transaction then it's stuck on Fee step: 1. There is no fee value shown 2. If tap on send button wile input filed contains `/send username 0.2` then nothing happens. Expected: Unsigned screen is shown with 2 transactions. ![img_0579](https://user-images.githubusercontent.com/6584810/30162725-9eded380-93de-11e7-856a-a29b99efa254.jpg) #### Expected behavior can send second time send transaction #### Actual behavior can't send 2nd transaction as tap on send does nothing ### Reproduction from 00:12 https://app.testfairy.com/projects/4803590-status/builds/6511547/sessions/5/?accessToken=1xw9BL-p5dEIONbItbxAqqE99xg - Open Status - Open 1-1 chat - tap on /send, select username and type some sum that can be send from your wallet (enough balance), e.g `/send username 0.1` - on unsigned `transaction screen tap on x in top left corner to close the screen - in 1-1 chat tap on /send, select username and type some other sum to send. As a result Fee will be shown. Tap on send button ### Additional Information * Status version: PR `feature/upgrade-status-go-g0270590` 0.9.11d60 (1429.188) * Operating System: Real device iPhone 6s, iOS 10.3.3
priority
second send transaction can t be sent as send button does nothing and fee value is not shown description type bug summary user can create send transaction in chat if user does not confirm this transaction and returns to chat in order to make send transaction then it s stuck on fee step there is no fee value shown if tap on send button wile input filed contains send username then nothing happens expected unsigned screen is shown with transactions expected behavior can send second time send transaction actual behavior can t send transaction as tap on send does nothing reproduction from open status open chat tap on send select username and type some sum that can be send from your wallet enough balance e g send username on unsigned transaction screen tap on x in top left corner to close the screen in chat tap on send select username and type some other sum to send as a result fee will be shown tap on send button additional information status version pr feature upgrade status go operating system real device iphone ios
1
383,517
11,357,067,150
IssuesEvent
2020-01-25 01:39:28
v0idp/Mellow
https://api.github.com/repos/v0idp/Mellow
closed
save password in config file with encryption
Hight Priority Rework
right now the password is in plaintext which is a no go. needs to be reworked alongside the auth and session code
1.0
save password in config file with encryption - right now the password is in plaintext which is a no go. needs to be reworked alongside the auth and session code
priority
save password in config file with encryption right now the password is in plaintext which is a no go needs to be reworked alongside the auth and session code
1
614,834
19,190,541,514
IssuesEvent
2021-12-05 22:43:07
MarkMcGregor-dev/BloodySimulatorGame
https://api.github.com/repos/MarkMcGregor-dev/BloodySimulatorGame
closed
Fix score not being reset after after death
bug high priority
Score looks like it resets after respawning (goes back to zero) but immediately shows the previous round's score after increasing
1.0
Fix score not being reset after after death - Score looks like it resets after respawning (goes back to zero) but immediately shows the previous round's score after increasing
priority
fix score not being reset after after death score looks like it resets after respawning goes back to zero but immediately shows the previous round s score after increasing
1
491,450
14,164,251,297
IssuesEvent
2020-11-12 04:32:58
wso2/docs-ei
https://api.github.com/repos/wso2/docs-ei
closed
Update call template and sequence template docs with new features
Priority/High micro-integrator
**Description:** Update documentation with the new features introduced with https://github.com/wso2/micro-integrator/issues/1673 https://ei.docs.wso2.com/en/latest/micro-integrator/references/mediators/call-Template-Mediator/
1.0
Update call template and sequence template docs with new features - **Description:** Update documentation with the new features introduced with https://github.com/wso2/micro-integrator/issues/1673 https://ei.docs.wso2.com/en/latest/micro-integrator/references/mediators/call-Template-Mediator/
priority
update call template and sequence template docs with new features description update documentation with the new features introduced with
1
23,855
2,664,564,320
IssuesEvent
2015-03-20 15:16:52
cwrc/CWRC-Writer
https://api.github.com/repos/cwrc/CWRC-Writer
opened
Avoiding tag redundancy when adding annotations to existing tag
enhancement high priority
http://apps.testing.cwrc.ca/editor/dev/editor_dev.htm#templates/ceww_new_entry_template Using this template brings home the problems of annotating existing markup: you end up with duplicate tags and it's hard to figure out which one has the entity and which one should be deleted. James or Andrew, can you think of any creative way of dealing with this? Could we add a right click functionality along the lines of "add entity to this tag"?
1.0
Avoiding tag redundancy when adding annotations to existing tag - http://apps.testing.cwrc.ca/editor/dev/editor_dev.htm#templates/ceww_new_entry_template Using this template brings home the problems of annotating existing markup: you end up with duplicate tags and it's hard to figure out which one has the entity and which one should be deleted. James or Andrew, can you think of any creative way of dealing with this? Could we add a right click functionality along the lines of "add entity to this tag"?
priority
avoiding tag redundancy when adding annotations to existing tag using this template brings home the problems of annotating existing markup you end up with duplicate tags and it s hard to figure out which one has the entity and which one should be deleted james or andrew can you think of any creative way of dealing with this could we add a right click functionality along the lines of add entity to this tag
1
731,903
25,236,040,321
IssuesEvent
2022-11-15 01:01:28
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
New Request - DOI's for Projects
Priority-High (Needed for work) Function-PublicationOrProject Enhancement
With journals requiring more publicly available data, we would like to be able to cite Arctos datasets used in original research. Arctos Projects are already set up to associate data loans and other transactions and we can link to journal etc. So being able to then share a DOI for the Project URL would be a huge boon. Currently the project does have a stable URI but it is not recognized as a digital object. (GBIF is not a good option for a variety of reasons...) This would be available across all collection types and has a lot of potential Can we add a new function button in the edit Project-- generate DOI? This would need to be displayed in the public view when available. This would not be required.
1.0
New Request - DOI's for Projects - With journals requiring more publicly available data, we would like to be able to cite Arctos datasets used in original research. Arctos Projects are already set up to associate data loans and other transactions and we can link to journal etc. So being able to then share a DOI for the Project URL would be a huge boon. Currently the project does have a stable URI but it is not recognized as a digital object. (GBIF is not a good option for a variety of reasons...) This would be available across all collection types and has a lot of potential Can we add a new function button in the edit Project-- generate DOI? This would need to be displayed in the public view when available. This would not be required.
priority
new request doi s for projects with journals requiring more publicly available data we would like to be able to cite arctos datasets used in original research arctos projects are already set up to associate data loans and other transactions and we can link to journal etc so being able to then share a doi for the project url would be a huge boon currently the project does have a stable uri but it is not recognized as a digital object gbif is not a good option for a variety of reasons this would be available across all collection types and has a lot of potential can we add a new function button in the edit project generate doi this would need to be displayed in the public view when available this would not be required
1
715,206
24,590,081,818
IssuesEvent
2022-10-14 00:44:45
OregonDigital/OD2
https://api.github.com/repos/OregonDigital/OD2
opened
Cache cacheable content and investigate additional caching options
Enhancement Priority - High MVP Review ruby
### Descriptive summary OD1 is still able to take advantage of Rails 4's Page and Action caching, as you can see with the following command on OD1: `memcached-tool localhost display` Rails 5 removed the integrated Page and Action caching, which means OD2 does not yet have the same level of caching that OD1 has. Without widespread caching, OD2 is going to have a difficult time handling large amounts of traffic. We are able to take advantage of other caches like ActiveFedora's LDP cache for Fedora content. We have Blazegraph for caching triples. We have caching for SQL queries to our primary PostgreSQL database. But to handle larger amounts of traffic and decrease overall response times we need more caching for our Rails frontend application. Without caching, Rails has to query backend services for every request, which manifests as longer and longer response times as backend services get more busy. Our average response time when the system is relatively idle ranges from 1.5 - 2.5 seconds. While large amounts of Sidekiq jobs are running this can increase to 5 seconds to 20+ seconds. With more caching we can maintain a more stable User Experience regardless of how busy the overall system is. This will also help to lighten the load on backend services. In addition to the extra resources that background jobs will have available, caching can be added to background jobs where it makes sense to speed them up even further. Since earlier this week we're now hooked up to `memcached` via `Rails.application.config.cache_store`. We can `write` and `fetch` keys from the `Rails.cache`. Without anything taking advantage of that our `cache_store` is empty, or mostly empty. ``` irb(main):034:0> Rails.cache.stats["memcache.od2-prod.svc.cluster.local:11211"]["total_items"] => "1" ``` This was after creating and clearing keys manually, otherwise it would still be returning 0. Out of the box we can [cache fragments and partials](https://guides.rubyonrails.org/caching_with_rails.html#fragment-caching) To make use of caching for OD2, we will need to add caching code for content that should be cached. Additionally, Rack Cache can cache content without changes to code. See the References section for a link. The Page and Action caching that are missing in Rails 5 are available as gems, which we should test on Staging. See the References section for links. We can also do low level caching with `Rails.cache.fetch()` and `Rails.cache.write()` for data from backends that won't change often, or are acceptable to be eventually correct. Whether that's SOLR query results, or responses from other external APIs, if the response changes infrequently or eventually correct results are ok we should be caching them. See the low level caching section of the Ruby on Rails Caching Guide for details. ### Expected behavior Infrequently changed content should be cached wherever possible. We should be seeing increasing numbers of items in the memcache cache_store. ### References [https://guides.rubyonrails.org/caching_with_rails.html](https://guides.rubyonrails.org/caching_with_rails.html) [https://rtomayko.github.io/rack-cache/configuration](https://rtomayko.github.io/rack-cache/configuration) [https://github.com/rails/actionpack-page_caching](https://github.com/rails/actionpack-page_caching)
1.0
Cache cacheable content and investigate additional caching options - ### Descriptive summary OD1 is still able to take advantage of Rails 4's Page and Action caching, as you can see with the following command on OD1: `memcached-tool localhost display` Rails 5 removed the integrated Page and Action caching, which means OD2 does not yet have the same level of caching that OD1 has. Without widespread caching, OD2 is going to have a difficult time handling large amounts of traffic. We are able to take advantage of other caches like ActiveFedora's LDP cache for Fedora content. We have Blazegraph for caching triples. We have caching for SQL queries to our primary PostgreSQL database. But to handle larger amounts of traffic and decrease overall response times we need more caching for our Rails frontend application. Without caching, Rails has to query backend services for every request, which manifests as longer and longer response times as backend services get more busy. Our average response time when the system is relatively idle ranges from 1.5 - 2.5 seconds. While large amounts of Sidekiq jobs are running this can increase to 5 seconds to 20+ seconds. With more caching we can maintain a more stable User Experience regardless of how busy the overall system is. This will also help to lighten the load on backend services. In addition to the extra resources that background jobs will have available, caching can be added to background jobs where it makes sense to speed them up even further. Since earlier this week we're now hooked up to `memcached` via `Rails.application.config.cache_store`. We can `write` and `fetch` keys from the `Rails.cache`. Without anything taking advantage of that our `cache_store` is empty, or mostly empty. ``` irb(main):034:0> Rails.cache.stats["memcache.od2-prod.svc.cluster.local:11211"]["total_items"] => "1" ``` This was after creating and clearing keys manually, otherwise it would still be returning 0. Out of the box we can [cache fragments and partials](https://guides.rubyonrails.org/caching_with_rails.html#fragment-caching) To make use of caching for OD2, we will need to add caching code for content that should be cached. Additionally, Rack Cache can cache content without changes to code. See the References section for a link. The Page and Action caching that are missing in Rails 5 are available as gems, which we should test on Staging. See the References section for links. We can also do low level caching with `Rails.cache.fetch()` and `Rails.cache.write()` for data from backends that won't change often, or are acceptable to be eventually correct. Whether that's SOLR query results, or responses from other external APIs, if the response changes infrequently or eventually correct results are ok we should be caching them. See the low level caching section of the Ruby on Rails Caching Guide for details. ### Expected behavior Infrequently changed content should be cached wherever possible. We should be seeing increasing numbers of items in the memcache cache_store. ### References [https://guides.rubyonrails.org/caching_with_rails.html](https://guides.rubyonrails.org/caching_with_rails.html) [https://rtomayko.github.io/rack-cache/configuration](https://rtomayko.github.io/rack-cache/configuration) [https://github.com/rails/actionpack-page_caching](https://github.com/rails/actionpack-page_caching)
priority
cache cacheable content and investigate additional caching options descriptive summary is still able to take advantage of rails s page and action caching as you can see with the following command on memcached tool localhost display rails removed the integrated page and action caching which means does not yet have the same level of caching that has without widespread caching is going to have a difficult time handling large amounts of traffic we are able to take advantage of other caches like activefedora s ldp cache for fedora content we have blazegraph for caching triples we have caching for sql queries to our primary postgresql database but to handle larger amounts of traffic and decrease overall response times we need more caching for our rails frontend application without caching rails has to query backend services for every request which manifests as longer and longer response times as backend services get more busy our average response time when the system is relatively idle ranges from seconds while large amounts of sidekiq jobs are running this can increase to seconds to seconds with more caching we can maintain a more stable user experience regardless of how busy the overall system is this will also help to lighten the load on backend services in addition to the extra resources that background jobs will have available caching can be added to background jobs where it makes sense to speed them up even further since earlier this week we re now hooked up to memcached via rails application config cache store we can write and fetch keys from the rails cache without anything taking advantage of that our cache store is empty or mostly empty irb main rails cache stats this was after creating and clearing keys manually otherwise it would still be returning out of the box we can to make use of caching for we will need to add caching code for content that should be cached additionally rack cache can cache content without changes to code see the references section for a link the page and action caching that are missing in rails are available as gems which we should test on staging see the references section for links we can also do low level caching with rails cache fetch and rails cache write for data from backends that won t change often or are acceptable to be eventually correct whether that s solr query results or responses from other external apis if the response changes infrequently or eventually correct results are ok we should be caching them see the low level caching section of the ruby on rails caching guide for details expected behavior infrequently changed content should be cached wherever possible we should be seeing increasing numbers of items in the memcache cache store references
1
816,803
30,612,984,631
IssuesEvent
2023-07-23 21:02:44
silverstripe/addons.silverstripe.org
https://api.github.com/repos/silverstripe/addons.silverstripe.org
closed
Create a generic SS community sites base CSS
type/enhancement priority-high
To be used with bootstrap-based sites, for now primarily api.ss.org, extensions.ss.org and eventually ss.org. Note that this is different from a SilverStripe theme, since it should be useable in non-SS projects like the staticically generated api.ss.org as well.
1.0
Create a generic SS community sites base CSS - To be used with bootstrap-based sites, for now primarily api.ss.org, extensions.ss.org and eventually ss.org. Note that this is different from a SilverStripe theme, since it should be useable in non-SS projects like the staticically generated api.ss.org as well.
priority
create a generic ss community sites base css to be used with bootstrap based sites for now primarily api ss org extensions ss org and eventually ss org note that this is different from a silverstripe theme since it should be useable in non ss projects like the staticically generated api ss org as well
1
536,568
15,710,817,132
IssuesEvent
2021-03-27 03:49:02
Blackoutburst/Wally
https://api.github.com/repos/Blackoutburst/Wally
closed
Search commands does not support player title and custom background
bug high priority incorrect
**Describe the bug** When displaying player stats using `!search` their title and background does not display **To Reproduce** Steps to reproduce the behavior: 1. use `!search` on any player with a custom title and/or background **Expected behavior** It should display everything like we used `!stats` on the player **Screenshots** Output ![stats](https://user-images.githubusercontent.com/30992311/112703474-5a3bb380-8e97-11eb-865d-4daac855afe6.png) Expected ![stats](https://user-images.githubusercontent.com/30992311/112703477-6293ee80-8e97-11eb-8f1b-30382b9c0fb1.png)
1.0
Search commands does not support player title and custom background - **Describe the bug** When displaying player stats using `!search` their title and background does not display **To Reproduce** Steps to reproduce the behavior: 1. use `!search` on any player with a custom title and/or background **Expected behavior** It should display everything like we used `!stats` on the player **Screenshots** Output ![stats](https://user-images.githubusercontent.com/30992311/112703474-5a3bb380-8e97-11eb-865d-4daac855afe6.png) Expected ![stats](https://user-images.githubusercontent.com/30992311/112703477-6293ee80-8e97-11eb-8f1b-30382b9c0fb1.png)
priority
search commands does not support player title and custom background describe the bug when displaying player stats using search their title and background does not display to reproduce steps to reproduce the behavior use search on any player with a custom title and or background expected behavior it should display everything like we used stats on the player screenshots output expected
1
575,977
17,067,898,579
IssuesEvent
2021-07-07 09:36:51
geosolutions-it/MapStore2-C040
https://api.github.com/repos/geosolutions-it/MapStore2-C040
opened
Install/Configure Grid Shift Transform files in GeoServer
C040-COMUNE_GE-2021-SUPPORT Priority: High Task
### Description Install the Grid Shift Transform files in GeoServer for the CRS 7791 using the transform file provided by the client. Files are available at this path: \\vm-linuxgeofe\MapStore2\grigliati_GE.zip The documentation to install Grid Shift Transform files is available in the [GeoServer online doc](https://docs.geoserver.org/stable/en/user/configuration/crshandling/coordtransforms.html#add-grid-shift-transform-files). ### Other useful information (optional):
1.0
Install/Configure Grid Shift Transform files in GeoServer - ### Description Install the Grid Shift Transform files in GeoServer for the CRS 7791 using the transform file provided by the client. Files are available at this path: \\vm-linuxgeofe\MapStore2\grigliati_GE.zip The documentation to install Grid Shift Transform files is available in the [GeoServer online doc](https://docs.geoserver.org/stable/en/user/configuration/crshandling/coordtransforms.html#add-grid-shift-transform-files). ### Other useful information (optional):
priority
install configure grid shift transform files in geoserver description install the grid shift transform files in geoserver for the crs using the transform file provided by the client files are available at this path vm linuxgeofe grigliati ge zip the documentation to install grid shift transform files is available in the other useful information optional
1
109,456
4,387,744,036
IssuesEvent
2016-08-08 16:41:12
danya02/public-bodyguard
https://api.github.com/repos/danya02/public-bodyguard
closed
Presentation
meta Very-High-Priority
We need a presentation for the project to explain the project, as per the rules of the project for which this one is being developed.
1.0
Presentation - We need a presentation for the project to explain the project, as per the rules of the project for which this one is being developed.
priority
presentation we need a presentation for the project to explain the project as per the rules of the project for which this one is being developed
1
539,619
15,792,099,832
IssuesEvent
2021-04-02 06:20:32
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
closed
Alarm policies that do not set thresholds and monitor targets can still generate alarms
area/alerting kind/bug kind/need-to-verify priority/high
<!-- You don't need to remove this comment section, it's invisible on the issues page. ## General remarks * Attention, please fill out this issues form using English only! * 注意!GitHub Issue 仅支持英文,中文 Issue 请在 [论坛](https://kubesphere.com.cn/forum/) 提交。 * This form is to report bugs. For general usage questions you can join our Slack channel [KubeSphere-users](https://join.slack.com/t/kubesphere/shared_invite/enQtNTE3MDIxNzUxNzQ0LTZkNTdkYWNiYTVkMTM5ZThhODY1MjAyZmVlYWEwZmQ3ODQ1NmM1MGVkNWEzZTRhNzk0MzM5MmY4NDc3ZWVhMjE) --> **Describe the Bug** Alarm policies that do not set thresholds and monitor targets can still generate alarms **Environment** http://xxxx:79:30104 **Preset conditions** 1、There is account A, which has cluster management privileges. 2、Use account A to log in to KS, and go to Platform management--》Cluster management--》Alerting Policies. **How To Reproduce** Steps to reproduce the behavior: 1. click "create" 2. Enter the name "test" 3. Set the alarm duration to 1 minute 4. Set the alarm rule expression to "node:node_cpu_utilisation:avg1m" 5. save policy 6. Check whether policy “test” produces an alert message **Expected behavior** Policy “test” does not produce an alert message ![image](https://user-images.githubusercontent.com/68640256/109524343-5bbdca00-7aeb-11eb-9c61-a9c6cd9b791c.png) ![image](https://user-images.githubusercontent.com/68640256/109524376-68422280-7aeb-11eb-9ea3-09618ba85f5c.png) /priority high /area alerting /cc @junotx /kind bug /milestone 3.1.0
1.0
Alarm policies that do not set thresholds and monitor targets can still generate alarms - <!-- You don't need to remove this comment section, it's invisible on the issues page. ## General remarks * Attention, please fill out this issues form using English only! * 注意!GitHub Issue 仅支持英文,中文 Issue 请在 [论坛](https://kubesphere.com.cn/forum/) 提交。 * This form is to report bugs. For general usage questions you can join our Slack channel [KubeSphere-users](https://join.slack.com/t/kubesphere/shared_invite/enQtNTE3MDIxNzUxNzQ0LTZkNTdkYWNiYTVkMTM5ZThhODY1MjAyZmVlYWEwZmQ3ODQ1NmM1MGVkNWEzZTRhNzk0MzM5MmY4NDc3ZWVhMjE) --> **Describe the Bug** Alarm policies that do not set thresholds and monitor targets can still generate alarms **Environment** http://xxxx:79:30104 **Preset conditions** 1、There is account A, which has cluster management privileges. 2、Use account A to log in to KS, and go to Platform management--》Cluster management--》Alerting Policies. **How To Reproduce** Steps to reproduce the behavior: 1. click "create" 2. Enter the name "test" 3. Set the alarm duration to 1 minute 4. Set the alarm rule expression to "node:node_cpu_utilisation:avg1m" 5. save policy 6. Check whether policy “test” produces an alert message **Expected behavior** Policy “test” does not produce an alert message ![image](https://user-images.githubusercontent.com/68640256/109524343-5bbdca00-7aeb-11eb-9c61-a9c6cd9b791c.png) ![image](https://user-images.githubusercontent.com/68640256/109524376-68422280-7aeb-11eb-9ea3-09618ba85f5c.png) /priority high /area alerting /cc @junotx /kind bug /milestone 3.1.0
priority
alarm policies that do not set thresholds and monitor targets can still generate alarms you don t need to remove this comment section it s invisible on the issues page general remarks attention please fill out this issues form using english only 注意!github issue 仅支持英文,中文 issue 请在 提交。 this form is to report bugs for general usage questions you can join our slack channel describe the bug alarm policies that do not set thresholds and monitor targets can still generate alarms environment preset conditions 、there is account a which has cluster management privileges 、use account a to log in to ks and go to platform management 》cluster management 》alerting policies how to reproduce steps to reproduce the behavior click create enter the name test set the alarm duration to minute set the alarm rule expression to node node cpu utilisation save policy check whether policy “test” produces an alert message expected behavior policy “test” does not produce an alert message priority high area alerting cc junotx kind bug milestone
1
192,443
6,849,982,590
IssuesEvent
2017-11-14 00:37:04
eriq-augustine/psl
https://api.github.com/repos/eriq-augustine/psl
opened
More Aggressive Atom Type Conversion
Difficulty - Easy Priority - High Type - Bug
We validate and do some mild type conversions for atoms (not ground atom) that come from rules. https://github.com/eriq-augustine/psl/blob/develop/psl-core/src/main/java/org/linqs/psl/model/atom/Atom.java#L113 Like: ``` Foo(+A, 'Dan') = 1 . ``` But we should be doing more aggressive type conversions from strings (since all constants must be quoted). A string can always be attempted to be converted into whatever type.
1.0
More Aggressive Atom Type Conversion - We validate and do some mild type conversions for atoms (not ground atom) that come from rules. https://github.com/eriq-augustine/psl/blob/develop/psl-core/src/main/java/org/linqs/psl/model/atom/Atom.java#L113 Like: ``` Foo(+A, 'Dan') = 1 . ``` But we should be doing more aggressive type conversions from strings (since all constants must be quoted). A string can always be attempted to be converted into whatever type.
priority
more aggressive atom type conversion we validate and do some mild type conversions for atoms not ground atom that come from rules like foo a dan but we should be doing more aggressive type conversions from strings since all constants must be quoted a string can always be attempted to be converted into whatever type
1
183,101
6,677,221,544
IssuesEvent
2017-10-05 09:35:15
metasfresh/metasfresh
https://api.github.com/repos/metasfresh/metasfresh
closed
Make letter print preview work with AD_BoilerPlate jasper process
branch:master priority:high
### Is this a bug or feature request? Bug ### What is the current behavior? jasper process is not used feature was originally implemented here: https://github.com/metasfresh/metasfresh-webui-api/issues/178 #### Which are the steps to reproduce? 1. define a text tempate (boilerplate) 1. select jasper process in template 1. go in existing sales order 1. press letter action 1. select template 1. click print preview 1. OK: letter is opened showing the text of template 1. NOK: jasper process is not used ### What is the expected or desired behavior? When you print preview form the letter window, the jasper assign to the boiler plate to be included in the preview.
1.0
Make letter print preview work with AD_BoilerPlate jasper process - ### Is this a bug or feature request? Bug ### What is the current behavior? jasper process is not used feature was originally implemented here: https://github.com/metasfresh/metasfresh-webui-api/issues/178 #### Which are the steps to reproduce? 1. define a text tempate (boilerplate) 1. select jasper process in template 1. go in existing sales order 1. press letter action 1. select template 1. click print preview 1. OK: letter is opened showing the text of template 1. NOK: jasper process is not used ### What is the expected or desired behavior? When you print preview form the letter window, the jasper assign to the boiler plate to be included in the preview.
priority
make letter print preview work with ad boilerplate jasper process is this a bug or feature request bug what is the current behavior jasper process is not used feature was originally implemented here which are the steps to reproduce define a text tempate boilerplate select jasper process in template go in existing sales order press letter action select template click print preview ok letter is opened showing the text of template nok jasper process is not used what is the expected or desired behavior when you print preview form the letter window the jasper assign to the boiler plate to be included in the preview
1
249,999
7,966,331,840
IssuesEvent
2018-07-14 20:38:52
City-Bureau/city-scrapers
https://api.github.com/repos/City-Bureau/city-scrapers
opened
Non-coding Contributions: Local Government Spreadsheet (Chicago and Detroit)
help wanted non-coding non-coding: agency research non-coding: good first issue priority: high (must have)
We're collecting public meeting information from local governing bodies but there's a good deal of supporting information that'll help us increase access to information on those bodies. If you're interested in a non-coding task this one can be done at your own pace and would help us survey the local political landscape. Information we need help collecting: - Phone - Email - Twitter Handle - Meeting Agenda URL - Meeting Minutes URL - Board/Commission Members - Annual Budget - Officer Selection Process See the Chicago and Detroit tabs of the Public Meetings Agency Table—Columns J-Q (i.e. the blue columns). Contribute info for any department by adding the comments: https://docs.google.com/spreadsheets/d/1Uu9FmcQAy8AXhBR2bj84LbSdlfgMNNroa9P89rsDJSQ/edit#gid=0 (You can request access to the document to contribute information directly).
1.0
Non-coding Contributions: Local Government Spreadsheet (Chicago and Detroit) - We're collecting public meeting information from local governing bodies but there's a good deal of supporting information that'll help us increase access to information on those bodies. If you're interested in a non-coding task this one can be done at your own pace and would help us survey the local political landscape. Information we need help collecting: - Phone - Email - Twitter Handle - Meeting Agenda URL - Meeting Minutes URL - Board/Commission Members - Annual Budget - Officer Selection Process See the Chicago and Detroit tabs of the Public Meetings Agency Table—Columns J-Q (i.e. the blue columns). Contribute info for any department by adding the comments: https://docs.google.com/spreadsheets/d/1Uu9FmcQAy8AXhBR2bj84LbSdlfgMNNroa9P89rsDJSQ/edit#gid=0 (You can request access to the document to contribute information directly).
priority
non coding contributions local government spreadsheet chicago and detroit we re collecting public meeting information from local governing bodies but there s a good deal of supporting information that ll help us increase access to information on those bodies if you re interested in a non coding task this one can be done at your own pace and would help us survey the local political landscape information we need help collecting phone email twitter handle meeting agenda url meeting minutes url board commission members annual budget officer selection process see the chicago and detroit tabs of the public meetings agency table—columns j q i e the blue columns contribute info for any department by adding the comments you can request access to the document to contribute information directly
1
87,189
3,739,015,615
IssuesEvent
2016-03-09 01:46:41
projectcalico/calico-containers
https://api.github.com/repos/projectcalico/calico-containers
closed
Calico Node/Bird routes through default interface(eth0) and not tunnelled interface (tunl0)
Priority: High
How did i notice the issue - docker instance on Machine1 was not able to communicate with docker instance on Machine2 inspite of they being part of same profile and the profile had default settings which is allow all communication within the profile. Background: I was running a few docker instances on Machine 1(192.168.0.12 , 13 and 14) and they were part of same profile so obviously as expected they were able to communicate with each other. Now i configured calico on another machine(Machine 2) and created a new docker instance (192.168.0.15) and adding it to same profile as the above mentioned docker instances, But now the docker instances on Machine1 were not able to reach instances on Machine2 and viceversa. ip route show output from both boxes: **Machine 1: ** default via 172.31.16.1 dev eth0 10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1 172.18.0.0/16 dev docker_gwbridge proto kernel scope link src 172.18.0.1 172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.27.45 blackhole 192.168.0.0/26 proto bird 192.168.0.12 dev calie8beb266c79 scope link 192.168.0.13 dev caliefab0610c79 scope link 192.168.0.14 dev calif2c7a52ec79 scope link 192.168.0.15 via 172.31.29.77 dev tunl0 proto bird onlink 192.168.0.64/26 via 172.31.29.77 dev tunl0 proto bird onlink **Machine 2: ** default via 172.31.16.1 dev eth0 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1 172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77 192.168.0.0/26 via 172.31.27.45 dev eth0 proto bird 192.168.0.15 dev cali7fcbe484c7a scope link blackhole 192.168.0.64/26 proto bird As we can see on machine 2 192.168.0.0/26 is being routed from interface eth0 and with bird proto, so obviously calico must have configured it, which is the bug! it should have configured it on tunl0 interface which was up and running on Machine 2. tunl0 Link encap:IPIP Tunnel HWaddr inet addr:192.168.0.64 Mask:255.255.255.255 UP RUNNING NOARP MTU:1440 Metric:1 RX packets:0 errors:0 dropped:0 overruns:0 frame:0 TX packets:4 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:0 (0.0 B) TX bytes:336 (336.0 B) I tried a few silly things to fix it like removing tunl0 interface and restarting calico node on Machine 2, but none of them helped.So finally i manually deleted the route entry for 192.168.0.0/26 from the route table on Machine2 and bingo! , it learned the new route through tunl0 and the instances were now able to communicate. ip route show on Machine 2: default via 172.31.16.1 dev eth0 10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1 172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77 192.168.0.0/26 via 172.31.27.45 dev tunl0 proto bird onlink 192.168.0.16 dev calie209b94cc7a scope link blackhole 192.168.0.64/26 proto bird So shouldnt the bird have configured proper route through tunl0 initially and not through eth0. If the bird dint do it then shouldnt bird be flushing the wrong entries after creating the new interface so the new routes can be learned?
1.0
Calico Node/Bird routes through default interface(eth0) and not tunnelled interface (tunl0) - How did i notice the issue - docker instance on Machine1 was not able to communicate with docker instance on Machine2 inspite of they being part of same profile and the profile had default settings which is allow all communication within the profile. Background: I was running a few docker instances on Machine 1(192.168.0.12 , 13 and 14) and they were part of same profile so obviously as expected they were able to communicate with each other. Now i configured calico on another machine(Machine 2) and created a new docker instance (192.168.0.15) and adding it to same profile as the above mentioned docker instances, But now the docker instances on Machine1 were not able to reach instances on Machine2 and viceversa. ip route show output from both boxes: **Machine 1: ** default via 172.31.16.1 dev eth0 10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1 172.18.0.0/16 dev docker_gwbridge proto kernel scope link src 172.18.0.1 172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.27.45 blackhole 192.168.0.0/26 proto bird 192.168.0.12 dev calie8beb266c79 scope link 192.168.0.13 dev caliefab0610c79 scope link 192.168.0.14 dev calif2c7a52ec79 scope link 192.168.0.15 via 172.31.29.77 dev tunl0 proto bird onlink 192.168.0.64/26 via 172.31.29.77 dev tunl0 proto bird onlink **Machine 2: ** default via 172.31.16.1 dev eth0 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1 172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77 192.168.0.0/26 via 172.31.27.45 dev eth0 proto bird 192.168.0.15 dev cali7fcbe484c7a scope link blackhole 192.168.0.64/26 proto bird As we can see on machine 2 192.168.0.0/26 is being routed from interface eth0 and with bird proto, so obviously calico must have configured it, which is the bug! it should have configured it on tunl0 interface which was up and running on Machine 2. tunl0 Link encap:IPIP Tunnel HWaddr inet addr:192.168.0.64 Mask:255.255.255.255 UP RUNNING NOARP MTU:1440 Metric:1 RX packets:0 errors:0 dropped:0 overruns:0 frame:0 TX packets:4 errors:0 dropped:0 overruns:0 carrier:0 collisions:0 txqueuelen:0 RX bytes:0 (0.0 B) TX bytes:336 (336.0 B) I tried a few silly things to fix it like removing tunl0 interface and restarting calico node on Machine 2, but none of them helped.So finally i manually deleted the route entry for 192.168.0.0/26 from the route table on Machine2 and bingo! , it learned the new route through tunl0 and the instances were now able to communicate. ip route show on Machine 2: default via 172.31.16.1 dev eth0 10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1 172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77 192.168.0.0/26 via 172.31.27.45 dev tunl0 proto bird onlink 192.168.0.16 dev calie209b94cc7a scope link blackhole 192.168.0.64/26 proto bird So shouldnt the bird have configured proper route through tunl0 initially and not through eth0. If the bird dint do it then shouldnt bird be flushing the wrong entries after creating the new interface so the new routes can be learned?
priority
calico node bird routes through default interface and not tunnelled interface how did i notice the issue docker instance on was not able to communicate with docker instance on inspite of they being part of same profile and the profile had default settings which is allow all communication within the profile background i was running a few docker instances on machine and and they were part of same profile so obviously as expected they were able to communicate with each other now i configured calico on another machine machine and created a new docker instance and adding it to same profile as the above mentioned docker instances but now the docker instances on were not able to reach instances on and viceversa ip route show output from both boxes machine default via dev dev proto kernel scope link src dev proto kernel scope link src dev docker gwbridge proto kernel scope link src dev proto kernel scope link src blackhole proto bird dev scope link dev scope link dev scope link via dev proto bird onlink via dev proto bird onlink machine default via dev dev proto kernel scope link src dev proto kernel scope link src via dev proto bird dev scope link blackhole proto bird as we can see on machine is being routed from interface and with bird proto so obviously calico must have configured it which is the bug it should have configured it on interface which was up and running on machine link encap ipip tunnel hwaddr inet addr mask up running noarp mtu metric rx packets errors dropped overruns frame tx packets errors dropped overruns carrier collisions txqueuelen rx bytes b tx bytes b i tried a few silly things to fix it like removing interface and restarting calico node on machine but none of them helped so finally i manually deleted the route entry for from the route table on and bingo it learned the new route through and the instances were now able to communicate ip route show on machine default via dev dev proto kernel scope link src dev proto kernel scope link src dev proto kernel scope link src via dev proto bird onlink dev scope link blackhole proto bird so shouldnt the bird have configured proper route through initially and not through if the bird dint do it then shouldnt bird be flushing the wrong entries after creating the new interface so the new routes can be learned
1
380,247
11,256,105,734
IssuesEvent
2020-01-12 14:09:37
williamjcm/wxMASSManager
https://api.github.com/repos/williamjcm/wxMASSManager
closed
Code reorganisation
enhancement high priority
When I started the app, I put all code, even the backend stuff, into EvtMainFrame. Before releasing 1.0 of the app, I should put as much backend stuff as possible into a separate class.
1.0
Code reorganisation - When I started the app, I put all code, even the backend stuff, into EvtMainFrame. Before releasing 1.0 of the app, I should put as much backend stuff as possible into a separate class.
priority
code reorganisation when i started the app i put all code even the backend stuff into evtmainframe before releasing of the app i should put as much backend stuff as possible into a separate class
1
175,529
6,551,786,638
IssuesEvent
2017-09-05 15:51:17
opencurrents/opencurrents
https://api.github.com/repos/opencurrents/opencurrents
closed
time-tracker: Display error if admin is not selected or "Not sure"
priority high
Display error: "Please select your admin to continue"
1.0
time-tracker: Display error if admin is not selected or "Not sure" - Display error: "Please select your admin to continue"
priority
time tracker display error if admin is not selected or not sure display error please select your admin to continue
1
141,633
5,439,239,319
IssuesEvent
2017-03-06 12:53:34
Cadasta/cadasta-platform
https://api.github.com/repos/Cadasta/cadasta-platform
closed
Hardcoded S3 paths
bug priority: high security
Hardcoded S3 bucket paths appear across multiple files: https://github.com/Cadasta/cadasta-platform/search?utf8=%E2%9C%93&q=platformprod&type=Code This results in all environments (staging, demo, and production) hitting the production S3 bucket, which violates our isolation requirements. It also exposes the bucket names. We should remove hardcoded references to bucket names/paths and change bucket names ASAP. We should also avoid commingling static assets like platform icons with project data, to allow for stricter security policies.
1.0
Hardcoded S3 paths - Hardcoded S3 bucket paths appear across multiple files: https://github.com/Cadasta/cadasta-platform/search?utf8=%E2%9C%93&q=platformprod&type=Code This results in all environments (staging, demo, and production) hitting the production S3 bucket, which violates our isolation requirements. It also exposes the bucket names. We should remove hardcoded references to bucket names/paths and change bucket names ASAP. We should also avoid commingling static assets like platform icons with project data, to allow for stricter security policies.
priority
hardcoded paths hardcoded bucket paths appear across multiple files this results in all environments staging demo and production hitting the production bucket which violates our isolation requirements it also exposes the bucket names we should remove hardcoded references to bucket names paths and change bucket names asap we should also avoid commingling static assets like platform icons with project data to allow for stricter security policies
1
768,875
26,985,024,425
IssuesEvent
2023-02-09 15:35:16
ballerina-platform/ballerina-standard-library
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
closed
NATS module build is failing more frequently
Points/2 Priority/High Type/Task module/nats Team/PCM
**Description:** NATS build is intermittently failing for quite some time and now is more frequently failing. Trying the approach of moving most of the tests into a separate test module.
1.0
NATS module build is failing more frequently - **Description:** NATS build is intermittently failing for quite some time and now is more frequently failing. Trying the approach of moving most of the tests into a separate test module.
priority
nats module build is failing more frequently description nats build is intermittently failing for quite some time and now is more frequently failing trying the approach of moving most of the tests into a separate test module
1
252,080
8,031,547,601
IssuesEvent
2018-07-28 03:04:17
crslab/cloud-vision-explorer
https://api.github.com/repos/crslab/cloud-vision-explorer
closed
[HIGH] Update build webpack build process to build quicker
enhancement high priority
Build process takes forever right now. Rewrite the webpack config, remove gulp from the process, and build for dev not for prod. Hopefully this will make things faster.
1.0
[HIGH] Update build webpack build process to build quicker - Build process takes forever right now. Rewrite the webpack config, remove gulp from the process, and build for dev not for prod. Hopefully this will make things faster.
priority
update build webpack build process to build quicker build process takes forever right now rewrite the webpack config remove gulp from the process and build for dev not for prod hopefully this will make things faster
1
47,327
2,974,767,124
IssuesEvent
2015-07-15 04:16:19
OpenSourceMalaria/OSM_To_Do_List
https://api.github.com/repos/OpenSourceMalaria/OSM_To_Do_List
reopened
Synthesis of 6-chloropyrazin-2-amine
Being Synthesised Now High Priority Series 4 Synthetic Chemistry Needed
![tz 1-1 target](https://cloud.githubusercontent.com/assets/13214148/8585610/a466cda2-2627-11e5-8344-0c9106694461.png) InChI=1S/C4H4ClN3/c5-3-1-7-2-4(6)8-3/h1-2H,(H2,6,8) ClC1=CN=CC(N)=N1 Hi, I am Tianyi. I am a student on exchange from Nanjing University and I am working with Alice on the lab. We try to make 6-chloropyrazin-2-amine under the same conditions used to make 2-Chloro-6-hydrazinylpyrazine (AEW 85-7), but replacing hydrazine hydrate with ammonia (http://malaria.ourexperiment.org/uri/810). The TLC so far suggests only starting material but we will confirm today. I did literature search and found a patent describing synthesis of 6-chloropyrazin-2-amine using the same starting material but with aqueous solvent at 100 degrees in a sealed tube. We will try that today.http://worldwide.espacenet.com/publicationDetails/originalDocument?CC=WO&NR=2013068755A1&KC=A1&FT=D&ND=&date=20130516&DB=&&locale=en_EP Ref https://github.com/tyzhengsydney
1.0
Synthesis of 6-chloropyrazin-2-amine - ![tz 1-1 target](https://cloud.githubusercontent.com/assets/13214148/8585610/a466cda2-2627-11e5-8344-0c9106694461.png) InChI=1S/C4H4ClN3/c5-3-1-7-2-4(6)8-3/h1-2H,(H2,6,8) ClC1=CN=CC(N)=N1 Hi, I am Tianyi. I am a student on exchange from Nanjing University and I am working with Alice on the lab. We try to make 6-chloropyrazin-2-amine under the same conditions used to make 2-Chloro-6-hydrazinylpyrazine (AEW 85-7), but replacing hydrazine hydrate with ammonia (http://malaria.ourexperiment.org/uri/810). The TLC so far suggests only starting material but we will confirm today. I did literature search and found a patent describing synthesis of 6-chloropyrazin-2-amine using the same starting material but with aqueous solvent at 100 degrees in a sealed tube. We will try that today.http://worldwide.espacenet.com/publicationDetails/originalDocument?CC=WO&NR=2013068755A1&KC=A1&FT=D&ND=&date=20130516&DB=&&locale=en_EP Ref https://github.com/tyzhengsydney
priority
synthesis of chloropyrazin amine inchi cn cc n hi i am tianyi i am a student on exchange from nanjing university and i am working with alice on the lab we try to make chloropyrazin amine under the same conditions used to make chloro hydrazinylpyrazine aew but replacing hydrazine hydrate with ammonia the tlc so far suggests only starting material but we will confirm today i did literature search and found a patent describing synthesis of chloropyrazin amine using the same starting material but with aqueous solvent at degrees in a sealed tube we will try that today ref
1
684,831
23,434,381,534
IssuesEvent
2022-08-15 08:08:28
BenjamooseCalto/SlasherBot
https://api.github.com/repos/BenjamooseCalto/SlasherBot
opened
BUG: The Steam game update check occasionally hangs, causing the check to fail.
bug high priority
This is sort of high prio, prevents the game update notifications from working.
1.0
BUG: The Steam game update check occasionally hangs, causing the check to fail. - This is sort of high prio, prevents the game update notifications from working.
priority
bug the steam game update check occasionally hangs causing the check to fail this is sort of high prio prevents the game update notifications from working
1
281,686
8,698,241,283
IssuesEvent
2018-12-04 22:42:35
broadinstitute/cromwell
https://api.github.com/repos/broadinstitute/cromwell
closed
Fix flakey test: KeyValueServiceActorSpec
HIGH PRIORITY QA Triage
https://broadinstitute.atlassian.net/browse/GAWB-3950 https://fc-jenkins.dsp-techops.broadinstitute.org/view/Testing/view/Test%20Runners/job/cromwell-test-runner/790/ https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/833/ https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/804/ tc: KeyValueServiceActor should insert a key/value tc: KeyValueServiceActor should return error if key doesn't exist tc: KeyValueServiceActor should be able to overwrite values {quote} org.scalatest.exceptions.TestFailedException: The future returned an exception of type: akka.pattern.AskTimeoutException, with message: Ask timed out on [Actor[akka://KeyValueServiceActorSpec/user/$a#-1019375090]] after [200000 ms]. Sender[null] sent message of type "cromwell.services.keyvalue.KeyValueServiceActor$KvPut".. at org.scalatest.concurrent.Futures$FutureConcept.tryTryAgain$1(Futures.scala:531) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl(Futures.scala:550) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl$(Futures.scala:479) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValueImpl(ScalaFutures.scala:275) at org.scalatest.concurrent.Futures$FutureConcept.futureValue(Futures.scala:476) at org.scalatest.concurrent.Futures$FutureConcept.futureValue$(Futures.scala:475) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValue(ScalaFutures.scala:275) at cromwell.services.keyvalue.impl.KeyValueServiceActorSpec.$anonfun$new$2(KeyValueServiceActorSpec.scala:46) at {quote}
1.0
Fix flakey test: KeyValueServiceActorSpec - https://broadinstitute.atlassian.net/browse/GAWB-3950 https://fc-jenkins.dsp-techops.broadinstitute.org/view/Testing/view/Test%20Runners/job/cromwell-test-runner/790/ https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/833/ https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/804/ tc: KeyValueServiceActor should insert a key/value tc: KeyValueServiceActor should return error if key doesn't exist tc: KeyValueServiceActor should be able to overwrite values {quote} org.scalatest.exceptions.TestFailedException: The future returned an exception of type: akka.pattern.AskTimeoutException, with message: Ask timed out on [Actor[akka://KeyValueServiceActorSpec/user/$a#-1019375090]] after [200000 ms]. Sender[null] sent message of type "cromwell.services.keyvalue.KeyValueServiceActor$KvPut".. at org.scalatest.concurrent.Futures$FutureConcept.tryTryAgain$1(Futures.scala:531) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl(Futures.scala:550) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl$(Futures.scala:479) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValueImpl(ScalaFutures.scala:275) at org.scalatest.concurrent.Futures$FutureConcept.futureValue(Futures.scala:476) at org.scalatest.concurrent.Futures$FutureConcept.futureValue$(Futures.scala:475) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValue(ScalaFutures.scala:275) at cromwell.services.keyvalue.impl.KeyValueServiceActorSpec.$anonfun$new$2(KeyValueServiceActorSpec.scala:46) at {quote}
priority
fix flakey test keyvalueserviceactorspec tc keyvalueserviceactor should insert a key value tc keyvalueserviceactor should return error if key doesn t exist tc keyvalueserviceactor should be able to overwrite values quote org scalatest exceptions testfailedexception the future returned an exception of type akka pattern asktimeoutexception with message ask timed out on after sender sent message of type cromwell services keyvalue keyvalueserviceactor kvput at org scalatest concurrent futures futureconcept trytryagain futures scala at org scalatest concurrent futures futureconcept futurevalueimpl futures scala at org scalatest concurrent futures futureconcept futurevalueimpl futures scala at org scalatest concurrent scalafutures anon futurevalueimpl scalafutures scala at org scalatest concurrent futures futureconcept futurevalue futures scala at org scalatest concurrent futures futureconcept futurevalue futures scala at org scalatest concurrent scalafutures anon futurevalue scalafutures scala at cromwell services keyvalue impl keyvalueserviceactorspec anonfun new keyvalueserviceactorspec scala at quote
1
687,598
23,532,999,058
IssuesEvent
2022-08-19 17:17:20
canonical/cos-lite-bundle
https://api.github.com/repos/canonical/cos-lite-bundle
closed
Integrate Loki and Alertmanager
Type: Enhancement Priority: High
Loki has an integration with Alertmanager that is currently not leveraged by the bundle. Besides the one-liner to add the missing relation, we need integration tests where Loki-generated alerts are ported through Alertmanager.
1.0
Integrate Loki and Alertmanager - Loki has an integration with Alertmanager that is currently not leveraged by the bundle. Besides the one-liner to add the missing relation, we need integration tests where Loki-generated alerts are ported through Alertmanager.
priority
integrate loki and alertmanager loki has an integration with alertmanager that is currently not leveraged by the bundle besides the one liner to add the missing relation we need integration tests where loki generated alerts are ported through alertmanager
1
740,308
25,743,675,746
IssuesEvent
2022-12-08 08:20:08
HackerN64/HackerSM64
https://api.github.com/repos/HackerN64/HackerSM64
opened
Split `segment2.c` into multiple files
high priority
The file contains way too many assets that could be split into multiple different categorized files. ASCII and paintings need to be merged before this split can happen.
1.0
Split `segment2.c` into multiple files - The file contains way too many assets that could be split into multiple different categorized files. ASCII and paintings need to be merged before this split can happen.
priority
split c into multiple files the file contains way too many assets that could be split into multiple different categorized files ascii and paintings need to be merged before this split can happen
1
781,006
27,418,288,044
IssuesEvent
2023-03-01 15:07:47
jj248/RealmsInExile
https://api.github.com/repos/jj248/RealmsInExile
closed
Evil Isengard event chain rework
localization lore priority: high events scripting
Revamp event chain to be more robust and to have multiple branching paths. Note that good Saruman altlore will be covered in a separate issue.
1.0
Evil Isengard event chain rework - Revamp event chain to be more robust and to have multiple branching paths. Note that good Saruman altlore will be covered in a separate issue.
priority
evil isengard event chain rework revamp event chain to be more robust and to have multiple branching paths note that good saruman altlore will be covered in a separate issue
1
260,170
8,204,713,571
IssuesEvent
2018-09-03 07:43:47
kowala-tech/kcoin
https://api.github.com/repos/kowala-tech/kcoin
closed
Review minimum validator stake
High priority blocker bootes-launch question research
The minimum stake from the whitepaper is 30,000 mTokens. The amount will likely increase over time naturally as the registry (with its default of 500 slots) fills up. What should the hard minimum be? Consideration factors: - The stake should represent a significant financial interest for anyone, such that it can't be trivially thrown in an attack. - All kUSD backers and other interested parties should be able to afford to be validators if they want to be. They shouldn't be locked out of validation over time. - The stake shouldn't be so low that a handful of substantial token holders could occupy all the available slots by splitting their tokens and making many validators. Possible solutions: - Increase the number of validation slots. Since the network will slow down with a large number of validators, they can't all be active at the same time. This solution involves somehow changing the validation and consensus systems so that an unpredictable subgroup of all the registered validators in get to participate in the validation at any time. This will constitute a major change and require the delivery of a new consensus protocol on a running network. - Introduce delegation into validation. That is, allow existing validators to accept tokens staked on their behalf in exchange for some pre-agreed, contractual validation revenue. This requires a much smaller change and no protocol updates.
1.0
Review minimum validator stake - The minimum stake from the whitepaper is 30,000 mTokens. The amount will likely increase over time naturally as the registry (with its default of 500 slots) fills up. What should the hard minimum be? Consideration factors: - The stake should represent a significant financial interest for anyone, such that it can't be trivially thrown in an attack. - All kUSD backers and other interested parties should be able to afford to be validators if they want to be. They shouldn't be locked out of validation over time. - The stake shouldn't be so low that a handful of substantial token holders could occupy all the available slots by splitting their tokens and making many validators. Possible solutions: - Increase the number of validation slots. Since the network will slow down with a large number of validators, they can't all be active at the same time. This solution involves somehow changing the validation and consensus systems so that an unpredictable subgroup of all the registered validators in get to participate in the validation at any time. This will constitute a major change and require the delivery of a new consensus protocol on a running network. - Introduce delegation into validation. That is, allow existing validators to accept tokens staked on their behalf in exchange for some pre-agreed, contractual validation revenue. This requires a much smaller change and no protocol updates.
priority
review minimum validator stake the minimum stake from the whitepaper is mtokens the amount will likely increase over time naturally as the registry with its default of slots fills up what should the hard minimum be consideration factors the stake should represent a significant financial interest for anyone such that it can t be trivially thrown in an attack all kusd backers and other interested parties should be able to afford to be validators if they want to be they shouldn t be locked out of validation over time the stake shouldn t be so low that a handful of substantial token holders could occupy all the available slots by splitting their tokens and making many validators possible solutions increase the number of validation slots since the network will slow down with a large number of validators they can t all be active at the same time this solution involves somehow changing the validation and consensus systems so that an unpredictable subgroup of all the registered validators in get to participate in the validation at any time this will constitute a major change and require the delivery of a new consensus protocol on a running network introduce delegation into validation that is allow existing validators to accept tokens staked on their behalf in exchange for some pre agreed contractual validation revenue this requires a much smaller change and no protocol updates
1
318,498
9,693,388,643
IssuesEvent
2019-05-24 15:57:12
pilosus/pilosus-pelican-theme
https://api.github.com/repos/pilosus/pilosus-pelican-theme
closed
Pagination for blog entries
bug priority: high
Pagination does work, but a template doesn't render url for pages: [prev] [2] [next]
1.0
Pagination for blog entries - Pagination does work, but a template doesn't render url for pages: [prev] [2] [next]
priority
pagination for blog entries pagination does work but a template doesn t render url for pages
1
546,764
16,018,846,397
IssuesEvent
2021-04-20 19:40:35
ORNL-AMO/VERIFI
https://api.github.com/repos/ORNL-AMO/VERIFI
closed
Update Examples
Data Input High Priority
Need more generic names for the example data. What we should do is create the example accounts within verifi, including facilities/meters/data etc. Then you can export the accounts and send them to me to use for the examples.
1.0
Update Examples - Need more generic names for the example data. What we should do is create the example accounts within verifi, including facilities/meters/data etc. Then you can export the accounts and send them to me to use for the examples.
priority
update examples need more generic names for the example data what we should do is create the example accounts within verifi including facilities meters data etc then you can export the accounts and send them to me to use for the examples
1
95,924
3,962,285,585
IssuesEvent
2016-05-02 16:14:01
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
"SQL Error : Incorrect syntax near 'LIMIT'" when using MS SQL Server 2014
bug High Priority
"SQL Error : Incorrect syntax near 'LIMIT'. " is thrown when viewing existing report output using SuiteCRM 7.5.1, connecting to MS SQL Server 2014. Log file: "SELECT id FROM aor_fields WHERE aor_report_id = N'981b609d-47a9-9efd-c193-56bdc9373c10' AND deleted = 0 ORDER BY field_order ASC: SQL Server error: Incorrect syntax near 'LIMIT'" The LIMIT clause is not supported by MS SQL & the appropriate connector/query needs to be updated
1.0
"SQL Error : Incorrect syntax near 'LIMIT'" when using MS SQL Server 2014 - "SQL Error : Incorrect syntax near 'LIMIT'. " is thrown when viewing existing report output using SuiteCRM 7.5.1, connecting to MS SQL Server 2014. Log file: "SELECT id FROM aor_fields WHERE aor_report_id = N'981b609d-47a9-9efd-c193-56bdc9373c10' AND deleted = 0 ORDER BY field_order ASC: SQL Server error: Incorrect syntax near 'LIMIT'" The LIMIT clause is not supported by MS SQL & the appropriate connector/query needs to be updated
priority
sql error incorrect syntax near limit when using ms sql server sql error incorrect syntax near limit is thrown when viewing existing report output using suitecrm connecting to ms sql server log file select id from aor fields where aor report id n and deleted order by field order asc sql server error incorrect syntax near limit the limit clause is not supported by ms sql the appropriate connector query needs to be updated
1
273,461
8,530,931,749
IssuesEvent
2018-11-04 05:34:31
cyberperspectives/sagacity
https://api.github.com/repos/cyberperspectives/sagacity
opened
Ingesting eChecklists results in duplicate findings in DB
High Priority bug
I load SCC results (and can do so repeatedly) and it works fine, but ingesting eChecklists results in duplicate findings in the DB. As a result, I can't export the final .ckl files. I think there is a problem with the get_Finding() routine in database.inc that doesn't return the info when there is a finding in the database. parse_excel_echecklist.php never goes into the updated_findings routine -only the new_findings section. I put an error_log() at the end of get_Finding() and it doesn't seem to return anything. This is urgent. We are testing this week and will need to export .ckl files at the end of the week. If we can get a patch we will be able to finish the assessment.
1.0
Ingesting eChecklists results in duplicate findings in DB - I load SCC results (and can do so repeatedly) and it works fine, but ingesting eChecklists results in duplicate findings in the DB. As a result, I can't export the final .ckl files. I think there is a problem with the get_Finding() routine in database.inc that doesn't return the info when there is a finding in the database. parse_excel_echecklist.php never goes into the updated_findings routine -only the new_findings section. I put an error_log() at the end of get_Finding() and it doesn't seem to return anything. This is urgent. We are testing this week and will need to export .ckl files at the end of the week. If we can get a patch we will be able to finish the assessment.
priority
ingesting echecklists results in duplicate findings in db i load scc results and can do so repeatedly and it works fine but ingesting echecklists results in duplicate findings in the db as a result i can t export the final ckl files i think there is a problem with the get finding routine in database inc that doesn t return the info when there is a finding in the database parse excel echecklist php never goes into the updated findings routine only the new findings section i put an error log at the end of get finding and it doesn t seem to return anything this is urgent we are testing this week and will need to export ckl files at the end of the week if we can get a patch we will be able to finish the assessment
1
780,835
27,410,318,348
IssuesEvent
2023-03-01 10:04:49
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Remove adding unnecessary any cast when creating tuple values #32891
Type/Improvement Priority/High Team/CompilerFE
**Description:** Every time we create a tuple value, we have added to cast to any. This is not needed for primitive types. This has 10% when we create tuple arrays.
1.0
Remove adding unnecessary any cast when creating tuple values #32891 - **Description:** Every time we create a tuple value, we have added to cast to any. This is not needed for primitive types. This has 10% when we create tuple arrays.
priority
remove adding unnecessary any cast when creating tuple values description every time we create a tuple value we have added to cast to any this is not needed for primitive types this has when we create tuple arrays
1
344,274
10,342,289,457
IssuesEvent
2019-09-04 05:56:04
redeclipse/base
https://api.github.com/repos/redeclipse/base
closed
[mac os][bug] Couldn't launch the game with last update
branch: stable difficulty: easy priority: high status: bug
MacOS Mojave 14.6 I don't know what libsteam license is, but I thought you can link it statically. Console output: ``` bins: 2e9ede6132ca3ee7748c6919e7be15db6b638924 is the current version. dyld: Library not loaded: @rpath/libsteam_api.dylib Referenced from: /Applications/redeclipse.app/Contents/Resources/bin/redeclipse.app/Contents/MacOS/redeclipse_universal Reason: image not found Abort trap: 6 logout ...bla-bla-bla ``` - Version and build: [2e9ede6132ca3ee7748c6919e7be15db6b638924]
1.0
[mac os][bug] Couldn't launch the game with last update - MacOS Mojave 14.6 I don't know what libsteam license is, but I thought you can link it statically. Console output: ``` bins: 2e9ede6132ca3ee7748c6919e7be15db6b638924 is the current version. dyld: Library not loaded: @rpath/libsteam_api.dylib Referenced from: /Applications/redeclipse.app/Contents/Resources/bin/redeclipse.app/Contents/MacOS/redeclipse_universal Reason: image not found Abort trap: 6 logout ...bla-bla-bla ``` - Version and build: [2e9ede6132ca3ee7748c6919e7be15db6b638924]
priority
couldn t launch the game with last update macos mojave i don t know what libsteam license is but i thought you can link it statically console output bins is the current version dyld library not loaded rpath libsteam api dylib referenced from applications redeclipse app contents resources bin redeclipse app contents macos redeclipse universal reason image not found abort trap logout bla bla bla version and build
1
463,185
13,261,427,118
IssuesEvent
2020-08-20 19:52:46
ChainSafe/forest
https://api.github.com/repos/ChainSafe/forest
closed
Change to using sha2-256 hashing instead of blake2b
IPLD Priority: 2 - High Spec Change Status: On Ice
This is a change to the spec for performance reasons.
1.0
Change to using sha2-256 hashing instead of blake2b - This is a change to the spec for performance reasons.
priority
change to using hashing instead of this is a change to the spec for performance reasons
1
546,659
16,016,965,985
IssuesEvent
2021-04-20 17:13:31
ESCOMP/CTSM
https://api.github.com/repos/ESCOMP/CTSM
closed
Add landunit_mask (formerly l2g_scale_type) metadata to history file
priority: high type: enhancement
Connected with #1297 - We should add the landunit_mask attribute (currently named l2g_scale_type, but intended to be renamed in #1342) to each field in the history file. Since #1342 suggests changing strings to integer constants, we'll need to have a translation from these integer constants back to human-readable strings when writing out the history metadata. I'm not sure what to do for fields that don't have a l2g_scale_type / landunit_mask attribute. Long-term, once we have made everything consistent, I think we probably want this attribute on all history fields (it will just be something like 'all' for many fields). But for now, that would be misleading, because many fields currently do masking via the initial setting to spval everywhere. So probably for now we should only add this attribute to fields for which l2g_scale_type / landunit_mask is explicitly defined.
1.0
Add landunit_mask (formerly l2g_scale_type) metadata to history file - Connected with #1297 - We should add the landunit_mask attribute (currently named l2g_scale_type, but intended to be renamed in #1342) to each field in the history file. Since #1342 suggests changing strings to integer constants, we'll need to have a translation from these integer constants back to human-readable strings when writing out the history metadata. I'm not sure what to do for fields that don't have a l2g_scale_type / landunit_mask attribute. Long-term, once we have made everything consistent, I think we probably want this attribute on all history fields (it will just be something like 'all' for many fields). But for now, that would be misleading, because many fields currently do masking via the initial setting to spval everywhere. So probably for now we should only add this attribute to fields for which l2g_scale_type / landunit_mask is explicitly defined.
priority
add landunit mask formerly scale type metadata to history file connected with we should add the landunit mask attribute currently named scale type but intended to be renamed in to each field in the history file since suggests changing strings to integer constants we ll need to have a translation from these integer constants back to human readable strings when writing out the history metadata i m not sure what to do for fields that don t have a scale type landunit mask attribute long term once we have made everything consistent i think we probably want this attribute on all history fields it will just be something like all for many fields but for now that would be misleading because many fields currently do masking via the initial setting to spval everywhere so probably for now we should only add this attribute to fields for which scale type landunit mask is explicitly defined
1
290,451
8,895,386,565
IssuesEvent
2019-01-16 08:30:37
gluster/glusterd2
https://api.github.com/repos/gluster/glusterd2
closed
Volume creation is failing after 59 volumes with 30 node cluster
FW: Volume Management GCS/1.0-Blocker priority: high
-> Created 30 node cluster -> Tried to create 200 volumes through script -> It's distributed-replica volume 10x3 (means each volume will be having 30 bricks ) -> One brick from each node that means 30 bricks form 30 nodes -> Volume creation failing with below error Volume creation failed Error: Request failed with HTTP Status code 500 Transaction step vol-create.StoreVolume failed on peer 5a011251-efb4-4efe-b9dc-b25b345c6634 with error: etcdserver: request is too large i didn't find any logs for volume creation failed After that I have tried to create single brick volume manually without script on 30 node cluster setup. actually volume creation is successful, but got error message like below ./glustercli volume create vol1 10.70.42.199:/bricks/brick0/vol1 error message: Volume creation failed Error: Post http://127.0.0.1:24007/v1/volumes: net/http: request canceled (Client.Timeout exceeded while awaiting headers) Then tried to start the volume, failed with below error /glustercli volume start vol1 volume start failed Error: Post http://127.0.0.1:24007/v1/volumes/vol1/start: net/http: request canceled (Client.Timeout exceeded while awaiting headers) Again tried to start the volume, this time volume started successfully ./glustercli volume start vol1 Volume vol1 started successfully
1.0
Volume creation is failing after 59 volumes with 30 node cluster - -> Created 30 node cluster -> Tried to create 200 volumes through script -> It's distributed-replica volume 10x3 (means each volume will be having 30 bricks ) -> One brick from each node that means 30 bricks form 30 nodes -> Volume creation failing with below error Volume creation failed Error: Request failed with HTTP Status code 500 Transaction step vol-create.StoreVolume failed on peer 5a011251-efb4-4efe-b9dc-b25b345c6634 with error: etcdserver: request is too large i didn't find any logs for volume creation failed After that I have tried to create single brick volume manually without script on 30 node cluster setup. actually volume creation is successful, but got error message like below ./glustercli volume create vol1 10.70.42.199:/bricks/brick0/vol1 error message: Volume creation failed Error: Post http://127.0.0.1:24007/v1/volumes: net/http: request canceled (Client.Timeout exceeded while awaiting headers) Then tried to start the volume, failed with below error /glustercli volume start vol1 volume start failed Error: Post http://127.0.0.1:24007/v1/volumes/vol1/start: net/http: request canceled (Client.Timeout exceeded while awaiting headers) Again tried to start the volume, this time volume started successfully ./glustercli volume start vol1 Volume vol1 started successfully
priority
volume creation is failing after volumes with node cluster created node cluster tried to create volumes through script it s distributed replica volume means each volume will be having bricks one brick from each node that means bricks form nodes volume creation failing with below error volume creation failed error request failed with http status code transaction step vol create storevolume failed on peer with error etcdserver request is too large i didn t find any logs for volume creation failed after that i have tried to create single brick volume manually without script on node cluster setup actually volume creation is successful but got error message like below glustercli volume create bricks error message volume creation failed error post net http request canceled client timeout exceeded while awaiting headers then tried to start the volume failed with below error glustercli volume start volume start failed error post net http request canceled client timeout exceeded while awaiting headers again tried to start the volume this time volume started successfully glustercli volume start volume started successfully
1
203,708
7,077,876,747
IssuesEvent
2018-01-10 00:11:21
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Provide EndpointRequest for configuring WebFlux-based Security
priority: high theme: security type: enhancement
Using Spring Boot, SpringMVC and Spring Security I can configure the security part of my Spring Boot app as follows: ``` ...requestMatchers(EndpointRequest.to("status", "info"))... ``` Now, I'm migrating from SpringMVC to Spring WebFlux. I see that I've to use `pathMatchers()` instead of `requestMatchers` from Spring Security. However, `EndpointRequest` is depending on `HttpServletRequest` (and thus on Spring MVC). In https://stackoverflow.com/questions/47287312/spring-security-with-webflux-how-to-migrate-requestmatchersendpointrequest-to#comment-81531115 @bclozel suggests to open an issue.
1.0
Provide EndpointRequest for configuring WebFlux-based Security - Using Spring Boot, SpringMVC and Spring Security I can configure the security part of my Spring Boot app as follows: ``` ...requestMatchers(EndpointRequest.to("status", "info"))... ``` Now, I'm migrating from SpringMVC to Spring WebFlux. I see that I've to use `pathMatchers()` instead of `requestMatchers` from Spring Security. However, `EndpointRequest` is depending on `HttpServletRequest` (and thus on Spring MVC). In https://stackoverflow.com/questions/47287312/spring-security-with-webflux-how-to-migrate-requestmatchersendpointrequest-to#comment-81531115 @bclozel suggests to open an issue.
priority
provide endpointrequest for configuring webflux based security using spring boot springmvc and spring security i can configure the security part of my spring boot app as follows requestmatchers endpointrequest to status info now i m migrating from springmvc to spring webflux i see that i ve to use pathmatchers instead of requestmatchers from spring security however endpointrequest is depending on httpservletrequest and thus on spring mvc in bclozel suggests to open an issue
1
455,270
13,123,276,703
IssuesEvent
2020-08-06 00:03:02
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
opened
[amp-story-player] PE: only preload distance = 1 iframes
P1: High Priority Type: Feature Request WG: stories
Today the Player always preloads 3 stories, when it should only preload N - 1, N, N + 1. This means one less story to load for the initial render. cc @ampproject/wg-stories
1.0
[amp-story-player] PE: only preload distance = 1 iframes - Today the Player always preloads 3 stories, when it should only preload N - 1, N, N + 1. This means one less story to load for the initial render. cc @ampproject/wg-stories
priority
pe only preload distance iframes today the player always preloads stories when it should only preload n n n this means one less story to load for the initial render cc ampproject wg stories
1
456,299
13,148,898,988
IssuesEvent
2020-08-09 00:59:07
HumanExposure/factotum
https://api.github.com/repos/HumanExposure/factotum
opened
Security Issue with Pillow
Priority :: High To point security
Out of Bounds read in Pillow. Upgrade Pillow to version 7.1.0 or later. For example: `Pillow>=7.1.0`
1.0
Security Issue with Pillow - Out of Bounds read in Pillow. Upgrade Pillow to version 7.1.0 or later. For example: `Pillow>=7.1.0`
priority
security issue with pillow out of bounds read in pillow upgrade pillow to version or later for example pillow
1
369,807
10,918,207,852
IssuesEvent
2019-11-21 16:28:41
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
reopened
AVT 1 - Code Snippet -> Inline has a DAP violation
package: react priority: high type: a11y ♿
## Environment macOS Mojave version 10.14.6 Chrome Version Version 77.0.3865.90 Carbon v10 - React DAP IBM Accessibility 1.8.0.0 Sept 2019 Ruleset ## Detailed Description Go to the React Component Run DAP and the following violations are present (see screenshot): <img width="914" alt="Screen Shot 2019-11-01 at 12 20 22 PM" src="https://user-images.githubusercontent.com/21676914/68042886-02991880-fca2-11e9-910a-943cb256dd6f.png"> Note: "node -v" which is the button text does not match the aria-label="copyable code snippet".
1.0
AVT 1 - Code Snippet -> Inline has a DAP violation - ## Environment macOS Mojave version 10.14.6 Chrome Version Version 77.0.3865.90 Carbon v10 - React DAP IBM Accessibility 1.8.0.0 Sept 2019 Ruleset ## Detailed Description Go to the React Component Run DAP and the following violations are present (see screenshot): <img width="914" alt="Screen Shot 2019-11-01 at 12 20 22 PM" src="https://user-images.githubusercontent.com/21676914/68042886-02991880-fca2-11e9-910a-943cb256dd6f.png"> Note: "node -v" which is the button text does not match the aria-label="copyable code snippet".
priority
avt code snippet inline has a dap violation environment macos mojave version chrome version version carbon react dap ibm accessibility sept ruleset detailed description go to the react component run dap and the following violations are present see screenshot img width alt screen shot at pm src note node v which is the button text does not match the aria label copyable code snippet
1
719,957
24,774,463,145
IssuesEvent
2022-10-23 15:01:29
signum-network/signum-node
https://api.github.com/repos/signum-network/signum-node
closed
New API call getTradeJournal
enhancement priority-high API
Currently we have single API calls to request orders or trades for account(s) and token(s). This is technical okay, but those calls are not fitting into the nedd for a DeFi portal. To handle the requirement we propose to create a new API call : `getTradeJournal` Parameter to enter : - Account-ID - Token-ID The call is always for a given account + token. The JSON output looks like this: ` { account: "ID", token:"ID", tradeJournal: [ order-id:"tx-id", timestamp:"tx-timestamp", status: integer, tradeType:"buy or sell" limitNQT: "limit from tx", quantityQNT:"quantity from tx", averagePriceOfOneQNT: float, executedAmountQNT:float, executionSlides: [ trade-id:"trade-id", timestamp:"trade-tx", height:integer, quantityQNT:"quantity from trade", priceOfOneQNT:"price from trade" ], ] } ` **Description** - account: The account from the request - token: The token from the request - `tradeJournal`: The section with the list of open/closed and canceled orders for the given account/token - order-id: The ID from the transaction which creates the buy or sell order - timestamp: The timestamp from the transaction which creates the buy or sell order - status : A integer value with the following range: - (-1) order is caneled - (0) order is open - (1 until 99) partial fill ofd the order in % - (100) order is closed ( 100% filled) - tradeType: Direction of the order (buy/sell) - limitNQT: The limit set on the transaction in Planck ( converted) - quantityQNT: The amount of tokens to buy or sell from the original transaction - averagePriceOfOneQNT : The calcualted execution price by all `executionSlides` - executedAmountQNT: The sum of all amount by the `executionSlides` - `trades` : In this section we show all trades related to the order-id - trade-id : ID of the trade - timestamp: Timestamp of the trade - height : Block-ID - quantityQNT : Executed amount - priceOfOneQNT: Excecuted price for the quantityQNT **Logic** We should loop over all transaction from the account for Place BID/ASK orders filtered by asset(token) In this case we have the main list of all orders from the user. Now we loop over all trades by the order-ids. We filter on asset-id (token) and check ask_order_id/bid_order_id. Finally we check all Cancel transaction and set those order-ids to status -1 The NQT fields should be expressed in Planck ( see #654 ) This ticket covers the following requirements: - #657 - #656 - #655
1.0
New API call getTradeJournal - Currently we have single API calls to request orders or trades for account(s) and token(s). This is technical okay, but those calls are not fitting into the nedd for a DeFi portal. To handle the requirement we propose to create a new API call : `getTradeJournal` Parameter to enter : - Account-ID - Token-ID The call is always for a given account + token. The JSON output looks like this: ` { account: "ID", token:"ID", tradeJournal: [ order-id:"tx-id", timestamp:"tx-timestamp", status: integer, tradeType:"buy or sell" limitNQT: "limit from tx", quantityQNT:"quantity from tx", averagePriceOfOneQNT: float, executedAmountQNT:float, executionSlides: [ trade-id:"trade-id", timestamp:"trade-tx", height:integer, quantityQNT:"quantity from trade", priceOfOneQNT:"price from trade" ], ] } ` **Description** - account: The account from the request - token: The token from the request - `tradeJournal`: The section with the list of open/closed and canceled orders for the given account/token - order-id: The ID from the transaction which creates the buy or sell order - timestamp: The timestamp from the transaction which creates the buy or sell order - status : A integer value with the following range: - (-1) order is caneled - (0) order is open - (1 until 99) partial fill ofd the order in % - (100) order is closed ( 100% filled) - tradeType: Direction of the order (buy/sell) - limitNQT: The limit set on the transaction in Planck ( converted) - quantityQNT: The amount of tokens to buy or sell from the original transaction - averagePriceOfOneQNT : The calcualted execution price by all `executionSlides` - executedAmountQNT: The sum of all amount by the `executionSlides` - `trades` : In this section we show all trades related to the order-id - trade-id : ID of the trade - timestamp: Timestamp of the trade - height : Block-ID - quantityQNT : Executed amount - priceOfOneQNT: Excecuted price for the quantityQNT **Logic** We should loop over all transaction from the account for Place BID/ASK orders filtered by asset(token) In this case we have the main list of all orders from the user. Now we loop over all trades by the order-ids. We filter on asset-id (token) and check ask_order_id/bid_order_id. Finally we check all Cancel transaction and set those order-ids to status -1 The NQT fields should be expressed in Planck ( see #654 ) This ticket covers the following requirements: - #657 - #656 - #655
priority
new api call gettradejournal currently we have single api calls to request orders or trades for account s and token s this is technical okay but those calls are not fitting into the nedd for a defi portal to handle the requirement we propose to create a new api call gettradejournal parameter to enter account id token id the call is always for a given account token the json output looks like this account id token id tradejournal order id tx id timestamp tx timestamp status integer tradetype buy or sell limitnqt limit from tx quantityqnt quantity from tx averagepriceofoneqnt float executedamountqnt float executionslides trade id trade id timestamp trade tx height integer quantityqnt quantity from trade priceofoneqnt price from trade description account the account from the request token the token from the request tradejournal the section with the list of open closed and canceled orders for the given account token order id the id from the transaction which creates the buy or sell order timestamp the timestamp from the transaction which creates the buy or sell order status a integer value with the following range order is caneled order is open until partial fill ofd the order in order is closed filled tradetype direction of the order buy sell limitnqt the limit set on the transaction in planck converted quantityqnt the amount of tokens to buy or sell from the original transaction averagepriceofoneqnt the calcualted execution price by all executionslides executedamountqnt the sum of all amount by the executionslides trades in this section we show all trades related to the order id trade id id of the trade timestamp timestamp of the trade height block id quantityqnt executed amount priceofoneqnt excecuted price for the quantityqnt logic we should loop over all transaction from the account for place bid ask orders filtered by asset token in this case we have the main list of all orders from the user now we loop over all trades by the order ids we filter on asset id token and check ask order id bid order id finally we check all cancel transaction and set those order ids to status the nqt fields should be expressed in planck see this ticket covers the following requirements
1
577,163
17,104,515,216
IssuesEvent
2021-07-09 15:40:13
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Filter and Aggregate by Inherited metadata (30pts)
Priority: High
Expected features: - Filter by inherited metadata instead of the title of the inherited entity - Show appropriate aggregations for each value. - Inherited filters should match with common filters. Implementation decisions - ~Template properties will have type (text, relationship...) and filterType (relationships will use inherited type)~ - ~The inheritedValue will be moved from inheritedValue to value~ - ~The value will be moved to ref {value: label:} or ref / refLabel~ - ~The label should be replaced with the inherited property label~ - Do not allow inheriting from inherited Tips - Make sure copy from copies everything and not just the value - check permissions ?? - Check if inherited properties are exported with the correct value and just report it but do not develop Example of inherited values ``` { sharedId: 456, title: 'Konz the famous magician', template: A, metadata: { favourite-colors: [{value: 28s7d3, label: 'green'}, {value: 234js9d3, label: 'blue'}], }, } { sharedId: 859, title: 'Rafa the copy cat', template: B, metadata: { inherited-color: [{value: [28s7d3, 234js9d3], label: ['green', 'blue'], ref: {value: '456', label: 'Konz the famous magician'}}], }, } ```
1.0
Filter and Aggregate by Inherited metadata (30pts) - Expected features: - Filter by inherited metadata instead of the title of the inherited entity - Show appropriate aggregations for each value. - Inherited filters should match with common filters. Implementation decisions - ~Template properties will have type (text, relationship...) and filterType (relationships will use inherited type)~ - ~The inheritedValue will be moved from inheritedValue to value~ - ~The value will be moved to ref {value: label:} or ref / refLabel~ - ~The label should be replaced with the inherited property label~ - Do not allow inheriting from inherited Tips - Make sure copy from copies everything and not just the value - check permissions ?? - Check if inherited properties are exported with the correct value and just report it but do not develop Example of inherited values ``` { sharedId: 456, title: 'Konz the famous magician', template: A, metadata: { favourite-colors: [{value: 28s7d3, label: 'green'}, {value: 234js9d3, label: 'blue'}], }, } { sharedId: 859, title: 'Rafa the copy cat', template: B, metadata: { inherited-color: [{value: [28s7d3, 234js9d3], label: ['green', 'blue'], ref: {value: '456', label: 'Konz the famous magician'}}], }, } ```
priority
filter and aggregate by inherited metadata expected features filter by inherited metadata instead of the title of the inherited entity show appropriate aggregations for each value inherited filters should match with common filters implementation decisions template properties will have type text relationship and filtertype relationships will use inherited type the inheritedvalue will be moved from inheritedvalue to value the value will be moved to ref value label or ref reflabel the label should be replaced with the inherited property label do not allow inheriting from inherited tips make sure copy from copies everything and not just the value check permissions check if inherited properties are exported with the correct value and just report it but do not develop example of inherited values sharedid title konz the famous magician template a metadata favourite colors sharedid title rafa the copy cat template b metadata inherited color label ref value label konz the famous magician
1
819,507
30,739,158,946
IssuesEvent
2023-07-28 10:01:56
MaFi-Games/Captain-of-Industry-issues
https://api.github.com/repos/MaFi-Games/Captain-of-Industry-issues
closed
Its dark i cant see.
could not reproduce high priority
### Issue description So with the newest update i found a bug that crashes my game every time. So i unlocked the transformer for my ppl but every time i place it i get the white boxs that has the Captain of Industry logo on it and my game crashes. thankfully autosave is on but it dark out here my ppl cant see when they wake up at 4 am to go to work and no coffee OMG there so slow now haha. ### Expected behavior just expected it to place. i did this 2 times till i figured out what it was. ### Reproduction steps 1. click on transformer 2. place transformer 3. disapointment has arrived 4. game crashed ### Game version v0.4.13 ### Attachments https://1drv.ms/u/s!ArMq3s42zGYsgQUB0hvAyNvBGQXi?e=rXHZpr
1.0
Its dark i cant see. - ### Issue description So with the newest update i found a bug that crashes my game every time. So i unlocked the transformer for my ppl but every time i place it i get the white boxs that has the Captain of Industry logo on it and my game crashes. thankfully autosave is on but it dark out here my ppl cant see when they wake up at 4 am to go to work and no coffee OMG there so slow now haha. ### Expected behavior just expected it to place. i did this 2 times till i figured out what it was. ### Reproduction steps 1. click on transformer 2. place transformer 3. disapointment has arrived 4. game crashed ### Game version v0.4.13 ### Attachments https://1drv.ms/u/s!ArMq3s42zGYsgQUB0hvAyNvBGQXi?e=rXHZpr
priority
its dark i cant see issue description so with the newest update i found a bug that crashes my game every time so i unlocked the transformer for my ppl but every time i place it i get the white boxs that has the captain of industry logo on it and my game crashes thankfully autosave is on but it dark out here my ppl cant see when they wake up at am to go to work and no coffee omg there so slow now haha expected behavior just expected it to place i did this times till i figured out what it was reproduction steps click on transformer place transformer disapointment has arrived game crashed game version attachments
1
172,827
6,516,704,776
IssuesEvent
2017-08-27 13:16:54
fossasia/open-event
https://api.github.com/repos/fossasia/open-event
closed
Update all the samples according to the new API specifications [ On Hold ]
has-PR Priority HIGH
The samples in the repo have become inconsistent. [Mozilla All Hands 2017](https://github.com/fossasia/open-event/pull/191/files) and [FOSSASIA 2016](https://github.com/fossasia/open-event/commit/77783374aca8b9cc2849484564dbd9cab54aff37) are using the new attributes as defined by the new API specifications while the other events are using the old attributes. Since we are removing support for the old attributes from the web and Android generators too, it would be good to convert all the remaining samples too according to the new API specs. Otherwise, a part of the samples (either old or new) won't be working anymore. Please suggest. Thanks :) @mariobehling @niranjan94 @aayusharora @Shailesh351
1.0
Update all the samples according to the new API specifications [ On Hold ] - The samples in the repo have become inconsistent. [Mozilla All Hands 2017](https://github.com/fossasia/open-event/pull/191/files) and [FOSSASIA 2016](https://github.com/fossasia/open-event/commit/77783374aca8b9cc2849484564dbd9cab54aff37) are using the new attributes as defined by the new API specifications while the other events are using the old attributes. Since we are removing support for the old attributes from the web and Android generators too, it would be good to convert all the remaining samples too according to the new API specs. Otherwise, a part of the samples (either old or new) won't be working anymore. Please suggest. Thanks :) @mariobehling @niranjan94 @aayusharora @Shailesh351
priority
update all the samples according to the new api specifications the samples in the repo have become inconsistent and are using the new attributes as defined by the new api specifications while the other events are using the old attributes since we are removing support for the old attributes from the web and android generators too it would be good to convert all the remaining samples too according to the new api specs otherwise a part of the samples either old or new won t be working anymore please suggest thanks mariobehling aayusharora
1
354,736
10,571,552,139
IssuesEvent
2019-10-07 07:27:50
hotosm/tasking-manager
https://api.github.com/repos/hotosm/tasking-manager
closed
Cannot execute migrations for staging deployment
Component: Backend Difficulty: Medium Priority: High Status: In Progress Type: Bug
Per chat with @dakotabenjamin and @thadk. Migration script returns the following error within staging server: ```ascii' codec can't decode byte 0xc3 in position 5067619: ordinal not in range(128)``` This error is present within migration file `migrations/versions/0eee8c1abd3a_.py", line 60`
1.0
Cannot execute migrations for staging deployment - Per chat with @dakotabenjamin and @thadk. Migration script returns the following error within staging server: ```ascii' codec can't decode byte 0xc3 in position 5067619: ordinal not in range(128)``` This error is present within migration file `migrations/versions/0eee8c1abd3a_.py", line 60`
priority
cannot execute migrations for staging deployment per chat with dakotabenjamin and thadk migration script returns the following error within staging server ascii codec can t decode byte in position ordinal not in range this error is present within migration file migrations versions py line
1
342,471
10,317,599,834
IssuesEvent
2019-08-30 13:07:13
pmem/issues
https://api.github.com/repos/pmem/issues
closed
Test: pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck)
Exposure: Medium OS: Linux Priority: 2 high Type: Bug
<!-- Before creating new issue, ensure that similar issue wasn't already created * Search: https://github.com/pmem/issues/issues Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report. Remember this is just a minimal template. You can extend it with data you think may be useful. --> # ISSUE: <!-- fill the title of issue --> ## Environment Information - PMDK package version(s): 1.4.3-rc1 - OS(es) version(s): SLES 12.4 - ndctl version(s): 61.2 - kernel version(s): 4.12.14-95.29-default ## Please provide a reproduction of the bug: ``` ./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all ``` ## How often bug is revealed: (always, often, rare): always <!-- describe special circumstances in section above --> ``` ./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck) pmempool_transform/TEST18 failed with Valgrind. See memcheck18.log. First 20 lines below. pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x2f5c00000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1bae00000, 0x1fae00000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x33ae00000, 0x37adff000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== pmempool_transform/TEST18 memcheck18.log ==31132== HEAP SUMMARY: pmempool_transform/TEST18 memcheck18.log ==31132== in use at exit: 0 bytes in 0 blocks pmempool_transform/TEST18 memcheck18.log ==31132== total heap usage: 1,148 allocs, 1,148 frees, 3,619,892 bytes allocated pmempool_transform/TEST18 memcheck18.log ==31132== pmempool_transform/TEST18 memcheck18.log ==31132== All heap blocks were freed -- no leaks are possible pmempool_transform/TEST18 memcheck18.log ==31132== pmempool_transform/TEST18 memcheck18.log ==31132== For counts of detected and suppressed errors, rerun with: -v pmempool_transform/TEST18 memcheck18.log ==31132== ERROR SUMMARY: 6 errors from 3 contexts (suppressed: 0 from 0) RUNTESTS: stopping: pmempool_transform/TEST18 failed, TEST=all FS=any BUILD=debug ``` <!-- fill this out --> ## Expected behavior: Test should pass. ## Details <!-- fill this out --> ## Additional information about Priority and Help Requested: Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible --> Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
1.0
Test: pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck) - <!-- Before creating new issue, ensure that similar issue wasn't already created * Search: https://github.com/pmem/issues/issues Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report. Remember this is just a minimal template. You can extend it with data you think may be useful. --> # ISSUE: <!-- fill the title of issue --> ## Environment Information - PMDK package version(s): 1.4.3-rc1 - OS(es) version(s): SLES 12.4 - ndctl version(s): 61.2 - kernel version(s): 4.12.14-95.29-default ## Please provide a reproduction of the bug: ``` ./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all ``` ## How often bug is revealed: (always, often, rare): always <!-- describe special circumstances in section above --> ``` ./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck) pmempool_transform/TEST18 failed with Valgrind. See memcheck18.log. First 20 lines below. pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x2f5c00000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1bae00000, 0x1fae00000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x33ae00000, 0x37adff000) (defined) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (noaccess) pmempool_transform/TEST18 memcheck18.log ==31132== pmempool_transform/TEST18 memcheck18.log ==31132== HEAP SUMMARY: pmempool_transform/TEST18 memcheck18.log ==31132== in use at exit: 0 bytes in 0 blocks pmempool_transform/TEST18 memcheck18.log ==31132== total heap usage: 1,148 allocs, 1,148 frees, 3,619,892 bytes allocated pmempool_transform/TEST18 memcheck18.log ==31132== pmempool_transform/TEST18 memcheck18.log ==31132== All heap blocks were freed -- no leaks are possible pmempool_transform/TEST18 memcheck18.log ==31132== pmempool_transform/TEST18 memcheck18.log ==31132== For counts of detected and suppressed errors, rerun with: -v pmempool_transform/TEST18 memcheck18.log ==31132== ERROR SUMMARY: 6 errors from 3 contexts (suppressed: 0 from 0) RUNTESTS: stopping: pmempool_transform/TEST18 failed, TEST=all FS=any BUILD=debug ``` <!-- fill this out --> ## Expected behavior: Test should pass. ## Details <!-- fill this out --> ## Additional information about Priority and Help Requested: Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible --> Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
priority
test pmempool transform setup all pmem debug memcheck before creating new issue ensure that similar issue wasn t already created search note that if you do not provide enough information to reproduce the issue we may not be able to take action on your report remember this is just a minimal template you can extend it with data you think may be useful issue environment information pmdk package version s os es version s sles ndctl version s kernel version s default please provide a reproduction of the bug runtests pmempool transform s m force enable t all how often bug is revealed always often rare always runtests pmempool transform s m force enable t all pmempool transform setup all pmem debug memcheck pmempool transform failed with valgrind see log first lines below pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range noaccess pmempool transform log pmempool transform log heap summary pmempool transform log in use at exit bytes in blocks pmempool transform log total heap usage allocs frees bytes allocated pmempool transform log pmempool transform log all heap blocks were freed no leaks are possible pmempool transform log pmempool transform log for counts of detected and suppressed errors rerun with v pmempool transform log error summary errors from contexts suppressed from runtests stopping pmempool transform failed test all fs any build debug expected behavior test should pass details additional information about priority and help requested are you willing to submit a pull request with a proposed change yes no requested priority showstopper high medium low
1
215,543
7,295,073,962
IssuesEvent
2018-02-26 04:29:18
utopian-io/utopian.io
https://api.github.com/repos/utopian-io/utopian.io
closed
Improve GitHub Integration
feature high-priority review
Show the synced github account on the contribution details and only allow to add a contribution on the development category if github is synced. If a user has no github account connected show a notice if he tries to create a development contribution
1.0
Improve GitHub Integration - Show the synced github account on the contribution details and only allow to add a contribution on the development category if github is synced. If a user has no github account connected show a notice if he tries to create a development contribution
priority
improve github integration show the synced github account on the contribution details and only allow to add a contribution on the development category if github is synced if a user has no github account connected show a notice if he tries to create a development contribution
1
205,760
7,105,799,394
IssuesEvent
2018-01-16 14:49:37
StratoDem/sd-material-ui
https://api.github.com/repos/StratoDem/sd-material-ui
opened
Buttons width is not adjusting correctly
Priority: High Tech: JS Type: Bug
The SDRaisedButton does not respond properly to width adjustment. The button itself can be reduced in width, but its containing Div does not shrink, and can disrupt the page's styling when it leaves a blank space. Change the SDRaisedButton (and SDFlatButton if necessary) to allow it to change style properly with the button. ![button-style](https://user-images.githubusercontent.com/16080662/34994702-612efc38-faa2-11e7-9124-97e2698460ce.png)
1.0
Buttons width is not adjusting correctly - The SDRaisedButton does not respond properly to width adjustment. The button itself can be reduced in width, but its containing Div does not shrink, and can disrupt the page's styling when it leaves a blank space. Change the SDRaisedButton (and SDFlatButton if necessary) to allow it to change style properly with the button. ![button-style](https://user-images.githubusercontent.com/16080662/34994702-612efc38-faa2-11e7-9124-97e2698460ce.png)
priority
buttons width is not adjusting correctly the sdraisedbutton does not respond properly to width adjustment the button itself can be reduced in width but its containing div does not shrink and can disrupt the page s styling when it leaves a blank space change the sdraisedbutton and sdflatbutton if necessary to allow it to change style properly with the button
1
659,634
21,935,585,039
IssuesEvent
2022-05-23 13:32:37
CarnegieLearningWeb/UpGrade
https://api.github.com/repos/CarnegieLearningWeb/UpGrade
reopened
Patch code for XPRIZE and Demonstration phase
priority: high
We need to decide what to do with the current xprize patch code and how it should be maintained if we're advanced to the demonstration phase (probably starting in Sept 22). Do we need to create a more permanent solution to the issue of experiment point/id conflicts?
1.0
Patch code for XPRIZE and Demonstration phase - We need to decide what to do with the current xprize patch code and how it should be maintained if we're advanced to the demonstration phase (probably starting in Sept 22). Do we need to create a more permanent solution to the issue of experiment point/id conflicts?
priority
patch code for xprize and demonstration phase we need to decide what to do with the current xprize patch code and how it should be maintained if we re advanced to the demonstration phase probably starting in sept do we need to create a more permanent solution to the issue of experiment point id conflicts
1
723,549
24,901,315,059
IssuesEvent
2022-10-28 21:15:13
virtualcell/vcell
https://api.github.com/repos/virtualcell/vcell
closed
Orphan overrides due to init/runtime naming issues
bug High Priority VCell-7.5.0
This issue affects clamped variables. Let's assume species EGF is clamped with some number or expression. Previously math generation would generate two constants, named EGF and EGF_init, both of which would be listed in the math and in the table for overrides. But only one of them (the first one) would actually be used in the equations. If the user would override EGF_init, nothing would actually happen (which in itself is a serious issue). But in most cases both of them would be overridden, usually with the same value/expression (if different, the first would be actually applied). Currently math generation will generate only one constant, named EGF_init_uM (whatever the units may be are suffixed). We are applying the pre-existing math override for EGF_init to the constant EGF_init_uM. This is wrong for two reasons: - first, whatever the override previously was on EGF_init was being ignored anyway; we should in fact apply the pre-existing math override for EGF to the constant EGF_init_uM in order to get the same results - second, the previous override on EGF would now be "oprhaned" and an error generated which would block any save/run/export until the user fixes it; however, the user does not have any way of understanding this issue and apply the correct solution manually
1.0
Orphan overrides due to init/runtime naming issues - This issue affects clamped variables. Let's assume species EGF is clamped with some number or expression. Previously math generation would generate two constants, named EGF and EGF_init, both of which would be listed in the math and in the table for overrides. But only one of them (the first one) would actually be used in the equations. If the user would override EGF_init, nothing would actually happen (which in itself is a serious issue). But in most cases both of them would be overridden, usually with the same value/expression (if different, the first would be actually applied). Currently math generation will generate only one constant, named EGF_init_uM (whatever the units may be are suffixed). We are applying the pre-existing math override for EGF_init to the constant EGF_init_uM. This is wrong for two reasons: - first, whatever the override previously was on EGF_init was being ignored anyway; we should in fact apply the pre-existing math override for EGF to the constant EGF_init_uM in order to get the same results - second, the previous override on EGF would now be "oprhaned" and an error generated which would block any save/run/export until the user fixes it; however, the user does not have any way of understanding this issue and apply the correct solution manually
priority
orphan overrides due to init runtime naming issues this issue affects clamped variables let s assume species egf is clamped with some number or expression previously math generation would generate two constants named egf and egf init both of which would be listed in the math and in the table for overrides but only one of them the first one would actually be used in the equations if the user would override egf init nothing would actually happen which in itself is a serious issue but in most cases both of them would be overridden usually with the same value expression if different the first would be actually applied currently math generation will generate only one constant named egf init um whatever the units may be are suffixed we are applying the pre existing math override for egf init to the constant egf init um this is wrong for two reasons first whatever the override previously was on egf init was being ignored anyway we should in fact apply the pre existing math override for egf to the constant egf init um in order to get the same results second the previous override on egf would now be oprhaned and an error generated which would block any save run export until the user fixes it however the user does not have any way of understanding this issue and apply the correct solution manually
1
461,670
13,234,182,442
IssuesEvent
2020-08-18 15:53:18
sugarlabs-appstore/aslo-v4
https://api.github.com/repos/sugarlabs-appstore/aslo-v4
opened
Rewrite static resources
high-priority
The static (HTML, CSS, JS) libraries, following #37 , #35 , #28, #34, and other performance issues, it is important to rewrite the Appstore again in a more performance optimized Javascript framework. Possibly, by using the powerful, but new CSS library Bulma, replacing the Bootstrap 4 library.
1.0
Rewrite static resources - The static (HTML, CSS, JS) libraries, following #37 , #35 , #28, #34, and other performance issues, it is important to rewrite the Appstore again in a more performance optimized Javascript framework. Possibly, by using the powerful, but new CSS library Bulma, replacing the Bootstrap 4 library.
priority
rewrite static resources the static html css js libraries following and other performance issues it is important to rewrite the appstore again in a more performance optimized javascript framework possibly by using the powerful but new css library bulma replacing the bootstrap library
1
138,090
5,327,864,311
IssuesEvent
2017-02-15 10:22:51
odalic/sti
https://api.github.com/repos/odalic/sti
closed
Cannot process one of the testing files - it was working before!
bug priority: High
Please set up knowledge base as defined at: https://grips.semantic-web.at/display/ADEQ/Odalic+-+Connection+to+the+Knowledge+Base Use file: [STMK_2012_LIVEBIRTHS-10lines.csv.zip](https://github.com/odalic/sti/files/776542/STMK_2012_LIVEBIRTHS-10lines.csv.zip) Cannot further process that file: but it was working before (approx 2 weeks back) The problematic query ``` SELECT DISTINCT ?subject WHERE { ?subject <http://www.w3.org/2004/02/skos/core#prefLabel> "Gro� Sankt Florian"@de ; a <http://www.w3.org/2004/02/skos/core#Concept> } ``` I guess that there may be an issue with that special char? The log: ``` java.util.concurrent.ExecutionException: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at java.util.concurrent.FutureTask.report(FutureTask.java:122) at java.util.concurrent.FutureTask.get(FutureTask.java:192) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.getResultForTaskId(DbCachedFutureBasedExecutionService.java:227) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:63) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:78) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.glassfish.jersey.server.model.internal.ResourceMethodInvocationHandlerFactory$1.invoke(ResourceMethodInvocationHandlerFactory.java:81) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher$1.run(AbstractJavaResourceMethodDispatcher.java:144) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.invoke(AbstractJavaResourceMethodDispatcher.java:161) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$ResponseOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:160) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:99) at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:389) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:347) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:102) at org.glassfish.jersey.server.ServerRuntime$2.run(ServerRuntime.java:326) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at org.glassfish.jersey.internal.Errors.process(Errors.java:315) at org.glassfish.jersey.internal.Errors.process(Errors.java:297) at org.glassfish.jersey.internal.Errors.process(Errors.java:267) at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317) at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:305) at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:1154) at org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:473) at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:427) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:388) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:341) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:228) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:292) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:240) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212) at org.apache.catalina.core.StandardContextValve.__invoke(StandardContextValve.java:106) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:528) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1099) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:670) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1520) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1476) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:745) Caused by: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:370) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:313) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:227) at javax.xml.stream.XMLInputFactory.newInstance(XMLInputFactory.java:154) at org.apache.jena.sparql.resultset.XMLInputStAX.<init>(XMLInputStAX.java:101) at org.apache.jena.sparql.resultset.XMLInput.make(XMLInput.java:74) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:38) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:34) at org.apache.jena.query.ResultSetFactory.fromXML(ResultSetFactory.java:308) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execResultSetInner(QueryEngineHTTP.java:365) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execSelect(QueryEngineHTTP.java:334) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:228) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:219) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryEntityCandidates(SPARQLProxy.java:631) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.findEntityCandidatesInternal(SPARQLProxy.java:364) at uk.ac.shef.dcs.kbproxy.KBProxy.lambda$findEntityCandidates$0(KBProxy.java:144) at uk.ac.shef.dcs.kbproxy.KBProxy.Do(KBProxy.java:387) at uk.ac.shef.dcs.kbproxy.KBProxy.findEntityCandidates(KBProxy.java:144) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNINGPreliminaryColumnClassifier.runPreliminaryColumnClassifier(LEARNINGPreliminaryColumnClassifier.java:120) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNING.learn(LEARNING.java:37) at uk.ac.shef.dcs.sti.core.algorithm.tmp.TMPOdalicInterpreter.start(TMPOdalicInterpreter.java:114) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.lambda$submitForTaskId$29(DbCachedFutureBasedExecutionService.java:160) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ... 1 more Caused by: java.lang.RuntimeException: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:367) ... 25 more Caused by: java.util.ServiceConfigurationError: javax.xml.stream.XMLInputFactory: Error reading configuration file at java.util.ServiceLoader.fail(ServiceLoader.java:232) at java.util.ServiceLoader.parse(ServiceLoader.java:309) at java.util.ServiceLoader.access$200(ServiceLoader.java:185) at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) at javax.xml.stream.FactoryFinder$1.run(FactoryFinder.java:352) at java.security.AccessController.doPrivileged(Native Method) at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:341) ... 25 more Caused by: java.io.FileNotFoundException: /usr/local/apache-tomcat-8.0.36/webapps/odalic/WEB-INF/lib/woodstox-core-asl-4.4.1.jar (No such file or directory) at java.util.zip.ZipFile.open(Native Method) at java.util.zip.ZipFile.<init>(ZipFile.java:219) at java.util.zip.ZipFile.<init>(ZipFile.java:149) at java.util.jar.JarFile.<init>(JarFile.java:166) at java.util.jar.JarFile.<init>(JarFile.java:103) at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) at java.net.URL.openStream(URL.java:1038) at java.util.ServiceLoader.parse(ServiceLoader.java:304) ... 32 more ``` This also means that #166 does not work properly??
1.0
Cannot process one of the testing files - it was working before! - Please set up knowledge base as defined at: https://grips.semantic-web.at/display/ADEQ/Odalic+-+Connection+to+the+Knowledge+Base Use file: [STMK_2012_LIVEBIRTHS-10lines.csv.zip](https://github.com/odalic/sti/files/776542/STMK_2012_LIVEBIRTHS-10lines.csv.zip) Cannot further process that file: but it was working before (approx 2 weeks back) The problematic query ``` SELECT DISTINCT ?subject WHERE { ?subject <http://www.w3.org/2004/02/skos/core#prefLabel> "Gro� Sankt Florian"@de ; a <http://www.w3.org/2004/02/skos/core#Concept> } ``` I guess that there may be an issue with that special char? The log: ``` java.util.concurrent.ExecutionException: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at java.util.concurrent.FutureTask.report(FutureTask.java:122) at java.util.concurrent.FutureTask.get(FutureTask.java:192) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.getResultForTaskId(DbCachedFutureBasedExecutionService.java:227) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:63) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:78) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.glassfish.jersey.server.model.internal.ResourceMethodInvocationHandlerFactory$1.invoke(ResourceMethodInvocationHandlerFactory.java:81) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher$1.run(AbstractJavaResourceMethodDispatcher.java:144) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.invoke(AbstractJavaResourceMethodDispatcher.java:161) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$ResponseOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:160) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:99) at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:389) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:347) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:102) at org.glassfish.jersey.server.ServerRuntime$2.run(ServerRuntime.java:326) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at org.glassfish.jersey.internal.Errors.process(Errors.java:315) at org.glassfish.jersey.internal.Errors.process(Errors.java:297) at org.glassfish.jersey.internal.Errors.process(Errors.java:267) at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317) at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:305) at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:1154) at org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:473) at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:427) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:388) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:341) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:228) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:292) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:240) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212) at org.apache.catalina.core.StandardContextValve.__invoke(StandardContextValve.java:106) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:528) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1099) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:670) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1520) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1476) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:745) Caused by: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:370) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:313) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:227) at javax.xml.stream.XMLInputFactory.newInstance(XMLInputFactory.java:154) at org.apache.jena.sparql.resultset.XMLInputStAX.<init>(XMLInputStAX.java:101) at org.apache.jena.sparql.resultset.XMLInput.make(XMLInput.java:74) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:38) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:34) at org.apache.jena.query.ResultSetFactory.fromXML(ResultSetFactory.java:308) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execResultSetInner(QueryEngineHTTP.java:365) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execSelect(QueryEngineHTTP.java:334) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:228) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:219) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryEntityCandidates(SPARQLProxy.java:631) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.findEntityCandidatesInternal(SPARQLProxy.java:364) at uk.ac.shef.dcs.kbproxy.KBProxy.lambda$findEntityCandidates$0(KBProxy.java:144) at uk.ac.shef.dcs.kbproxy.KBProxy.Do(KBProxy.java:387) at uk.ac.shef.dcs.kbproxy.KBProxy.findEntityCandidates(KBProxy.java:144) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNINGPreliminaryColumnClassifier.runPreliminaryColumnClassifier(LEARNINGPreliminaryColumnClassifier.java:120) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNING.learn(LEARNING.java:37) at uk.ac.shef.dcs.sti.core.algorithm.tmp.TMPOdalicInterpreter.start(TMPOdalicInterpreter.java:114) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.lambda$submitForTaskId$29(DbCachedFutureBasedExecutionService.java:160) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ... 1 more Caused by: java.lang.RuntimeException: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:367) ... 25 more Caused by: java.util.ServiceConfigurationError: javax.xml.stream.XMLInputFactory: Error reading configuration file at java.util.ServiceLoader.fail(ServiceLoader.java:232) at java.util.ServiceLoader.parse(ServiceLoader.java:309) at java.util.ServiceLoader.access$200(ServiceLoader.java:185) at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) at javax.xml.stream.FactoryFinder$1.run(FactoryFinder.java:352) at java.security.AccessController.doPrivileged(Native Method) at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:341) ... 25 more Caused by: java.io.FileNotFoundException: /usr/local/apache-tomcat-8.0.36/webapps/odalic/WEB-INF/lib/woodstox-core-asl-4.4.1.jar (No such file or directory) at java.util.zip.ZipFile.open(Native Method) at java.util.zip.ZipFile.<init>(ZipFile.java:219) at java.util.zip.ZipFile.<init>(ZipFile.java:149) at java.util.jar.JarFile.<init>(JarFile.java:166) at java.util.jar.JarFile.<init>(JarFile.java:103) at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) at java.net.URL.openStream(URL.java:1038) at java.util.ServiceLoader.parse(ServiceLoader.java:304) ... 32 more ``` This also means that #166 does not work properly??
priority
cannot process one of the testing files it was working before please set up knowledge base as defined at use file cannot further process that file but it was working before approx weeks back the problematic query select distinct subject where subject gro� sankt florian de a i guess that there may be an issue with that special char the log java util concurrent executionexception javax xml stream factoryconfigurationerror provider for class javax xml stream xmlinputfactory cannot be created at java util concurrent futuretask report futuretask java at java util concurrent futuretask get futuretask java at cz cuni mff xrg odalic tasks executions dbcachedfuturebasedexecutionservice getresultfortaskid dbcachedfuturebasedexecutionservice java at cz cuni mff xrg odalic api rest resources resultresource getresult resultresource java at cz cuni mff xrg odalic api rest resources resultresource getresult resultresource java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org glassfish jersey server model internal resourcemethodinvocationhandlerfactory invoke resourcemethodinvocationhandlerfactory java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher run abstractjavaresourcemethoddispatcher java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher invoke abstractjavaresourcemethoddispatcher java at org glassfish jersey server model internal javaresourcemethoddispatcherprovider responseoutinvoker dodispatch javaresourcemethoddispatcherprovider java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher dispatch abstractjavaresourcemethoddispatcher java at org glassfish jersey server model resourcemethodinvoker invoke resourcemethodinvoker java at org glassfish jersey server model resourcemethodinvoker apply resourcemethodinvoker java at org glassfish jersey server model resourcemethodinvoker apply resourcemethodinvoker java at org glassfish jersey server serverruntime run serverruntime java at org glassfish jersey internal errors call errors java at org glassfish jersey internal errors call errors java at org glassfish jersey internal errors process errors java at org glassfish jersey internal errors process errors java at org glassfish jersey internal errors process errors java at org glassfish jersey process internal requestscope runinscope requestscope java at org glassfish jersey server serverruntime process serverruntime java at org glassfish jersey server applicationhandler handle applicationhandler java at org glassfish jersey servlet webcomponent serviceimpl webcomponent java at org glassfish jersey servlet webcomponent service webcomponent java at org glassfish jersey servlet servletcontainer service servletcontainer java at org glassfish jersey servlet servletcontainer service servletcontainer java at org glassfish jersey servlet servletcontainer service servletcontainer java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote process java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net nioendpoint socketprocessor run nioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java caused by javax xml stream factoryconfigurationerror provider for class javax xml stream xmlinputfactory cannot be created at javax xml stream factoryfinder findserviceprovider factoryfinder java at javax xml stream factoryfinder find factoryfinder java at javax xml stream factoryfinder find factoryfinder java at javax xml stream xmlinputfactory newinstance xmlinputfactory java at org apache jena sparql resultset xmlinputstax xmlinputstax java at org apache jena sparql resultset xmlinput make xmlinput java at org apache jena sparql resultset xmlinput fromxml xmlinput java at org apache jena sparql resultset xmlinput fromxml xmlinput java at org apache jena query resultsetfactory fromxml resultsetfactory java at org apache jena sparql engine http queryenginehttp execresultsetinner queryenginehttp java at org apache jena sparql engine http queryenginehttp execselect queryenginehttp java at uk ac shef dcs kbproxy sparql sparqlproxy queryreturnsinglevalues sparqlproxy java at uk ac shef dcs kbproxy sparql sparqlproxy queryreturnsinglevalues sparqlproxy java at uk ac shef dcs kbproxy sparql sparqlproxy queryentitycandidates sparqlproxy java at uk ac shef dcs kbproxy sparql sparqlproxy findentitycandidatesinternal sparqlproxy java at uk ac shef dcs kbproxy kbproxy lambda findentitycandidates kbproxy java at uk ac shef dcs kbproxy kbproxy do kbproxy java at uk ac shef dcs kbproxy kbproxy findentitycandidates kbproxy java at uk ac shef dcs sti core algorithm tmp learningpreliminarycolumnclassifier runpreliminarycolumnclassifier learningpreliminarycolumnclassifier java at uk ac shef dcs sti core algorithm tmp learning learn learning java at uk ac shef dcs sti core algorithm tmp tmpodalicinterpreter start tmpodalicinterpreter java at cz cuni mff xrg odalic tasks executions dbcachedfuturebasedexecutionservice lambda submitfortaskid dbcachedfuturebasedexecutionservice java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java more caused by java lang runtimeexception provider for class javax xml stream xmlinputfactory cannot be created at javax xml stream factoryfinder findserviceprovider factoryfinder java more caused by java util serviceconfigurationerror javax xml stream xmlinputfactory error reading configuration file at java util serviceloader fail serviceloader java at java util serviceloader parse serviceloader java at java util serviceloader access serviceloader java at java util serviceloader lazyiterator hasnextservice serviceloader java at java util serviceloader lazyiterator hasnext serviceloader java at java util serviceloader hasnext serviceloader java at javax xml stream factoryfinder run factoryfinder java at java security accesscontroller doprivileged native method at javax xml stream factoryfinder findserviceprovider factoryfinder java more caused by java io filenotfoundexception usr local apache tomcat webapps odalic web inf lib woodstox core asl jar no such file or directory at java util zip zipfile open native method at java util zip zipfile zipfile java at java util zip zipfile zipfile java at java util jar jarfile jarfile java at java util jar jarfile jarfile java at sun net at sun net at sun net at sun net at sun net at java net url openstream url java at java util serviceloader parse serviceloader java more this also means that does not work properly
1
413,283
12,064,394,733
IssuesEvent
2020-04-16 08:12:46
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Read-only constraints not being consistently applied
Bug: development Docs: not needed Effort: small Feature Ivory Coast (phase 2) Module: dispensary Priority: high
## Is your feature request related to a problem? Please describe. A few problems/inconsistencies with dispensing for non-local patients: - the `supplyingStoreId` field is not being correctly set for patients created via the lookup API. - the `storeId` field is not being correctly set for patients created via the lookup API. - non-local patients/prescribers cannot be edited, but the edit form can still be accessed and the icon is not disabled as it is for policies. ## Describe the solution you'd like - correctly set supplying store fields for patients/prescribers. - disable edit icon in the dispensing window for non-local patients/prescribers ## Implementation - update `PatientActions.patientUpdate`. - update `PrescriberActions.updatePrescriber` - update `PrescriptionInfo` component. ## Describe alternatives you've considered N/A. ## Additional context See epic #2446.
1.0
Read-only constraints not being consistently applied - ## Is your feature request related to a problem? Please describe. A few problems/inconsistencies with dispensing for non-local patients: - the `supplyingStoreId` field is not being correctly set for patients created via the lookup API. - the `storeId` field is not being correctly set for patients created via the lookup API. - non-local patients/prescribers cannot be edited, but the edit form can still be accessed and the icon is not disabled as it is for policies. ## Describe the solution you'd like - correctly set supplying store fields for patients/prescribers. - disable edit icon in the dispensing window for non-local patients/prescribers ## Implementation - update `PatientActions.patientUpdate`. - update `PrescriberActions.updatePrescriber` - update `PrescriptionInfo` component. ## Describe alternatives you've considered N/A. ## Additional context See epic #2446.
priority
read only constraints not being consistently applied is your feature request related to a problem please describe a few problems inconsistencies with dispensing for non local patients the supplyingstoreid field is not being correctly set for patients created via the lookup api the storeid field is not being correctly set for patients created via the lookup api non local patients prescribers cannot be edited but the edit form can still be accessed and the icon is not disabled as it is for policies describe the solution you d like correctly set supplying store fields for patients prescribers disable edit icon in the dispensing window for non local patients prescribers implementation update patientactions patientupdate update prescriberactions updateprescriber update prescriptioninfo component describe alternatives you ve considered n a additional context see epic
1
745,005
25,965,761,941
IssuesEvent
2022-12-19 06:41:58
paperclip-ui/paperclip
https://api.github.com/repos/paperclip-ui/paperclip
closed
Consider dropping brackets
priority: low impact: low effort: high
E.g: ```javascript import "./common.pc" as common import "./theme.pc" as theme import "./input.pc" as input component ColorPickerPopup render div style extends common.panel position: fixed flex-shrink:0 transform: translate(calc(-100% - var(theme.space07))) z-index: 1024 div header style padding: var(theme.space04) display: flex justify-content: space-between font-weight: 600 padding-bottom: var(theme.space04) div text "Color" div text "&times" div style padding: var(theme.space04) display: flex flex-direction: column gap: var(theme.space04) color-picker(width: "220", height: "130") div text "Hue" div text "Opacity" div text "value" div style border-top: 1px solid var(theme.borderColor) padding: var(theme.space04) display: flex flex-direction: column gap: var(theme.space03) div text "color var" div text "color var" div text "color var" ```
1.0
Consider dropping brackets - E.g: ```javascript import "./common.pc" as common import "./theme.pc" as theme import "./input.pc" as input component ColorPickerPopup render div style extends common.panel position: fixed flex-shrink:0 transform: translate(calc(-100% - var(theme.space07))) z-index: 1024 div header style padding: var(theme.space04) display: flex justify-content: space-between font-weight: 600 padding-bottom: var(theme.space04) div text "Color" div text "&times" div style padding: var(theme.space04) display: flex flex-direction: column gap: var(theme.space04) color-picker(width: "220", height: "130") div text "Hue" div text "Opacity" div text "value" div style border-top: 1px solid var(theme.borderColor) padding: var(theme.space04) display: flex flex-direction: column gap: var(theme.space03) div text "color var" div text "color var" div text "color var" ```
priority
consider dropping brackets e g javascript import common pc as common import theme pc as theme import input pc as input component colorpickerpopup render div style extends common panel position fixed flex shrink transform translate calc var theme z index div header style padding var theme display flex justify content space between font weight padding bottom var theme div text color div text times div style padding var theme display flex flex direction column gap var theme color picker width height div text hue div text opacity div text value div style border top solid var theme bordercolor padding var theme display flex flex direction column gap var theme div text color var div text color var div text color var
1