Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
855
| labels
stringlengths 4
721
| body
stringlengths 1
261k
| index
stringclasses 13
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
240k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,048
| 2,507,024,719
|
IssuesEvent
|
2015-01-12 15:39:29
|
G-Node/GCA-Web
|
https://api.github.com/repos/G-Node/GCA-Web
|
closed
|
Editor does not work in IE
|
bug high priority
|
The editor does not update the abstract properly in internet explorer (version 11 and others).
|
1.0
|
Editor does not work in IE - The editor does not update the abstract properly in internet explorer (version 11 and others).
|
priority
|
editor does not work in ie the editor does not update the abstract properly in internet explorer version and others
| 1
|
239,601
| 7,799,873,036
|
IssuesEvent
|
2018-06-09 01:30:38
|
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
|
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
|
closed
|
0005720:
'my contacts' favorite does not work correctly
|
Addressbook Bug Mantis high priority
|
**Reported by pschuele on 15 Feb 2012 14:41**
**Version:** Milan (2012-03) Beta 4
'my contacts' favorite does not work correctly
it looks like the server does not send the correct node info, the users displayname is shown in the filter instead of the node/path ...
|
1.0
|
0005720:
'my contacts' favorite does not work correctly - **Reported by pschuele on 15 Feb 2012 14:41**
**Version:** Milan (2012-03) Beta 4
'my contacts' favorite does not work correctly
it looks like the server does not send the correct node info, the users displayname is shown in the filter instead of the node/path ...
|
priority
|
my contacts favorite does not work correctly reported by pschuele on feb version milan beta my contacts favorite does not work correctly it looks like the server does not send the correct node info the users displayname is shown in the filter instead of the node path
| 1
|
390,186
| 11,527,787,139
|
IssuesEvent
|
2020-02-16 00:23:26
|
tempor1s/oracli
|
https://api.github.com/repos/tempor1s/oracli
|
closed
|
Require user auth to be pulled from header
|
enhancement help wanted high priority
|
Currently, you need to send the token over the body, but I would prefer if we can move this over to the token being sent in the header as it is a much cleaner implementation of authentication :)
|
1.0
|
Require user auth to be pulled from header - Currently, you need to send the token over the body, but I would prefer if we can move this over to the token being sent in the header as it is a much cleaner implementation of authentication :)
|
priority
|
require user auth to be pulled from header currently you need to send the token over the body but i would prefer if we can move this over to the token being sent in the header as it is a much cleaner implementation of authentication
| 1
|
712,312
| 24,490,137,051
|
IssuesEvent
|
2022-10-09 23:48:10
|
python/mypy
|
https://api.github.com/repos/python/mypy
|
closed
|
Give an error if there is a variable annotation within a function but no signature
|
feature priority-0-high topic-usability good-first-issue
|
Mypy should perhaps give an error about a variable annotation in an otherwise unanotated function. Example:
```py
def f():
a: int = 'x' # Maybe this annotation should be an error?
```
The rationale is that mypy will ignore the type annotation since the function is still considered unannotated, but this is confusing because there is an annotation *within* the function so it can appear to be annotated.
If `--check-untyped-defs` is being used this error shouldn't be generated.
Originally reported in #3945.
|
1.0
|
Give an error if there is a variable annotation within a function but no signature - Mypy should perhaps give an error about a variable annotation in an otherwise unanotated function. Example:
```py
def f():
a: int = 'x' # Maybe this annotation should be an error?
```
The rationale is that mypy will ignore the type annotation since the function is still considered unannotated, but this is confusing because there is an annotation *within* the function so it can appear to be annotated.
If `--check-untyped-defs` is being used this error shouldn't be generated.
Originally reported in #3945.
|
priority
|
give an error if there is a variable annotation within a function but no signature mypy should perhaps give an error about a variable annotation in an otherwise unanotated function example py def f a int x maybe this annotation should be an error the rationale is that mypy will ignore the type annotation since the function is still considered unannotated but this is confusing because there is an annotation within the function so it can appear to be annotated if check untyped defs is being used this error shouldn t be generated originally reported in
| 1
|
716,371
| 24,630,274,496
|
IssuesEvent
|
2022-10-17 00:57:10
|
adisve/tumble-for-kronox
|
https://api.github.com/repos/adisve/tumble-for-kronox
|
closed
|
German and French translation
|
enhancement High Priority
|
Update to the translations with the new strings that were added.
|
1.0
|
German and French translation - Update to the translations with the new strings that were added.
|
priority
|
german and french translation update to the translations with the new strings that were added
| 1
|
260,961
| 8,221,636,317
|
IssuesEvent
|
2018-09-06 03:06:55
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
closed
|
[studio] items are not deleted in stage on delete
|
CI bug priority: high
|
### Expected behavior
Deletes should be pushed to both live and staging
### Actual behavior
In 3.0.x delete is an immediate deploy to live. When staging is enabled, the delete takes place on live but not staging.
### Steps to reproduce the problem
* Create an a page
* Publish it to staging
* Publish it to live
* Delete the item
* Not that it still exists in staging
### Log/stack trace (use https://gist.github.com)
N/A
### Specs
#### Version
Studio Version Number: 3.0.16-SNAPSHOT-4c989c
Build Number: 4c989c7cb50201155c988637a6db453257b7a1cc
Build Date/Time: 08-02-2018 14:57:36 -0400
#### OS
Any
#### Browser
Any
|
1.0
|
[studio] items are not deleted in stage on delete - ### Expected behavior
Deletes should be pushed to both live and staging
### Actual behavior
In 3.0.x delete is an immediate deploy to live. When staging is enabled, the delete takes place on live but not staging.
### Steps to reproduce the problem
* Create an a page
* Publish it to staging
* Publish it to live
* Delete the item
* Not that it still exists in staging
### Log/stack trace (use https://gist.github.com)
N/A
### Specs
#### Version
Studio Version Number: 3.0.16-SNAPSHOT-4c989c
Build Number: 4c989c7cb50201155c988637a6db453257b7a1cc
Build Date/Time: 08-02-2018 14:57:36 -0400
#### OS
Any
#### Browser
Any
|
priority
|
items are not deleted in stage on delete expected behavior deletes should be pushed to both live and staging actual behavior in x delete is an immediate deploy to live when staging is enabled the delete takes place on live but not staging steps to reproduce the problem create an a page publish it to staging publish it to live delete the item not that it still exists in staging log stack trace use n a specs version studio version number snapshot build number build date time os any browser any
| 1
|
290,459
| 8,895,463,430
|
IssuesEvent
|
2019-01-16 08:45:24
|
telstra/open-kilda
|
https://api.github.com/repos/telstra/open-kilda
|
closed
|
New API: Reroute all flows which go through a particular ISL
|
area/api area/arch feature priority/2-high
|
Initiate a reroute for all flows on this ISL.
API: PATCH `links/flows/reroute?<params>`
Related: #1548 (Maintenance mode for ISL)
|
1.0
|
New API: Reroute all flows which go through a particular ISL - Initiate a reroute for all flows on this ISL.
API: PATCH `links/flows/reroute?<params>`
Related: #1548 (Maintenance mode for ISL)
|
priority
|
new api reroute all flows which go through a particular isl initiate a reroute for all flows on this isl api patch links flows reroute related maintenance mode for isl
| 1
|
369,920
| 10,919,947,713
|
IssuesEvent
|
2019-11-21 20:08:49
|
flextype/flextype
|
https://api.github.com/repos/flextype/flextype
|
closed
|
Flextype Core: Forms - add ability to hide fieldsets from entries type select
|
priority: high type: feature
|
We should have ability to hide fieldsets from entries type select, this will be useful for nested fieldsets. Possible we should add new property `hide` - by default is `false`
Logic:
if **hide** property is **true** then hide fieldsets from entries type select.
if **hide** property is **false** then show fieldsets from entries type select.
if **hide** property is **is not exists** then show fieldsets from entries type select.
|
1.0
|
Flextype Core: Forms - add ability to hide fieldsets from entries type select - We should have ability to hide fieldsets from entries type select, this will be useful for nested fieldsets. Possible we should add new property `hide` - by default is `false`
Logic:
if **hide** property is **true** then hide fieldsets from entries type select.
if **hide** property is **false** then show fieldsets from entries type select.
if **hide** property is **is not exists** then show fieldsets from entries type select.
|
priority
|
flextype core forms add ability to hide fieldsets from entries type select we should have ability to hide fieldsets from entries type select this will be useful for nested fieldsets possible we should add new property hide by default is false logic if hide property is true then hide fieldsets from entries type select if hide property is false then show fieldsets from entries type select if hide property is is not exists then show fieldsets from entries type select
| 1
|
139,443
| 5,375,479,836
|
IssuesEvent
|
2017-02-23 05:02:26
|
ArctosDB/documentation-wiki
|
https://api.github.com/repos/ArctosDB/documentation-wiki
|
opened
|
remove hyphens from anchor links when creating from <h3>subtitles
|
bug Priority: High question
|
Here's the problem:
in the current site the anchors were not necessarily the subheader/subtitle phrase:
https://arctosdb.org/documentation/agent/#namesearch
BUT the subheader is "Searching Agents"
Opt#1 I can go through and update the subheaders to match but in some cases like this one is leaves wtih some awkward phrasing.
Opt#2? Anyway to create anchors not from <h3>tags?
In any case we need to get rid of hyphens when constructing the anchors from subtitles
(readability pfft!)
This blocks the launch if we dont have a solid fix
|
1.0
|
remove hyphens from anchor links when creating from <h3>subtitles - Here's the problem:
in the current site the anchors were not necessarily the subheader/subtitle phrase:
https://arctosdb.org/documentation/agent/#namesearch
BUT the subheader is "Searching Agents"
Opt#1 I can go through and update the subheaders to match but in some cases like this one is leaves wtih some awkward phrasing.
Opt#2? Anyway to create anchors not from <h3>tags?
In any case we need to get rid of hyphens when constructing the anchors from subtitles
(readability pfft!)
This blocks the launch if we dont have a solid fix
|
priority
|
remove hyphens from anchor links when creating from subtitles here s the problem in the current site the anchors were not necessarily the subheader subtitle phrase but the subheader is searching agents opt i can go through and update the subheaders to match but in some cases like this one is leaves wtih some awkward phrasing opt anyway to create anchors not from tags in any case we need to get rid of hyphens when constructing the anchors from subtitles readability pfft this blocks the launch if we dont have a solid fix
| 1
|
396,193
| 11,705,177,872
|
IssuesEvent
|
2020-03-07 14:22:28
|
localstack/localstack
|
https://api.github.com/repos/localstack/localstack
|
closed
|
Missing required parameter in input: "FunctionName"
|
bug needs-triaging priority-high
|
While creating a lambda using CFN, my command fails with the following message:
```
2019-11-08 15:52:29,358:API: Error on request:
Traceback (most recent call last):
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 304, in run_wsgi
execute(self.server.app)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 292, in execute
application_iter = app(environ, start_response)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/server.py", line 132, in __call__
return backend_app(environ, start_response)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__
return self.wsgi_app(environ, start_response)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app
response = self.handle_exception(e)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function
return cors_after_request(app.make_response(f(*args, **kwargs)))
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1741, in handle_exception
reraise(exc_type, exc_value, tb)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise
raise value
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app
response = self.full_dispatch_request()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function
return cors_after_request(app.make_response(f(*args, **kwargs)))
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise
raise value
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request
rv = self.dispatch_request()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 140, in __call__
result = self.callback(request, request.url, {})
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 168, in dispatch
return cls()._dispatch(*args, **kwargs)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 259, in _dispatch
return self.call_action()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 340, in call_action
response = method()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 264, in _wrapper
response = f(*args, **kwargs)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/responses.py", line 108, in create_change_set
change_set_type=update_or_create,
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 452, in create_change_set
cross_stack_resources=self.exports
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 295, in __init__
create_change_set=True,
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 187, in __init__
self.resource_map = self._create_resource_map()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 195, in _create_resource_map
resource_map.create()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 482, in create
if isinstance(self[resource], ec2_models.TaggedEC2Resource):
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 415, in __getitem__
resource_logical_id, resource_json, self, self._region_name)
File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 173, in parse_and_create_resource
return _parse_and_create_resource(logical_id, resource_json, resources_map, region_name)
File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 260, in _parse_and_create_resource
result = deploy_func(logical_id, resource_wrapped, stack_name=stack_name)
File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 671, in deploy_resource
result = deploy_resource_via_sdk_function(resource_id, resources, resource_type, func, stack_name)
File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 741, in deploy_resource_via_sdk_function
raise e
File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 738, in deploy_resource_via_sdk_function
result = function(**params)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 357, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 634, in _make_api_call
api_params, operation_model, context=request_context)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 682, in _convert_to_request_dict
api_params, operation_model)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/validate.py", line 297, in serialize_to_request
raise ParamValidationError(report=report.generate_report())
botocore.exceptions.ParamValidationError: Parameter validation failed:
Missing required parameter in input: "FunctionName"
```
Considering the fact that the `FunctionName` property is not mandatory from CFN perspective (it's even worse, since having it or not has significant implications), is it possible to create a name on the fly?
|
1.0
|
Missing required parameter in input: "FunctionName" - While creating a lambda using CFN, my command fails with the following message:
```
2019-11-08 15:52:29,358:API: Error on request:
Traceback (most recent call last):
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 304, in run_wsgi
execute(self.server.app)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/werkzeug/serving.py", line 292, in execute
application_iter = app(environ, start_response)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/server.py", line 132, in __call__
return backend_app(environ, start_response)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2309, in __call__
return self.wsgi_app(environ, start_response)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2295, in wsgi_app
response = self.handle_exception(e)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function
return cors_after_request(app.make_response(f(*args, **kwargs)))
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1741, in handle_exception
reraise(exc_type, exc_value, tb)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise
raise value
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 2292, in wsgi_app
response = self.full_dispatch_request()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1815, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask_cors/extension.py", line 161, in wrapped_function
return cors_after_request(app.make_response(f(*args, **kwargs)))
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1718, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/_compat.py", line 35, in reraise
raise value
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1813, in full_dispatch_request
rv = self.dispatch_request()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/flask/app.py", line 1799, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 140, in __call__
result = self.callback(request, request.url, {})
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 168, in dispatch
return cls()._dispatch(*args, **kwargs)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 259, in _dispatch
return self.call_action()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/responses.py", line 340, in call_action
response = method()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/core/utils.py", line 264, in _wrapper
response = f(*args, **kwargs)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/responses.py", line 108, in create_change_set
change_set_type=update_or_create,
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 452, in create_change_set
cross_stack_resources=self.exports
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 295, in __init__
create_change_set=True,
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 187, in __init__
self.resource_map = self._create_resource_map()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/models.py", line 195, in _create_resource_map
resource_map.create()
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 482, in create
if isinstance(self[resource], ec2_models.TaggedEC2Resource):
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/moto/cloudformation/parsing.py", line 415, in __getitem__
resource_logical_id, resource_json, self, self._region_name)
File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 173, in parse_and_create_resource
return _parse_and_create_resource(logical_id, resource_json, resources_map, region_name)
File "/opt/code/localstack/localstack/services/cloudformation/cloudformation_starter.py", line 260, in _parse_and_create_resource
result = deploy_func(logical_id, resource_wrapped, stack_name=stack_name)
File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 671, in deploy_resource
result = deploy_resource_via_sdk_function(resource_id, resources, resource_type, func, stack_name)
File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 741, in deploy_resource_via_sdk_function
raise e
File "/opt/code/localstack/localstack/utils/cloudformation/template_deployer.py", line 738, in deploy_resource_via_sdk_function
result = function(**params)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 357, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 634, in _make_api_call
api_params, operation_model, context=request_context)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/client.py", line 682, in _convert_to_request_dict
api_params, operation_model)
File "/opt/code/localstack/.venv/lib/python3.6/site-packages/botocore/validate.py", line 297, in serialize_to_request
raise ParamValidationError(report=report.generate_report())
botocore.exceptions.ParamValidationError: Parameter validation failed:
Missing required parameter in input: "FunctionName"
```
Considering the fact that the `FunctionName` property is not mandatory from CFN perspective (it's even worse, since having it or not has significant implications), is it possible to create a name on the fly?
|
priority
|
missing required parameter in input functionname while creating a lambda using cfn my command fails with the following message api error on request traceback most recent call last file opt code localstack venv lib site packages werkzeug serving py line in run wsgi execute self server app file opt code localstack venv lib site packages werkzeug serving py line in execute application iter app environ start response file opt code localstack venv lib site packages moto server py line in call return backend app environ start response file opt code localstack venv lib site packages flask app py line in call return self wsgi app environ start response file opt code localstack venv lib site packages flask app py line in wsgi app response self handle exception e file opt code localstack venv lib site packages flask cors extension py line in wrapped function return cors after request app make response f args kwargs file opt code localstack venv lib site packages flask app py line in handle exception reraise exc type exc value tb file opt code localstack venv lib site packages flask compat py line in reraise raise value file opt code localstack venv lib site packages flask app py line in wsgi app response self full dispatch request file opt code localstack venv lib site packages flask app py line in full dispatch request rv self handle user exception e file opt code localstack venv lib site packages flask cors extension py line in wrapped function return cors after request app make response f args kwargs file opt code localstack venv lib site packages flask app py line in handle user exception reraise exc type exc value tb file opt code localstack venv lib site packages flask compat py line in reraise raise value file opt code localstack venv lib site packages flask app py line in full dispatch request rv self dispatch request file opt code localstack venv lib site packages flask app py line in dispatch request return self view functions req view args file opt code localstack venv lib site packages moto core utils py line in call result self callback request request url file opt code localstack venv lib site packages moto core responses py line in dispatch return cls dispatch args kwargs file opt code localstack venv lib site packages moto core responses py line in dispatch return self call action file opt code localstack venv lib site packages moto core responses py line in call action response method file opt code localstack venv lib site packages moto core utils py line in wrapper response f args kwargs file opt code localstack venv lib site packages moto cloudformation responses py line in create change set change set type update or create file opt code localstack venv lib site packages moto cloudformation models py line in create change set cross stack resources self exports file opt code localstack venv lib site packages moto cloudformation models py line in init create change set true file opt code localstack venv lib site packages moto cloudformation models py line in init self resource map self create resource map file opt code localstack venv lib site packages moto cloudformation models py line in create resource map resource map create file opt code localstack venv lib site packages moto cloudformation parsing py line in create if isinstance self models file opt code localstack venv lib site packages moto cloudformation parsing py line in getitem resource logical id resource json self self region name file opt code localstack localstack services cloudformation cloudformation starter py line in parse and create resource return parse and create resource logical id resource json resources map region name file opt code localstack localstack services cloudformation cloudformation starter py line in parse and create resource result deploy func logical id resource wrapped stack name stack name file opt code localstack localstack utils cloudformation template deployer py line in deploy resource result deploy resource via sdk function resource id resources resource type func stack name file opt code localstack localstack utils cloudformation template deployer py line in deploy resource via sdk function raise e file opt code localstack localstack utils cloudformation template deployer py line in deploy resource via sdk function result function params file opt code localstack venv lib site packages botocore client py line in api call return self make api call operation name kwargs file opt code localstack venv lib site packages botocore client py line in make api call api params operation model context request context file opt code localstack venv lib site packages botocore client py line in convert to request dict api params operation model file opt code localstack venv lib site packages botocore validate py line in serialize to request raise paramvalidationerror report report generate report botocore exceptions paramvalidationerror parameter validation failed missing required parameter in input functionname considering the fact that the functionname property is not mandatory from cfn perspective it s even worse since having it or not has significant implications is it possible to create a name on the fly
| 1
|
319,617
| 9,747,282,420
|
IssuesEvent
|
2019-06-03 14:05:46
|
dojot/dojot
|
https://api.github.com/repos/dojot/dojot
|
closed
|
[DataBroker, device-manager, history] Real-time does not work on actuate attributes
|
Priority:High Status:In Progress Team:Backend Type:Bug
|
**Steps to reproduce the problem:**
_test 1_:
1. create a flow with attribute of actuation
2. activate the flow

3. view attribute detail

4. After F5

_test 2_:
1. actuate on the device

2. view attribute detail

3. publish data from another attribute


4. After F5

**Affected Version**: 61.1-20190423
|
1.0
|
[DataBroker, device-manager, history] Real-time does not work on actuate attributes - **Steps to reproduce the problem:**
_test 1_:
1. create a flow with attribute of actuation
2. activate the flow

3. view attribute detail

4. After F5

_test 2_:
1. actuate on the device

2. view attribute detail

3. publish data from another attribute


4. After F5

**Affected Version**: 61.1-20190423
|
priority
|
real time does not work on actuate attributes steps to reproduce the problem test create a flow with attribute of actuation activate the flow view attribute detail after test actuate on the device view attribute detail publish data from another attribute after affected version
| 1
|
530,665
| 15,435,525,670
|
IssuesEvent
|
2021-03-07 09:15:45
|
VelvetThePanda/Silk
|
https://api.github.com/repos/VelvetThePanda/Silk
|
closed
|
Daily command for newcomers doesn't save balance correctly
|
Bugged Priority: HIGH
|
**Describe the bug**
If you're new, and you run the daily command, it will say you've collected $500, but in reality it saves 0 dollars. However, if you have an account, it saves correctly.
**To Reproduce**
- Be new
- Run daily command
- Run cash command
- Observe
**Expected behavior**
An account is created, and $500 is deposited.
**Actual behavior**
An account is created, but $0 is saved.
|
1.0
|
Daily command for newcomers doesn't save balance correctly - **Describe the bug**
If you're new, and you run the daily command, it will say you've collected $500, but in reality it saves 0 dollars. However, if you have an account, it saves correctly.
**To Reproduce**
- Be new
- Run daily command
- Run cash command
- Observe
**Expected behavior**
An account is created, and $500 is deposited.
**Actual behavior**
An account is created, but $0 is saved.
|
priority
|
daily command for newcomers doesn t save balance correctly describe the bug if you re new and you run the daily command it will say you ve collected but in reality it saves dollars however if you have an account it saves correctly to reproduce be new run daily command run cash command observe expected behavior an account is created and is deposited actual behavior an account is created but is saved
| 1
|
580,146
| 17,210,729,901
|
IssuesEvent
|
2021-07-19 03:42:53
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
reopened
|
Allow negative learning rates
|
enhancement high priority module: optimizer triage review triaged
|
## 🚀 Feature
Currently, optimizers throw an assertion error when negative learning rates are supplied at construction. This proposal suggests removing this restriction.
## Motivation
Since maximization is equivalent to minimizing a negative loss function, negative learning rates are useful and make sense. For example, in GANs, the generator and discriminator can be trained adversarially by giving the discriminator a negative learning rate. This avoids having two backward passes, which improves both computational efficiency and conceptual clarity.
## Pitch
As learning rates are typically parameterized by constants, providing a negative rate by accident is highly unlikely. I argue that making this mistake is much less likely than wanting a negative learning rate, and so the defensive assertion is better removed.
## Alternatives
It is currently possible to set a negative LR through an ugly loop through the optimizer's `param_group`. An alternative would be to provide a cleaner way to do so, i.e., a `maximize=True` flag on optimizer construction.
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @vincentqb @iramazanli
|
1.0
|
Allow negative learning rates - ## 🚀 Feature
Currently, optimizers throw an assertion error when negative learning rates are supplied at construction. This proposal suggests removing this restriction.
## Motivation
Since maximization is equivalent to minimizing a negative loss function, negative learning rates are useful and make sense. For example, in GANs, the generator and discriminator can be trained adversarially by giving the discriminator a negative learning rate. This avoids having two backward passes, which improves both computational efficiency and conceptual clarity.
## Pitch
As learning rates are typically parameterized by constants, providing a negative rate by accident is highly unlikely. I argue that making this mistake is much less likely than wanting a negative learning rate, and so the defensive assertion is better removed.
## Alternatives
It is currently possible to set a negative LR through an ugly loop through the optimizer's `param_group`. An alternative would be to provide a cleaner way to do so, i.e., a `maximize=True` flag on optimizer construction.
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @vincentqb @iramazanli
|
priority
|
allow negative learning rates 🚀 feature currently optimizers throw an assertion error when negative learning rates are supplied at construction this proposal suggests removing this restriction motivation since maximization is equivalent to minimizing a negative loss function negative learning rates are useful and make sense for example in gans the generator and discriminator can be trained adversarially by giving the discriminator a negative learning rate this avoids having two backward passes which improves both computational efficiency and conceptual clarity pitch as learning rates are typically parameterized by constants providing a negative rate by accident is highly unlikely i argue that making this mistake is much less likely than wanting a negative learning rate and so the defensive assertion is better removed alternatives it is currently possible to set a negative lr through an ugly loop through the optimizer s param group an alternative would be to provide a cleaner way to do so i e a maximize true flag on optimizer construction cc ezyang gchanan bdhirsh jbschlosser vincentqb iramazanli
| 1
|
760,111
| 26,629,051,264
|
IssuesEvent
|
2023-01-24 16:28:26
|
valantic/vue-template
|
https://api.github.com/repos/valantic/vue-template
|
closed
|
Update api documentation
|
enhancement high priority vue-3
|
All attributes on the api helper are documented as required. But most of them are optional (e.g. `[notificationOptions]`). Update documentation and wrap optional attributes in square brackets.
|
1.0
|
Update api documentation - All attributes on the api helper are documented as required. But most of them are optional (e.g. `[notificationOptions]`). Update documentation and wrap optional attributes in square brackets.
|
priority
|
update api documentation all attributes on the api helper are documented as required but most of them are optional e g update documentation and wrap optional attributes in square brackets
| 1
|
423,538
| 12,298,203,675
|
IssuesEvent
|
2020-05-11 10:06:46
|
MyDataTaiwan/mylog14
|
https://api.github.com/repos/MyDataTaiwan/mylog14
|
reopened
|
[Todo] Style adjustment for daily-overview page
|
enhancement priority-high
|
# Issue description
1. Center the Main Header title
2. Fix that the lower part of the countdown animation overlap with the dividing line
|
1.0
|
[Todo] Style adjustment for daily-overview page - # Issue description
1. Center the Main Header title
2. Fix that the lower part of the countdown animation overlap with the dividing line
|
priority
|
style adjustment for daily overview page issue description center the main header title fix that the lower part of the countdown animation overlap with the dividing line
| 1
|
178,636
| 6,613,080,570
|
IssuesEvent
|
2017-09-20 07:51:23
|
OpenWebslides/OpenWebslides
|
https://api.github.com/repos/OpenWebslides/OpenWebslides
|
opened
|
Docker race condition
|
bug high priority operations
|
When restarting Docker/rebooting server, the NGINX container cannot reach the Docker-defined `app` host.
|
1.0
|
Docker race condition - When restarting Docker/rebooting server, the NGINX container cannot reach the Docker-defined `app` host.
|
priority
|
docker race condition when restarting docker rebooting server the nginx container cannot reach the docker defined app host
| 1
|
556,187
| 16,477,192,564
|
IssuesEvent
|
2021-05-24 07:14:51
|
MathiasReker/Delfinen
|
https://api.github.com/repos/MathiasReker/Delfinen
|
closed
|
View Member info
|
feature request high priority required
|
Feature Request
Display Member infomation nicely in terminal.
- [x] The user must choose which member he wants to view a. The User searches on ID or name
* If the user searches on ID, and it exists, the user is directed to the "member view"
* If there is more than one match, the matches are displayed, and the user must choose by picking a corresponding
number
- [x] The system displays member information (Name, Age, etc.)
- [x] The system displays different options
* Update Member info
* Anonymize the member
|
1.0
|
View Member info - Feature Request
Display Member infomation nicely in terminal.
- [x] The user must choose which member he wants to view a. The User searches on ID or name
* If the user searches on ID, and it exists, the user is directed to the "member view"
* If there is more than one match, the matches are displayed, and the user must choose by picking a corresponding
number
- [x] The system displays member information (Name, Age, etc.)
- [x] The system displays different options
* Update Member info
* Anonymize the member
|
priority
|
view member info feature request display member infomation nicely in terminal the user must choose which member he wants to view a the user searches on id or name if the user searches on id and it exists the user is directed to the member view if there is more than one match the matches are displayed and the user must choose by picking a corresponding number the system displays member information name age etc the system displays different options update member info anonymize the member
| 1
|
291,505
| 8,926,061,704
|
IssuesEvent
|
2019-01-22 02:16:03
|
SalvatoreTosti/spicy-bingo
|
https://api.github.com/repos/SalvatoreTosti/spicy-bingo
|
opened
|
Update validation on user-inputted name fields
|
high priority
|
User inputted fields should be checked to ensure they conform to the back-end specifications of the information.
Ex. space checking on new room names.
|
1.0
|
Update validation on user-inputted name fields - User inputted fields should be checked to ensure they conform to the back-end specifications of the information.
Ex. space checking on new room names.
|
priority
|
update validation on user inputted name fields user inputted fields should be checked to ensure they conform to the back end specifications of the information ex space checking on new room names
| 1
|
348,771
| 10,452,589,496
|
IssuesEvent
|
2019-09-19 14:57:06
|
ansible/galaxy
|
https://api.github.com/repos/ansible/galaxy
|
opened
|
Security Updates
|
area/backend priority/high status/new type/bug
|
[ ] Update packages/dependencies to fix failing security audit
[ ] Address the concerns identified by @chrismeyersfsu
|
1.0
|
Security Updates - [ ] Update packages/dependencies to fix failing security audit
[ ] Address the concerns identified by @chrismeyersfsu
|
priority
|
security updates update packages dependencies to fix failing security audit address the concerns identified by chrismeyersfsu
| 1
|
458,086
| 13,168,399,022
|
IssuesEvent
|
2020-08-11 12:03:46
|
kubesphere/kubesphere
|
https://api.github.com/repos/kubesphere/kubesphere
|
closed
|
custom project role permission denied
|
area/iam kind/bug kind/need-to-verify priority/high
|
**Describe the Bug**
A custom project role with `Application Workloads View / Application Workloads Management` configured HPA but denied


**Versions Used**
KubeSphere: 3.0.0 (2020-08-11)
|
1.0
|
custom project role permission denied - **Describe the Bug**
A custom project role with `Application Workloads View / Application Workloads Management` configured HPA but denied


**Versions Used**
KubeSphere: 3.0.0 (2020-08-11)
|
priority
|
custom project role permission denied describe the bug a custom project role with application workloads view application workloads management configured hpa but denied versions used kubesphere
| 1
|
425,412
| 12,339,932,104
|
IssuesEvent
|
2020-05-14 19:00:13
|
canonical-web-and-design/jaas-dashboard
|
https://api.github.com/repos/canonical-web-and-design/jaas-dashboard
|
closed
|
Empty table warning is partly in header
|
Bug 🐛 Models Listing Priority: High
|

To reproduce, run the dashboard with no models.
|
1.0
|
Empty table warning is partly in header - 
To reproduce, run the dashboard with no models.
|
priority
|
empty table warning is partly in header to reproduce run the dashboard with no models
| 1
|
631,396
| 20,151,422,599
|
IssuesEvent
|
2022-02-09 12:47:28
|
Blosc/caterva
|
https://api.github.com/repos/Blosc/caterva
|
closed
|
Implement a resize functionality
|
enhancement high priority
|
This would allow to extend/shrink an array in different dimensions. I suggest a new function with a signature similar to this:
```
/**
* @brief Resize a caterva array
*
* Changes the shape of the caterva array by growing or shrinking one or more dimensions.
*
* @param ctx The caterva context to be used.
* @param array The caterva array.
* @param new_dims New dimensions of the array.
*
* @return An error code
*/
int caterva_resize(caterva_ctx_t *ctx, caterva_array_t *array, int *new_dims);
```
|
1.0
|
Implement a resize functionality - This would allow to extend/shrink an array in different dimensions. I suggest a new function with a signature similar to this:
```
/**
* @brief Resize a caterva array
*
* Changes the shape of the caterva array by growing or shrinking one or more dimensions.
*
* @param ctx The caterva context to be used.
* @param array The caterva array.
* @param new_dims New dimensions of the array.
*
* @return An error code
*/
int caterva_resize(caterva_ctx_t *ctx, caterva_array_t *array, int *new_dims);
```
|
priority
|
implement a resize functionality this would allow to extend shrink an array in different dimensions i suggest a new function with a signature similar to this brief resize a caterva array changes the shape of the caterva array by growing or shrinking one or more dimensions param ctx the caterva context to be used param array the caterva array param new dims new dimensions of the array return an error code int caterva resize caterva ctx t ctx caterva array t array int new dims
| 1
|
659,192
| 21,919,302,097
|
IssuesEvent
|
2022-05-22 10:20:07
|
OpenRefine/OpenRefine
|
https://api.github.com/repos/OpenRefine/OpenRefine
|
closed
|
Snapshot releases (Mac and Linux versions) on MacOS are not properly loading
|
bug priority: High packaging
|
When opening the Mac OS (.dmg) snapshot release, it doesn't load OpenRefine any further than the logo, as seen in this screenshot:

When I try with the Linux download, that one doesn't even open a new browser tab at all.
### Expected Behavior
OpenRefine should fully load.
### Context
Terminal log when trying to start the MacOS .dmg (pretty minimal):
```
Sandras-MacBook-Air:~ fokky$ /Applications/OpenRefine\ snapshot\ 20220408.app/Contents/MacOS/JavaAppLauncher ; exit;
08:15:29.649 [ refine_server] Starting Server bound to '127.0.0.1:3333' (0ms)
08:15:29.717 [ refine_server] Initializing context: '/' from '/Applications/OpenRefine snapshot 20220408.app/Contents/Resources/webapp' (68ms)
08:15:31.931 [ refine] Starting OpenRefine 3.6-SNAPSHOT [0ba3b92]... (2214ms)
08:15:31.937 [ refine] initializing FileProjectManager with dir (6ms)
08:15:31.937 [ refine] /Users/fokky/Library/Application Support/OpenRefine (0ms)
```
When I try to open the Linux version of the snapshot release, that one doesn't even start up a new browser tab. Terminal log for that one is:
```
miniMac:~ fokky$ /Users/fokky/Desktop/openrefine-3.6-SNAPSHOT\ 3/refine ; exit;
Using refine.ini for configuration
cat: refine.ini: No such file or directory
-------------------------------------------------------------------------------------------------
You have 812M of free memory.
Your current configuration is set to use 1024M of memory.
OpenRefine can run better when given more memory. Read our FAQ on how to allocate more memory here:
https://docs.openrefine.org/manual/installing#increasing-memory-allocation
-------------------------------------------------------------------------------------------------
ls: server/target/lib: No such file or directory
\nCould not find Maven locally, starting download for Maven ...
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 8469k 100 8469k 0 0 10.7M 0 --:--:-- --:--:-- --:--:-- 10.7M
/Users/fokky/Desktop/openrefine-3.6-SNAPSHOT 3/refine: line 294: cd: main/webapp: No such file or directory
[INFO] Scanning for projects...
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 0.300 s
[INFO] Finished at: 2022-04-10T09:26:43+02:00
[INFO] ------------------------------------------------------------------------
[ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException
[INFO] Scanning for projects...
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 0.426 s
[INFO] Finished at: 2022-04-10T09:26:47+02:00
[INFO] ------------------------------------------------------------------------
[ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException
Error: Could not find or load main class com.google.refine.Refine
Caused by: java.lang.ClassNotFoundException: com.google.refine.Refine
logout
Saving session...
...copying shared history...
...saving history...truncating history files...
...completed.
[Process completed]
```
### Versions<!-- (please complete the following information)-->
- Operating System: MacOS 10.13.6, MacOS 12.2.1
- Browser Version: Chrome
- JRE or JDK Version: n/a?
- OpenRefine: 3.6 snapshot release
### Files used
I just tried with the April 8 and April 10 snapshot releases, but the problem has been around for longer (several weeks?).
I heard from a Linux user (via the Structured Data on Commons Telegram group) that they have the same issue of OpenRefine snapshot releases not loading for them.
|
1.0
|
Snapshot releases (Mac and Linux versions) on MacOS are not properly loading - When opening the Mac OS (.dmg) snapshot release, it doesn't load OpenRefine any further than the logo, as seen in this screenshot:

When I try with the Linux download, that one doesn't even open a new browser tab at all.
### Expected Behavior
OpenRefine should fully load.
### Context
Terminal log when trying to start the MacOS .dmg (pretty minimal):
```
Sandras-MacBook-Air:~ fokky$ /Applications/OpenRefine\ snapshot\ 20220408.app/Contents/MacOS/JavaAppLauncher ; exit;
08:15:29.649 [ refine_server] Starting Server bound to '127.0.0.1:3333' (0ms)
08:15:29.717 [ refine_server] Initializing context: '/' from '/Applications/OpenRefine snapshot 20220408.app/Contents/Resources/webapp' (68ms)
08:15:31.931 [ refine] Starting OpenRefine 3.6-SNAPSHOT [0ba3b92]... (2214ms)
08:15:31.937 [ refine] initializing FileProjectManager with dir (6ms)
08:15:31.937 [ refine] /Users/fokky/Library/Application Support/OpenRefine (0ms)
```
When I try to open the Linux version of the snapshot release, that one doesn't even start up a new browser tab. Terminal log for that one is:
```
miniMac:~ fokky$ /Users/fokky/Desktop/openrefine-3.6-SNAPSHOT\ 3/refine ; exit;
Using refine.ini for configuration
cat: refine.ini: No such file or directory
-------------------------------------------------------------------------------------------------
You have 812M of free memory.
Your current configuration is set to use 1024M of memory.
OpenRefine can run better when given more memory. Read our FAQ on how to allocate more memory here:
https://docs.openrefine.org/manual/installing#increasing-memory-allocation
-------------------------------------------------------------------------------------------------
ls: server/target/lib: No such file or directory
\nCould not find Maven locally, starting download for Maven ...
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 8469k 100 8469k 0 0 10.7M 0 --:--:-- --:--:-- --:--:-- 10.7M
/Users/fokky/Desktop/openrefine-3.6-SNAPSHOT 3/refine: line 294: cd: main/webapp: No such file or directory
[INFO] Scanning for projects...
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 0.300 s
[INFO] Finished at: 2022-04-10T09:26:43+02:00
[INFO] ------------------------------------------------------------------------
[ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException
[INFO] Scanning for projects...
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 0.426 s
[INFO] Finished at: 2022-04-10T09:26:47+02:00
[INFO] ------------------------------------------------------------------------
[ERROR] The goal you specified requires a project to execute but there is no POM in this directory (/Users/fokky/Desktop). Please verify you invoked Maven from the correct directory. -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MissingProjectException
Error: Could not find or load main class com.google.refine.Refine
Caused by: java.lang.ClassNotFoundException: com.google.refine.Refine
logout
Saving session...
...copying shared history...
...saving history...truncating history files...
...completed.
[Process completed]
```
### Versions<!-- (please complete the following information)-->
- Operating System: MacOS 10.13.6, MacOS 12.2.1
- Browser Version: Chrome
- JRE or JDK Version: n/a?
- OpenRefine: 3.6 snapshot release
### Files used
I just tried with the April 8 and April 10 snapshot releases, but the problem has been around for longer (several weeks?).
I heard from a Linux user (via the Structured Data on Commons Telegram group) that they have the same issue of OpenRefine snapshot releases not loading for them.
|
priority
|
snapshot releases mac and linux versions on macos are not properly loading when opening the mac os dmg snapshot release it doesn t load openrefine any further than the logo as seen in this screenshot when i try with the linux download that one doesn t even open a new browser tab at all expected behavior openrefine should fully load context terminal log when trying to start the macos dmg pretty minimal sandras macbook air fokky applications openrefine snapshot app contents macos javaapplauncher exit starting server bound to initializing context from applications openrefine snapshot app contents resources webapp starting openrefine snapshot initializing fileprojectmanager with dir users fokky library application support openrefine when i try to open the linux version of the snapshot release that one doesn t even start up a new browser tab terminal log for that one is minimac fokky users fokky desktop openrefine snapshot refine exit using refine ini for configuration cat refine ini no such file or directory you have of free memory your current configuration is set to use of memory openrefine can run better when given more memory read our faq on how to allocate more memory here ls server target lib no such file or directory ncould not find maven locally starting download for maven total received xferd average speed time time time current dload upload total spent left speed users fokky desktop openrefine snapshot refine line cd main webapp no such file or directory scanning for projects build failure total time s finished at the goal you specified requires a project to execute but there is no pom in this directory users fokky desktop please verify you invoked maven from the correct directory to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles scanning for projects build failure total time s finished at the goal you specified requires a project to execute but there is no pom in this directory users fokky desktop please verify you invoked maven from the correct directory to see the full stack trace of the errors re run maven with the e switch re run maven using the x switch to enable full debug logging for more information about the errors and possible solutions please read the following articles error could not find or load main class com google refine refine caused by java lang classnotfoundexception com google refine refine logout saving session copying shared history saving history truncating history files completed versions operating system macos macos browser version chrome jre or jdk version n a openrefine snapshot release files used i just tried with the april and april snapshot releases but the problem has been around for longer several weeks i heard from a linux user via the structured data on commons telegram group that they have the same issue of openrefine snapshot releases not loading for them
| 1
|
510,546
| 14,792,652,967
|
IssuesEvent
|
2021-01-12 15:00:06
|
staxrip/staxrip
|
https://api.github.com/repos/staxrip/staxrip
|
closed
|
Staxrip will try to append CHUNKS even if chunks = 1
|
added/fixed/done bug priority high
|
**BUG:**
When there are files called file_chunk2.hevc and file_chunk3.hevc (...) in the temp folder, Staxrip will attempt to append them when muxing, even if chunk encoding is not activated (i.e chunks =1).
These files could happen to be there due to a previous encoding.
**Expected behaviour**
If chunks = 1 then those files should be ignored, and muxing should not append them.
Muxing should work SMARTLY.
If chunks = 1 , look for file_out.hevc, ignore any other
If chunks = 2 look for file_out.hevc and file_chunk2.hevc, ignore any other
If chunks = 3 look for file_out.hevc and file_chunk2.hevc and file_chunk3.hevc, ignore any other
Here is the complete log for your pleasure: it shows that NO second chunk has been encoded, but muxer has appended a chunk from a previous processing.
```
------------------------- System Environment -------------------------
StaxRip : 2.1.7.1
Windows : Windows 10 Home 2004
Language : English (United States)
CPU : Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz
GPU : Intel(R) HD Graphics 530, NVIDIA GeForce GTX 960M
Resolution : 1920 x 1080
DPI : 96
----------------------- Media Info Source File -----------------------
D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4
General
Complete name : D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4
Format : MPEG-4
Format profile : Base Media / Version 2
Codec ID : mp42 (isom/mp42)
File size : 4.88 MiB
Duration : 1 min 18 s
Overall bit rate mode : Variable
Overall bit rate : 522 kb/s
Encoded date : UTC 2017-12-17 15:18:14
Tagged date : UTC 2017-12-17 15:18:14
gsst : 0
gstd : 78460
Video
ID : 1
Format : AVC
Format/Info : Advanced Video Codec
Format profile : High@L3.1
Format settings : CABAC / 1 Ref Frames
Format, CABAC : Yes
Format, Reference frames : 1 frame
Format, GOP : M=1, N=60
Codec ID : avc1
Codec ID/Info : Advanced Video Coding
Duration : 1 min 18 s
Bit rate : 327 kb/s
Maximum bit rate : 471 kb/s
Width : 1 280 pixels
Height : 720 pixels
Display aspect ratio : 16:9
Frame rate mode : Constant
Frame rate : 25.000 FPS
Color space : YUV
Chroma subsampling : 4:2:0
Bit depth : 8 bits
Scan type : Progressive
Bits/(Pixel*Frame) : 0.014
Stream size : 3.06 MiB (63%)
Tagged date : UTC 2017-12-17 15:18:15
Color range : Limited
Color primaries : BT.709
Transfer characteristics : BT.709
Matrix coefficients : BT.709
Codec configuration box : avcC
Audio
ID : 2
Format : AAC LC
Format/Info : Advanced Audio Codec Low Complexity
Codec ID : mp4a-40-2
Duration : 1 min 18 s
Bit rate mode : Variable
Bit rate : 192 kb/s
Maximum bit rate : 201 kb/s
Channel(s) : 2 channels
Channel layout : L R
Sampling rate : 44.1 kHz
Frame rate : 43.066 FPS (1024 SPF)
Compression mode : Lossy
Stream size : 1.79 MiB (37%)
Title : IsoMedia File Produced by Google, 5-11-2011
Encoded date : UTC 2017-12-17 15:18:15
Tagged date : UTC 2017-12-17 15:18:15
----------------------------- Demux audio -----------------------------
MP4Box 1.1.0-DEV-rev390-g4228658a9-x64-gcc10.2.0 Patman86
"C:\Program Files\StaxRip\Apps\Support\MP4Box\MP4Box.exe" -single 2 -out "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4"
Start: 11:12:51
End: 11:12:51
Duration: 00:00:00
General
Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a
Format : MPEG-4
Format profile : Base Media
Codec ID : isom (isom)
File size : 1.81 MiB
Duration : 1 min 18 s
Overall bit rate mode : Variable
Overall bit rate : 194 kb/s
Encoded date : UTC 2021-01-11 09:12:51
Tagged date : UTC 2021-01-11 09:12:51
Audio
ID : 2
Format : AAC LC
Format/Info : Advanced Audio Codec Low Complexity
Format profile : AAC@L2
Codec ID : mp4a-40-2
Duration : 1 min 18 s
Bit rate mode : Variable
Bit rate : 192 kb/s
Nominal bit rate : 728 b/s
Maximum bit rate : 201 kb/s
Channel(s) : 2 channels
Channel layout : L R
Sampling rate : 44.1 kHz
Frame rate : 43.066 FPS (1024 SPF)
Compression mode : Lossy
Stream size : 1.79 MiB (99%)
Encoded date : UTC 2021-01-11 09:12:51
Tagged date : UTC 2021-01-11 09:12:51
---------------------- Indexing using ffmsindex ----------------------
"C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffmsindex.exe" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex"
Writing index... done.
Start: 11:12:52
End: 11:12:52
Duration: 00:00:00
---------------------------- Configuration ----------------------------
Template : ZK x265 480p BT709 Base
Video Encoder Profile : x265
Container/Muxer Profile : MKV (mkvmerge)
--------------------------- AviSynth Script ---------------------------
AddAutoloadDir("C:\Program Files\StaxRip\Apps\FrameServer\AviSynth\plugins")
AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\AviSynth")
AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\Dual")
LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffms2.dll")
LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\AVS\JPSDR\Plugins_JPSDR.dll")
FFVideoSource("D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4", cachefile="C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex")
#AssumeFPS(25)
LanczosResizeMT(856, 480, prefetch=4)
Trim(100, 145) + Trim(1375, 1420)
------------------------- Source Script Info -------------------------
Width : 1280
Height : 720
Frames : 1960
Time : 01:18.400
Framerate : 25 (25/1)
Format : YUV420P8
------------------------- Target Script Info -------------------------
Width : 856
Height : 480
Frames : 92
Time : 00:03.680
Framerate : 25 (25/1)
Format : YUV420P8
--------------------------- Video encoding ---------------------------
x265 M-3.4+35-772bb4c84-x64-gcc10.2.0 Patman86
"C:\Program Files\StaxRip\Apps\Encoders\x265\x265.exe" --crf 24 --preset slow --output-depth 10 --output "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps.avs"
avs+ [info]: AviSynth+ 3.6.2 (r3341, master, x86_64)
avs+ [info]: Video colorspace: YUV420 (YV12)
avs+ [info]: Video depth: 8
avs+ [info]: Video resolution: 856x480
avs+ [info]: Video framerate: 25/1
avs+ [info]: Video framecount: 92
avs+ [info]: 856x480 fps 25/1 i420p8 frames 0 - 91 of 92
raw [info]: output file: C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc
x265 [info]: HEVC encoder version x265M 3.4+35-772bb4c84
x265 [info]: build info [Windows][MSVC 1928][64 bit] 10bit
x265 [info]: using cpu capabilities: MMX2 SSE2Fast LZCNT SSSE3 SSE4.2 AVX FMA3 BMI2 AVX2
x265 [info]: Main 10 profile, Level-3 (Main tier)
x265 [info]: Thread pool created using 8 threads
x265 [info]: Slices : 1
x265 [info]: frame threads / pool features : 3 / wpp(8 rows)
x265 [warning]: Source height < 720p; disabling lookahead-slices
x265 [info]: Coding QT: max CU size, min CU size : 64 / 8
x265 [info]: Residual QT: max TU size, max depth : 32 / 1 inter / 1 intra
x265 [info]: ME / range / subpel / merge : star / 57 / 3 / 3
x265 [info]: Keyframe min / max / scenecut / bias : 25 / 250 / 40 / 5.00
x265 [info]: Lookahead / bframes / badapt : 25 / 4 / 2
x265 [info]: b-pyramid / weightp / weightb : 1 / 1 / 0
x265 [info]: References / ref-limit cu / depth : 4 / on / on
x265 [info]: AQ: mode / str / qg-size / cu-tree : 2 / 1.0 / 32 / 1
x265 [info]: Rate Control / qCompress : CRF-24.0 / 0.60
x265 [info]: tools: rect limit-modes rd=4 psy-rd=2.00 rdoq=2 psy-rdoq=1.00
x265 [info]: tools: rskip mode=1 signhide tmvp strong-intra-smoothing deblock
x265 [info]: tools: sao
x265 [info]: frame I: 1, Avg QP:22.44 kb/s: 1123.60
x265 [info]: frame P: 20, Avg QP:27.63 kb/s: 119.84
x265 [info]: frame B: 71, Avg QP:31.24 kb/s: 58.88
x265 [info]: Weighted P-Frames: Y:0.0% UV:0.0%
x265 [info]: consecutive B-frames: 9.5% 4.8% 4.8% 0.0% 81.0%
encoded 92 frames in 1.98s (46.46 fps), 83.71 kb/s, Avg QP:30.36
Start: 23:13:42
End: 23:13:45
Duration: 00:00:02
General
Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc
Format : HEVC
Format/Info : High Efficiency Video Coding
File size : 40.3 KiB
Duration : 640 ms
Overall bit rate : 516 kb/s
Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit
Video
Format : HEVC
Format/Info : High Efficiency Video Coding
Format profile : Main 10@L3@Main
Duration : 640 ms
Width : 856 pixels
Height : 480 pixels
Display aspect ratio : 16:9
Frame rate : 25.000 FPS
Color space : YUV
Chroma subsampling : 4:2:0
Bit depth : 10 bits
Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit
---------------------------- Muxing to MKV ----------------------------
mkvmerge 52
"C:\Program Files\StaxRip\Apps\Support\MKVToolNix\mkvmerge.exe" -o "R:\Timecode sample - 25fps.mkv" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" + "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc" --ui-language en --title ""
mkvmerge v52.0.0 ('Secret For The Mad') 64-bit
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc': Using the demultiplexer for the format 'HEVC/H.265'.
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc': Using the demultiplexer for the format 'HEVC/H.265'.
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'.
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'.
No append mapping was given for the file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc'). A default mapping of 1:0:0:0 will be used instead. Please keep that in mind if mkvmerge aborts with an error message regarding invalid '--append-to' options.
The file 'R:\Timecode sample - 25fps.mkv' has been opened for writing.
Appending track 0 from file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc') to track 0 from file no. 0 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc').
The cue entries (the index) are being written...
Multiplexing took 0 seconds.
Start: 23:13:45
End: 23:13:45
Duration: 00:00:00
General
Complete name : R:\Timecode sample - 25fps.mkv
Format : Matroska
Format version : Version 4
File size : 73.1 KiB
Duration : 5 s 520 ms
Overall bit rate : 108 kb/s
Encoded date : UTC 2021-01-11 21:13:45
Writing application : mkvmerge v52.0.0 ('Secret For The Mad') 64-bit
Writing library : libebml v1.4.1 + libmatroska v1.6.2
Video
ID : 1
Format : HEVC
Format/Info : High Efficiency Video Coding
Format profile : Main 10@L3@Main
Codec ID : V_MPEGH/ISO/HEVC
Duration : 5 s 520 ms
Bit rate : 95.0 kb/s
Width : 856 pixels
Height : 480 pixels
Display aspect ratio : 16:9
Frame rate mode : Constant
Frame rate : 25.000 FPS
Color space : YUV
Chroma subsampling : 4:2:0
Bit depth : 10 bits
Bits/(Pixel*Frame) : 0.009
Stream size : 64.0 KiB (88%)
Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit
Default : Yes
Forced : No
---------------------------- Job Complete ----------------------------
Start: 23:13:42
End: 23:13:45
Duration: 00:00:03
```
|
1.0
|
Staxrip will try to append CHUNKS even if chunks = 1 - **BUG:**
When there are files called file_chunk2.hevc and file_chunk3.hevc (...) in the temp folder, Staxrip will attempt to append them when muxing, even if chunk encoding is not activated (i.e chunks =1).
These files could happen to be there due to a previous encoding.
**Expected behaviour**
If chunks = 1 then those files should be ignored, and muxing should not append them.
Muxing should work SMARTLY.
If chunks = 1 , look for file_out.hevc, ignore any other
If chunks = 2 look for file_out.hevc and file_chunk2.hevc, ignore any other
If chunks = 3 look for file_out.hevc and file_chunk2.hevc and file_chunk3.hevc, ignore any other
Here is the complete log for your pleasure: it shows that NO second chunk has been encoded, but muxer has appended a chunk from a previous processing.
```
------------------------- System Environment -------------------------
StaxRip : 2.1.7.1
Windows : Windows 10 Home 2004
Language : English (United States)
CPU : Intel(R) Core(TM) i7-6700HQ CPU @ 2.60GHz
GPU : Intel(R) HD Graphics 530, NVIDIA GeForce GTX 960M
Resolution : 1920 x 1080
DPI : 96
----------------------- Media Info Source File -----------------------
D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4
General
Complete name : D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4
Format : MPEG-4
Format profile : Base Media / Version 2
Codec ID : mp42 (isom/mp42)
File size : 4.88 MiB
Duration : 1 min 18 s
Overall bit rate mode : Variable
Overall bit rate : 522 kb/s
Encoded date : UTC 2017-12-17 15:18:14
Tagged date : UTC 2017-12-17 15:18:14
gsst : 0
gstd : 78460
Video
ID : 1
Format : AVC
Format/Info : Advanced Video Codec
Format profile : High@L3.1
Format settings : CABAC / 1 Ref Frames
Format, CABAC : Yes
Format, Reference frames : 1 frame
Format, GOP : M=1, N=60
Codec ID : avc1
Codec ID/Info : Advanced Video Coding
Duration : 1 min 18 s
Bit rate : 327 kb/s
Maximum bit rate : 471 kb/s
Width : 1 280 pixels
Height : 720 pixels
Display aspect ratio : 16:9
Frame rate mode : Constant
Frame rate : 25.000 FPS
Color space : YUV
Chroma subsampling : 4:2:0
Bit depth : 8 bits
Scan type : Progressive
Bits/(Pixel*Frame) : 0.014
Stream size : 3.06 MiB (63%)
Tagged date : UTC 2017-12-17 15:18:15
Color range : Limited
Color primaries : BT.709
Transfer characteristics : BT.709
Matrix coefficients : BT.709
Codec configuration box : avcC
Audio
ID : 2
Format : AAC LC
Format/Info : Advanced Audio Codec Low Complexity
Codec ID : mp4a-40-2
Duration : 1 min 18 s
Bit rate mode : Variable
Bit rate : 192 kb/s
Maximum bit rate : 201 kb/s
Channel(s) : 2 channels
Channel layout : L R
Sampling rate : 44.1 kHz
Frame rate : 43.066 FPS (1024 SPF)
Compression mode : Lossy
Stream size : 1.79 MiB (37%)
Title : IsoMedia File Produced by Google, 5-11-2011
Encoded date : UTC 2017-12-17 15:18:15
Tagged date : UTC 2017-12-17 15:18:15
----------------------------- Demux audio -----------------------------
MP4Box 1.1.0-DEV-rev390-g4228658a9-x64-gcc10.2.0 Patman86
"C:\Program Files\StaxRip\Apps\Support\MP4Box\MP4Box.exe" -single 2 -out "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4"
Start: 11:12:51
End: 11:12:51
Duration: 00:00:00
General
Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\ID1 {IsoMedia File Produced by Google, 5-11-2011}.m4a
Format : MPEG-4
Format profile : Base Media
Codec ID : isom (isom)
File size : 1.81 MiB
Duration : 1 min 18 s
Overall bit rate mode : Variable
Overall bit rate : 194 kb/s
Encoded date : UTC 2021-01-11 09:12:51
Tagged date : UTC 2021-01-11 09:12:51
Audio
ID : 2
Format : AAC LC
Format/Info : Advanced Audio Codec Low Complexity
Format profile : AAC@L2
Codec ID : mp4a-40-2
Duration : 1 min 18 s
Bit rate mode : Variable
Bit rate : 192 kb/s
Nominal bit rate : 728 b/s
Maximum bit rate : 201 kb/s
Channel(s) : 2 channels
Channel layout : L R
Sampling rate : 44.1 kHz
Frame rate : 43.066 FPS (1024 SPF)
Compression mode : Lossy
Stream size : 1.79 MiB (99%)
Encoded date : UTC 2021-01-11 09:12:51
Tagged date : UTC 2021-01-11 09:12:51
---------------------- Indexing using ffmsindex ----------------------
"C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffmsindex.exe" "D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex"
Writing index... done.
Start: 11:12:52
End: 11:12:52
Duration: 00:00:00
---------------------------- Configuration ----------------------------
Template : ZK x265 480p BT709 Base
Video Encoder Profile : x265
Container/Muxer Profile : MKV (mkvmerge)
--------------------------- AviSynth Script ---------------------------
AddAutoloadDir("C:\Program Files\StaxRip\Apps\FrameServer\AviSynth\plugins")
AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\AviSynth")
AddAutoloadDir("C:\Program Files\Staxrip\Settings\Plugins\Dual")
LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\Dual\ffms2\ffms2.dll")
LoadPlugin("C:\Program Files\StaxRip\Apps\Plugins\AVS\JPSDR\Plugins_JPSDR.dll")
FFVideoSource("D:\Downloads\_IDM Downloads\Timecode sample - 25fps.mp4", cachefile="C:\Temp\_StaxRip\Timecode sample - 25fps_temp\temp.ffindex")
#AssumeFPS(25)
LanczosResizeMT(856, 480, prefetch=4)
Trim(100, 145) + Trim(1375, 1420)
------------------------- Source Script Info -------------------------
Width : 1280
Height : 720
Frames : 1960
Time : 01:18.400
Framerate : 25 (25/1)
Format : YUV420P8
------------------------- Target Script Info -------------------------
Width : 856
Height : 480
Frames : 92
Time : 00:03.680
Framerate : 25 (25/1)
Format : YUV420P8
--------------------------- Video encoding ---------------------------
x265 M-3.4+35-772bb4c84-x64-gcc10.2.0 Patman86
"C:\Program Files\StaxRip\Apps\Encoders\x265\x265.exe" --crf 24 --preset slow --output-depth 10 --output "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps.avs"
avs+ [info]: AviSynth+ 3.6.2 (r3341, master, x86_64)
avs+ [info]: Video colorspace: YUV420 (YV12)
avs+ [info]: Video depth: 8
avs+ [info]: Video resolution: 856x480
avs+ [info]: Video framerate: 25/1
avs+ [info]: Video framecount: 92
avs+ [info]: 856x480 fps 25/1 i420p8 frames 0 - 91 of 92
raw [info]: output file: C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc
x265 [info]: HEVC encoder version x265M 3.4+35-772bb4c84
x265 [info]: build info [Windows][MSVC 1928][64 bit] 10bit
x265 [info]: using cpu capabilities: MMX2 SSE2Fast LZCNT SSSE3 SSE4.2 AVX FMA3 BMI2 AVX2
x265 [info]: Main 10 profile, Level-3 (Main tier)
x265 [info]: Thread pool created using 8 threads
x265 [info]: Slices : 1
x265 [info]: frame threads / pool features : 3 / wpp(8 rows)
x265 [warning]: Source height < 720p; disabling lookahead-slices
x265 [info]: Coding QT: max CU size, min CU size : 64 / 8
x265 [info]: Residual QT: max TU size, max depth : 32 / 1 inter / 1 intra
x265 [info]: ME / range / subpel / merge : star / 57 / 3 / 3
x265 [info]: Keyframe min / max / scenecut / bias : 25 / 250 / 40 / 5.00
x265 [info]: Lookahead / bframes / badapt : 25 / 4 / 2
x265 [info]: b-pyramid / weightp / weightb : 1 / 1 / 0
x265 [info]: References / ref-limit cu / depth : 4 / on / on
x265 [info]: AQ: mode / str / qg-size / cu-tree : 2 / 1.0 / 32 / 1
x265 [info]: Rate Control / qCompress : CRF-24.0 / 0.60
x265 [info]: tools: rect limit-modes rd=4 psy-rd=2.00 rdoq=2 psy-rdoq=1.00
x265 [info]: tools: rskip mode=1 signhide tmvp strong-intra-smoothing deblock
x265 [info]: tools: sao
x265 [info]: frame I: 1, Avg QP:22.44 kb/s: 1123.60
x265 [info]: frame P: 20, Avg QP:27.63 kb/s: 119.84
x265 [info]: frame B: 71, Avg QP:31.24 kb/s: 58.88
x265 [info]: Weighted P-Frames: Y:0.0% UV:0.0%
x265 [info]: consecutive B-frames: 9.5% 4.8% 4.8% 0.0% 81.0%
encoded 92 frames in 1.98s (46.46 fps), 83.71 kb/s, Avg QP:30.36
Start: 23:13:42
End: 23:13:45
Duration: 00:00:02
General
Complete name : C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc
Format : HEVC
Format/Info : High Efficiency Video Coding
File size : 40.3 KiB
Duration : 640 ms
Overall bit rate : 516 kb/s
Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit
Video
Format : HEVC
Format/Info : High Efficiency Video Coding
Format profile : Main 10@L3@Main
Duration : 640 ms
Width : 856 pixels
Height : 480 pixels
Display aspect ratio : 16:9
Frame rate : 25.000 FPS
Color space : YUV
Chroma subsampling : 4:2:0
Bit depth : 10 bits
Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit
---------------------------- Muxing to MKV ----------------------------
mkvmerge 52
"C:\Program Files\StaxRip\Apps\Support\MKVToolNix\mkvmerge.exe" -o "R:\Timecode sample - 25fps.mkv" "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc" + "C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc" --ui-language en --title ""
mkvmerge v52.0.0 ('Secret For The Mad') 64-bit
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc': Using the demultiplexer for the format 'HEVC/H.265'.
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc': Using the demultiplexer for the format 'HEVC/H.265'.
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'.
'C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc' track 0: Using the output module for the format 'HEVC/H.265 (unframed)'.
No append mapping was given for the file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc'). A default mapping of 1:0:0:0 will be used instead. Please keep that in mind if mkvmerge aborts with an error message regarding invalid '--append-to' options.
The file 'R:\Timecode sample - 25fps.mkv' has been opened for writing.
Appending track 0 from file no. 1 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out_chunk2.hevc') to track 0 from file no. 0 ('C:\Temp\_StaxRip\Timecode sample - 25fps_temp\Timecode sample - 25fps_out.hevc').
The cue entries (the index) are being written...
Multiplexing took 0 seconds.
Start: 23:13:45
End: 23:13:45
Duration: 00:00:00
General
Complete name : R:\Timecode sample - 25fps.mkv
Format : Matroska
Format version : Version 4
File size : 73.1 KiB
Duration : 5 s 520 ms
Overall bit rate : 108 kb/s
Encoded date : UTC 2021-01-11 21:13:45
Writing application : mkvmerge v52.0.0 ('Secret For The Mad') 64-bit
Writing library : libebml v1.4.1 + libmatroska v1.6.2
Video
ID : 1
Format : HEVC
Format/Info : High Efficiency Video Coding
Format profile : Main 10@L3@Main
Codec ID : V_MPEGH/ISO/HEVC
Duration : 5 s 520 ms
Bit rate : 95.0 kb/s
Width : 856 pixels
Height : 480 pixels
Display aspect ratio : 16:9
Frame rate mode : Constant
Frame rate : 25.000 FPS
Color space : YUV
Chroma subsampling : 4:2:0
Bit depth : 10 bits
Bits/(Pixel*Frame) : 0.009
Stream size : 64.0 KiB (88%)
Writing library : x265M - 3.4+35-772bb4c84:[Windows][MSVC 1928][64 bit] 10bit
Default : Yes
Forced : No
---------------------------- Job Complete ----------------------------
Start: 23:13:42
End: 23:13:45
Duration: 00:00:03
```
|
priority
|
staxrip will try to append chunks even if chunks bug when there are files called file hevc and file hevc in the temp folder staxrip will attempt to append them when muxing even if chunk encoding is not activated i e chunks these files could happen to be there due to a previous encoding expected behaviour if chunks then those files should be ignored and muxing should not append them muxing should work smartly if chunks look for file out hevc ignore any other if chunks look for file out hevc and file hevc ignore any other if chunks look for file out hevc and file hevc and file hevc ignore any other here is the complete log for your pleasure it shows that no second chunk has been encoded but muxer has appended a chunk from a previous processing system environment staxrip windows windows home language english united states cpu intel r core tm cpu gpu intel r hd graphics nvidia geforce gtx resolution x dpi media info source file d downloads idm downloads timecode sample general complete name d downloads idm downloads timecode sample format mpeg format profile base media version codec id isom file size mib duration min s overall bit rate mode variable overall bit rate kb s encoded date utc tagged date utc gsst gstd video id format avc format info advanced video codec format profile high format settings cabac ref frames format cabac yes format reference frames frame format gop m n codec id codec id info advanced video coding duration min s bit rate kb s maximum bit rate kb s width pixels height pixels display aspect ratio frame rate mode constant frame rate fps color space yuv chroma subsampling bit depth bits scan type progressive bits pixel frame stream size mib tagged date utc color range limited color primaries bt transfer characteristics bt matrix coefficients bt codec configuration box avcc audio id format aac lc format info advanced audio codec low complexity codec id duration min s bit rate mode variable bit rate kb s maximum bit rate kb s channel s channels channel layout l r sampling rate khz frame rate fps spf compression mode lossy stream size mib title isomedia file produced by google encoded date utc tagged date utc demux audio dev c program files staxrip apps support exe single out c temp staxrip timecode sample temp isomedia file produced by google d downloads idm downloads timecode sample start end duration general complete name c temp staxrip timecode sample temp isomedia file produced by google format mpeg format profile base media codec id isom isom file size mib duration min s overall bit rate mode variable overall bit rate kb s encoded date utc tagged date utc audio id format aac lc format info advanced audio codec low complexity format profile aac codec id duration min s bit rate mode variable bit rate kb s nominal bit rate b s maximum bit rate kb s channel s channels channel layout l r sampling rate khz frame rate fps spf compression mode lossy stream size mib encoded date utc tagged date utc indexing using ffmsindex c program files staxrip apps plugins dual ffmsindex exe d downloads idm downloads timecode sample c temp staxrip timecode sample temp temp ffindex writing index done start end duration configuration template zk base video encoder profile container muxer profile mkv mkvmerge avisynth script addautoloaddir c program files staxrip apps frameserver avisynth plugins addautoloaddir c program files staxrip settings plugins avisynth addautoloaddir c program files staxrip settings plugins dual loadplugin c program files staxrip apps plugins dual dll loadplugin c program files staxrip apps plugins avs jpsdr plugins jpsdr dll ffvideosource d downloads idm downloads timecode sample cachefile c temp staxrip timecode sample temp temp ffindex assumefps lanczosresizemt prefetch trim trim source script info width height frames time framerate format target script info width height frames time framerate format video encoding m c program files staxrip apps encoders exe crf preset slow output depth output c temp staxrip timecode sample temp timecode sample out hevc c temp staxrip timecode sample temp timecode sample avs avs avisynth master avs video colorspace avs video depth avs video resolution avs video framerate avs video framecount avs fps frames of raw output file c temp staxrip timecode sample temp timecode sample out hevc hevc encoder version build info using cpu capabilities lzcnt avx main profile level main tier thread pool created using threads slices frame threads pool features wpp rows source height disabling lookahead slices coding qt max cu size min cu size residual qt max tu size max depth inter intra me range subpel merge star keyframe min max scenecut bias lookahead bframes badapt b pyramid weightp weightb references ref limit cu depth on on aq mode str qg size cu tree rate control qcompress crf tools rect limit modes rd psy rd rdoq psy rdoq tools rskip mode signhide tmvp strong intra smoothing deblock tools sao frame i avg qp kb s frame p avg qp kb s frame b avg qp kb s weighted p frames y uv consecutive b frames encoded frames in fps kb s avg qp start end duration general complete name c temp staxrip timecode sample temp timecode sample out hevc format hevc format info high efficiency video coding file size kib duration ms overall bit rate kb s writing library video format hevc format info high efficiency video coding format profile main main duration ms width pixels height pixels display aspect ratio frame rate fps color space yuv chroma subsampling bit depth bits writing library muxing to mkv mkvmerge c program files staxrip apps support mkvtoolnix mkvmerge exe o r timecode sample mkv c temp staxrip timecode sample temp timecode sample out hevc c temp staxrip timecode sample temp timecode sample out hevc ui language en title mkvmerge secret for the mad bit c temp staxrip timecode sample temp timecode sample out hevc using the demultiplexer for the format hevc h c temp staxrip timecode sample temp timecode sample out hevc using the demultiplexer for the format hevc h c temp staxrip timecode sample temp timecode sample out hevc track using the output module for the format hevc h unframed c temp staxrip timecode sample temp timecode sample out hevc track using the output module for the format hevc h unframed no append mapping was given for the file no c temp staxrip timecode sample temp timecode sample out hevc a default mapping of will be used instead please keep that in mind if mkvmerge aborts with an error message regarding invalid append to options the file r timecode sample mkv has been opened for writing appending track from file no c temp staxrip timecode sample temp timecode sample out hevc to track from file no c temp staxrip timecode sample temp timecode sample out hevc the cue entries the index are being written multiplexing took seconds start end duration general complete name r timecode sample mkv format matroska format version version file size kib duration s ms overall bit rate kb s encoded date utc writing application mkvmerge secret for the mad bit writing library libebml libmatroska video id format hevc format info high efficiency video coding format profile main main codec id v mpegh iso hevc duration s ms bit rate kb s width pixels height pixels display aspect ratio frame rate mode constant frame rate fps color space yuv chroma subsampling bit depth bits bits pixel frame stream size kib writing library default yes forced no job complete start end duration
| 1
|
208,680
| 7,157,243,900
|
IssuesEvent
|
2018-01-26 19:11:32
|
unitystation/unitystation
|
https://api.github.com/repos/unitystation/unitystation
|
reopened
|
Clients cannot walk on tiles where explosions took place
|
High Priority bug
|
# Preliminary information
## Category
BUG
## Bug: Release or Development?
0.2.2 beta 4 steam
# Report
## Current Behavior
You cannot step on tiles (you lerp back) explosives were located on (prior to their explosion)
## Expected/Wanted/Requested Behavior
Want to walk free
## Steps to Reproduce
Please enter the steps to reproduce the bug or behaviour:
1. Explode a fuel tank from safe distance
2. Try to step on the tile where the tank was

## OS used:
OSX 10.9
## System specifications
MBP mid2010
## Network
beta server 60ms
|
1.0
|
Clients cannot walk on tiles where explosions took place - # Preliminary information
## Category
BUG
## Bug: Release or Development?
0.2.2 beta 4 steam
# Report
## Current Behavior
You cannot step on tiles (you lerp back) explosives were located on (prior to their explosion)
## Expected/Wanted/Requested Behavior
Want to walk free
## Steps to Reproduce
Please enter the steps to reproduce the bug or behaviour:
1. Explode a fuel tank from safe distance
2. Try to step on the tile where the tank was

## OS used:
OSX 10.9
## System specifications
MBP mid2010
## Network
beta server 60ms
|
priority
|
clients cannot walk on tiles where explosions took place preliminary information category bug bug release or development beta steam report current behavior you cannot step on tiles you lerp back explosives were located on prior to their explosion expected wanted requested behavior want to walk free steps to reproduce please enter the steps to reproduce the bug or behaviour explode a fuel tank from safe distance try to step on the tile where the tank was os used osx system specifications mbp network beta server
| 1
|
201,800
| 7,036,735,001
|
IssuesEvent
|
2017-12-28 10:32:11
|
akvo/akvo-rsr
|
https://api.github.com/repos/akvo/akvo-rsr
|
closed
|
Hide field labels from the qualitative indicator update for IUCN
|
Priority: High Type: Results framework
|
IUCN uses qualitative indicators for regular narrative reporting. They would like to avoid potential confusion for their Users by removing field names referring to a qualitative indictor.
Also, the 'Actual' field label will be changed to 'Narrative Reporting'.

|
1.0
|
Hide field labels from the qualitative indicator update for IUCN - IUCN uses qualitative indicators for regular narrative reporting. They would like to avoid potential confusion for their Users by removing field names referring to a qualitative indictor.
Also, the 'Actual' field label will be changed to 'Narrative Reporting'.

|
priority
|
hide field labels from the qualitative indicator update for iucn iucn uses qualitative indicators for regular narrative reporting they would like to avoid potential confusion for their users by removing field names referring to a qualitative indictor also the actual field label will be changed to narrative reporting
| 1
|
647,601
| 21,112,162,355
|
IssuesEvent
|
2022-04-05 03:47:54
|
ethnt/orchard
|
https://api.github.com/repos/ethnt/orchard
|
opened
|
HTPC VM not properly utilizing passthrough GPU
|
bug machine: htpc high priority
|
```
# lspci -nnk | grep -iA2 VG
00:10.0 VGA compatible controller [0300]: Intel Corporation TigerLake-LP GT2 [Iris Xe Graphics] [8086:9a49] (rev 01)
Subsystem: Intel Corporation Device [8086:3004]
Kernel driver in use: i915
```
```
# inxi -G
Graphics: Message: No device data found.
Display: server: No display server data found. Headless machine? tty: 198x39
Message: Advanced graphics data unavailable in console for root.
```
```
# lshw -C display
*-display
physical id: 10
bus info: pci@0000:00:10.0
version: 01
width: 64 bits
clock: 33MHz
capabilities: pciexpress msi pm bus_master cap_list rom
configuration: driver=i915 latency=0
resources: iomemory:80-7f iomemory:80-7f irq:40 memory:810000000-810ffffff memory:800000000-80fffffff ioport:f000(size=64) memory:c0000-dffff
```
|
1.0
|
HTPC VM not properly utilizing passthrough GPU - ```
# lspci -nnk | grep -iA2 VG
00:10.0 VGA compatible controller [0300]: Intel Corporation TigerLake-LP GT2 [Iris Xe Graphics] [8086:9a49] (rev 01)
Subsystem: Intel Corporation Device [8086:3004]
Kernel driver in use: i915
```
```
# inxi -G
Graphics: Message: No device data found.
Display: server: No display server data found. Headless machine? tty: 198x39
Message: Advanced graphics data unavailable in console for root.
```
```
# lshw -C display
*-display
physical id: 10
bus info: pci@0000:00:10.0
version: 01
width: 64 bits
clock: 33MHz
capabilities: pciexpress msi pm bus_master cap_list rom
configuration: driver=i915 latency=0
resources: iomemory:80-7f iomemory:80-7f irq:40 memory:810000000-810ffffff memory:800000000-80fffffff ioport:f000(size=64) memory:c0000-dffff
```
|
priority
|
htpc vm not properly utilizing passthrough gpu lspci nnk grep vg vga compatible controller intel corporation tigerlake lp rev subsystem intel corporation device kernel driver in use inxi g graphics message no device data found display server no display server data found headless machine tty message advanced graphics data unavailable in console for root lshw c display display physical id bus info pci version width bits clock capabilities pciexpress msi pm bus master cap list rom configuration driver latency resources iomemory iomemory irq memory memory ioport size memory dffff
| 1
|
111,216
| 4,466,884,802
|
IssuesEvent
|
2016-08-25 01:05:03
|
AtlasOfLivingAustralia/spatial-portal
|
https://api.github.com/repos/AtlasOfLivingAustralia/spatial-portal
|
closed
|
Add select box/generate selected layer for contextual layers with multi-class legends
|
enhancement priority-high
|
Currently, it is difficult to select and identify multiple polygons from one or more contextual layers that have multi-class legends. It seems consistent to offer the parallel to facet class selection and layer generation to contextual layers.
|
1.0
|
Add select box/generate selected layer for contextual layers with multi-class legends - Currently, it is difficult to select and identify multiple polygons from one or more contextual layers that have multi-class legends. It seems consistent to offer the parallel to facet class selection and layer generation to contextual layers.
|
priority
|
add select box generate selected layer for contextual layers with multi class legends currently it is difficult to select and identify multiple polygons from one or more contextual layers that have multi class legends it seems consistent to offer the parallel to facet class selection and layer generation to contextual layers
| 1
|
517,108
| 14,994,738,290
|
IssuesEvent
|
2021-01-29 13:22:04
|
lumian/grcentral
|
https://api.github.com/repos/lumian/grcentral
|
closed
|
Изменение логики отдачи прошивок
|
Priority: High Type: Planned improvement
|
По результатам общения с технической поддержкой Grandstream и нескольких неудачных обновлений было принято решение об изменении логики отдачи прошивок.
Суть изменения в следующем:
* Устройства всегда должны получать статус HTTP 200 и актуальный файл прошивки. Устройство само решает, будет оно обновляться или нет.
* Статус HTTP 404 должен отдаваться только в случае каких-либо ошибок или не правильных настроек сервера.
|
1.0
|
Изменение логики отдачи прошивок - По результатам общения с технической поддержкой Grandstream и нескольких неудачных обновлений было принято решение об изменении логики отдачи прошивок.
Суть изменения в следующем:
* Устройства всегда должны получать статус HTTP 200 и актуальный файл прошивки. Устройство само решает, будет оно обновляться или нет.
* Статус HTTP 404 должен отдаваться только в случае каких-либо ошибок или не правильных настроек сервера.
|
priority
|
изменение логики отдачи прошивок по результатам общения с технической поддержкой grandstream и нескольких неудачных обновлений было принято решение об изменении логики отдачи прошивок суть изменения в следующем устройства всегда должны получать статус http и актуальный файл прошивки устройство само решает будет оно обновляться или нет статус http должен отдаваться только в случае каких либо ошибок или не правильных настроек сервера
| 1
|
242,716
| 7,845,789,188
|
IssuesEvent
|
2018-06-19 13:54:00
|
nanoframework/Home
|
https://api.github.com/repos/nanoframework/Home
|
closed
|
MDP is not generating a valid PE when project has skeleton project enabled
|
Area: Visual Studio extension Priority: High Status: FIXED Type: Bug
|
The generated PE seems to be missing it's CRC32 so the resolving assemblies step (at boot time) is failing with a "missing" referenced assembly.
|
1.0
|
MDP is not generating a valid PE when project has skeleton project enabled - The generated PE seems to be missing it's CRC32 so the resolving assemblies step (at boot time) is failing with a "missing" referenced assembly.
|
priority
|
mdp is not generating a valid pe when project has skeleton project enabled the generated pe seems to be missing it s so the resolving assemblies step at boot time is failing with a missing referenced assembly
| 1
|
219,679
| 7,344,935,487
|
IssuesEvent
|
2018-03-07 16:00:53
|
MyCryptoHQ/MyCrypto
|
https://api.github.com/repos/MyCryptoHQ/MyCrypto
|
opened
|
Update Translations
|
blocking stable launch enhancement high priority
|
### Description of the Issue
With the release of the beta, we've added and update a lot of copy throughout the site. To provide translations, all the new 'raw' text needs to be added to the translation files.
|
1.0
|
Update Translations - ### Description of the Issue
With the release of the beta, we've added and update a lot of copy throughout the site. To provide translations, all the new 'raw' text needs to be added to the translation files.
|
priority
|
update translations description of the issue with the release of the beta we ve added and update a lot of copy throughout the site to provide translations all the new raw text needs to be added to the translation files
| 1
|
710,803
| 24,435,280,399
|
IssuesEvent
|
2022-10-06 11:01:40
|
HiAvatar/backend
|
https://api.github.com/repos/HiAvatar/backend
|
closed
|
신규 프로젝트를 생성할 때 아바타에 대한 default 값이 변경됨
|
Type: Enhancement Priority: High
|
### Description
기존에는 avatarName, avatarType, bgName에 대해 '미정', '미정', '배경0'으로 default 값으로 지정됐지만,
FE 요청에 따라 'avatar1', '미정', '미정'으로 변경한다.
<br>
### Todo List
- [x] Embeddable 아바타 클래스 수정하기
<br>
### e.t.c (Optional)
(참조 및 링크 첨부하기)
|
1.0
|
신규 프로젝트를 생성할 때 아바타에 대한 default 값이 변경됨 - ### Description
기존에는 avatarName, avatarType, bgName에 대해 '미정', '미정', '배경0'으로 default 값으로 지정됐지만,
FE 요청에 따라 'avatar1', '미정', '미정'으로 변경한다.
<br>
### Todo List
- [x] Embeddable 아바타 클래스 수정하기
<br>
### e.t.c (Optional)
(참조 및 링크 첨부하기)
|
priority
|
신규 프로젝트를 생성할 때 아바타에 대한 default 값이 변경됨 description 기존에는 avatarname avatartype bgname에 대해 미정 미정 으로 default 값으로 지정됐지만 fe 요청에 따라 미정 미정 으로 변경한다 todo list embeddable 아바타 클래스 수정하기 e t c optional 참조 및 링크 첨부하기
| 1
|
32,029
| 2,743,072,485
|
IssuesEvent
|
2015-04-21 19:44:37
|
afollestad/cabinet-issue-tracker
|
https://api.github.com/repos/afollestad/cabinet-issue-tracker
|
closed
|
Fast scroller crashes from Fabric.io
|
bug high priority
|
https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5534e10f5141dcfd8f847269
```groovy
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2334)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2392)
at android.app.ActivityThread.access$900(ActivityThread.java:169)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1280)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:146)
at android.app.ActivityThread.main(ActivityThread.java:5487)
at java.lang.reflect.Method.invokeNative(Method.java)
at java.lang.reflect.Method.invoke(Method.java:515)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1283)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1099)
at dalvik.system.NativeStart.main(NativeStart.java)
Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.view.LayoutInflater.createView(LayoutInflater.java:626)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:702)
```
https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568
```groovy
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087)
at android.app.ActivityThread.access$600(ActivityThread.java:133)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198)
at android.os.Handler.dispatchMessage(Handler.java:99)
at android.os.Looper.loop(Looper.java:146)
at android.app.ActivityThread.main(ActivityThread.java:4777)
at java.lang.reflect.Method.invokeNative(Method.java)
at java.lang.reflect.Method.invoke(Method.java:511)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556)
at dalvik.system.NativeStart.main(NativeStart.java)
Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.view.LayoutInflater.createView(LayoutInflater.java:613)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687)
```
https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568
```groovy
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087)
at android.app.ActivityThread.access$600(ActivityThread.java:133)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198)
at android.os.Handler.dispatchMessage(Handler.java:99)
at android.os.Looper.loop(Looper.java:146)
at android.app.ActivityThread.main(ActivityThread.java:4777)
at java.lang.reflect.Method.invokeNative(Method.java)
at java.lang.reflect.Method.invoke(Method.java:511)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556)
at dalvik.system.NativeStart.main(NativeStart.java)
Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.view.LayoutInflater.createView(LayoutInflater.java:613)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687)
```
|
1.0
|
Fast scroller crashes from Fabric.io - https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5534e10f5141dcfd8f847269
```groovy
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2334)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2392)
at android.app.ActivityThread.access$900(ActivityThread.java:169)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1280)
at android.os.Handler.dispatchMessage(Handler.java:102)
at android.os.Looper.loop(Looper.java:146)
at android.app.ActivityThread.main(ActivityThread.java:5487)
at java.lang.reflect.Method.invokeNative(Method.java)
at java.lang.reflect.Method.invoke(Method.java:515)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:1283)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1099)
at dalvik.system.NativeStart.main(NativeStart.java)
Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.view.LayoutInflater.createView(LayoutInflater.java:626)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:702)
```
https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568
```groovy
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087)
at android.app.ActivityThread.access$600(ActivityThread.java:133)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198)
at android.os.Handler.dispatchMessage(Handler.java:99)
at android.os.Looper.loop(Looper.java:146)
at android.app.ActivityThread.main(ActivityThread.java:4777)
at java.lang.reflect.Method.invokeNative(Method.java)
at java.lang.reflect.Method.invoke(Method.java:511)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556)
at dalvik.system.NativeStart.main(NativeStart.java)
Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.view.LayoutInflater.createView(LayoutInflater.java:613)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687)
```
https://fabric.io/aidan-follestads-projects/android/apps/com.afollestad.cabinet/issues/5535b3b65141dcfd8f865568
```groovy
java.lang.RuntimeException: Unable to start activity ComponentInfo{com.afollestad.cabinet/com.afollestad.cabinet.ui.MainActivity}: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.app.ActivityThread.performLaunchActivity(ActivityThread.java:2062)
at android.app.ActivityThread.handleLaunchActivity(ActivityThread.java:2087)
at android.app.ActivityThread.access$600(ActivityThread.java:133)
at android.app.ActivityThread$H.handleMessage(ActivityThread.java:1198)
at android.os.Handler.dispatchMessage(Handler.java:99)
at android.os.Looper.loop(Looper.java:146)
at android.app.ActivityThread.main(ActivityThread.java:4777)
at java.lang.reflect.Method.invokeNative(Method.java)
at java.lang.reflect.Method.invoke(Method.java:511)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run(ZygoteInit.java:789)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:556)
at dalvik.system.NativeStart.main(NativeStart.java)
Caused by: android.view.InflateException: Binary XML file line #20: Error inflating class xyz.danoz.recyclerviewfastscroller.FastScroller
at android.view.LayoutInflater.createView(LayoutInflater.java:613)
at android.view.LayoutInflater.createViewFromTag(LayoutInflater.java:687)
```
|
priority
|
fast scroller crashes from fabric io groovy java lang runtimeexception unable to start activity componentinfo com afollestad cabinet com afollestad cabinet ui mainactivity android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative method java at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main nativestart java caused by android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android view layoutinflater createview layoutinflater java at android view layoutinflater createviewfromtag layoutinflater java groovy java lang runtimeexception unable to start activity componentinfo com afollestad cabinet com afollestad cabinet ui mainactivity android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative method java at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main nativestart java caused by android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android view layoutinflater createview layoutinflater java at android view layoutinflater createviewfromtag layoutinflater java groovy java lang runtimeexception unable to start activity componentinfo com afollestad cabinet com afollestad cabinet ui mainactivity android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android app activitythread performlaunchactivity activitythread java at android app activitythread handlelaunchactivity activitythread java at android app activitythread access activitythread java at android app activitythread h handlemessage activitythread java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invokenative method java at java lang reflect method invoke method java at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java at dalvik system nativestart main nativestart java caused by android view inflateexception binary xml file line error inflating class xyz danoz recyclerviewfastscroller fastscroller at android view layoutinflater createview layoutinflater java at android view layoutinflater createviewfromtag layoutinflater java
| 1
|
464,733
| 13,339,134,609
|
IssuesEvent
|
2020-08-28 12:21:57
|
ruby-concurrency/concurrent-ruby
|
https://api.github.com/repos/ruby-concurrency/concurrent-ruby
|
reopened
|
Segmentation fault with ruby-2.6.3
|
bug high-priority
|
When updating an app to run with ruby 2.6.3 (from 2.4.4) I get the following error:
```
/home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87: [BUG] Segmentation fault at 0x000000002222002d
ruby 2.6.3p62 (2019-04-16 revision 67580) [x86_64-linux]
-- Control frame information -----------------------------------------------
c:0002 p:0006 s:0007 e:000005 BLOCK /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var. [FINISH]
c:0001 p:---- s:0003 e:000002 (none) [FINISH]
-- Ruby level backtrace information ----------------------------------------
/home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87:in `block (2 levels) in threadlocal_finalizer'
-- Machine register context ------------------------------------------------
RIP: 0x00007fde9b71bf49 RBP: 0x0000000000021deb RSP: 0x00007fde97cb77e8
RAX: 0x000000000b275620 RBX: 0x0000000001719630 RCX: 0x000000002222002d
RDX: 0x0000000000000000 RDI: 0x000000000b275620 RSI: 0x000000000a667f70
R8: 0x0000000001718d50 R9: 0x0000000001719618 R10: 0x0000000000000008
R11: 0x000000000a568038 R12: 0x000000000b2750f0 R13: 0x000000000a567ff8
R14: 0x0000000001719640 R15: 0x000000000a667f70 EFL: 0x0000000000010206
-- C level backtrace information -------------------------------------------
```
I tried using `Concurrent.use_stdlib_logger(Logger::DEBUG)` to get more information but nothing else is printed to the console. How can I start debugging this problem?
```
* Operating system: linux
* Ruby implementation: Ruby 2.6.3
* `concurrent-ruby` version: 1.1.5
* `concurrent-ruby-ext` installed: no
* `concurrent-ruby-edge` used: no
```
|
1.0
|
Segmentation fault with ruby-2.6.3 - When updating an app to run with ruby 2.6.3 (from 2.4.4) I get the following error:
```
/home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87: [BUG] Segmentation fault at 0x000000002222002d
ruby 2.6.3p62 (2019-04-16 revision 67580) [x86_64-linux]
-- Control frame information -----------------------------------------------
c:0002 p:0006 s:0007 e:000005 BLOCK /home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var. [FINISH]
c:0001 p:---- s:0003 e:000002 (none) [FINISH]
-- Ruby level backtrace information ----------------------------------------
/home/user/.rvm/gems/ruby-2.6.3@project/gems/concurrent-ruby-1.1.5/lib/concurrent/atomic/ruby_thread_local_var.rb:87:in `block (2 levels) in threadlocal_finalizer'
-- Machine register context ------------------------------------------------
RIP: 0x00007fde9b71bf49 RBP: 0x0000000000021deb RSP: 0x00007fde97cb77e8
RAX: 0x000000000b275620 RBX: 0x0000000001719630 RCX: 0x000000002222002d
RDX: 0x0000000000000000 RDI: 0x000000000b275620 RSI: 0x000000000a667f70
R8: 0x0000000001718d50 R9: 0x0000000001719618 R10: 0x0000000000000008
R11: 0x000000000a568038 R12: 0x000000000b2750f0 R13: 0x000000000a567ff8
R14: 0x0000000001719640 R15: 0x000000000a667f70 EFL: 0x0000000000010206
-- C level backtrace information -------------------------------------------
```
I tried using `Concurrent.use_stdlib_logger(Logger::DEBUG)` to get more information but nothing else is printed to the console. How can I start debugging this problem?
```
* Operating system: linux
* Ruby implementation: Ruby 2.6.3
* `concurrent-ruby` version: 1.1.5
* `concurrent-ruby-ext` installed: no
* `concurrent-ruby-edge` used: no
```
|
priority
|
segmentation fault with ruby when updating an app to run with ruby from i get the following error home user rvm gems ruby project gems concurrent ruby lib concurrent atomic ruby thread local var rb segmentation fault at ruby revision control frame information c p s e block home user rvm gems ruby project gems concurrent ruby lib concurrent atomic ruby thread local var c p s e none ruby level backtrace information home user rvm gems ruby project gems concurrent ruby lib concurrent atomic ruby thread local var rb in block levels in threadlocal finalizer machine register context rip rbp rsp rax rbx rcx rdx rdi rsi efl c level backtrace information i tried using concurrent use stdlib logger logger debug to get more information but nothing else is printed to the console how can i start debugging this problem operating system linux ruby implementation ruby concurrent ruby version concurrent ruby ext installed no concurrent ruby edge used no
| 1
|
78,518
| 3,510,761,726
|
IssuesEvent
|
2016-01-09 18:58:50
|
Co0sh/BetonQuest
|
https://api.github.com/repos/Co0sh/BetonQuest
|
closed
|
More journal options
|
Feature High Priority
|
In the _config.yml_ file there should be these additional options:
* configurable amount of characters per page
* one entry per page option
* reversed order of entries option
* hide date/time in journal entries
|
1.0
|
More journal options - In the _config.yml_ file there should be these additional options:
* configurable amount of characters per page
* one entry per page option
* reversed order of entries option
* hide date/time in journal entries
|
priority
|
more journal options in the config yml file there should be these additional options configurable amount of characters per page one entry per page option reversed order of entries option hide date time in journal entries
| 1
|
526,738
| 15,300,311,458
|
IssuesEvent
|
2021-02-24 12:08:29
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
Feature Gate
|
Priority/High Severity/Major feature identity-core improvement task
|
### Story - Why do we need the Feature Gate
- Half-baked features might propagate to staging or prod environments.
- Even if the feature development is completed, we might want to keep it disabled for time due to various reasons such as pricing.
- We need to enable/disable some features based on the tenant.
### Requirements
- Ability to turn on/off features on each environment separately.
- Via config file
- Via REST API
- Identified features that will be controlled by Feature Gate.
- https://docs.google.com/spreadsheets/d/14q4qD_O14Lvd78WZeyKy48tkHz0shzMwVWaOxdWwos8/edit#gid=0
### Implementation
- BE
- API level
- UI
- Levels of block in UI
- Hide from UI
- Disable from UI (Greyed icons)
- Need to show an indication for the feature status.
- Experimental
- New feature
Beta feature
|
1.0
|
Feature Gate - ### Story - Why do we need the Feature Gate
- Half-baked features might propagate to staging or prod environments.
- Even if the feature development is completed, we might want to keep it disabled for time due to various reasons such as pricing.
- We need to enable/disable some features based on the tenant.
### Requirements
- Ability to turn on/off features on each environment separately.
- Via config file
- Via REST API
- Identified features that will be controlled by Feature Gate.
- https://docs.google.com/spreadsheets/d/14q4qD_O14Lvd78WZeyKy48tkHz0shzMwVWaOxdWwos8/edit#gid=0
### Implementation
- BE
- API level
- UI
- Levels of block in UI
- Hide from UI
- Disable from UI (Greyed icons)
- Need to show an indication for the feature status.
- Experimental
- New feature
Beta feature
|
priority
|
feature gate story why do we need the feature gate half baked features might propagate to staging or prod environments even if the feature development is completed we might want to keep it disabled for time due to various reasons such as pricing we need to enable disable some features based on the tenant requirements ability to turn on off features on each environment separately via config file via rest api identified features that will be controlled by feature gate implementation be api level ui levels of block in ui hide from ui disable from ui greyed icons need to show an indication for the feature status experimental new feature beta feature
| 1
|
435,013
| 12,530,592,424
|
IssuesEvent
|
2020-06-04 13:19:35
|
zeebe-io/zeebe
|
https://api.github.com/repos/zeebe-io/zeebe
|
opened
|
Elasticsearch exporter ignores errors on flush
|
Impact: Integration Scope: broker Severity: High Status: Needs Priority Type: Bug
|
**Describe the bug**
If a failure occurs while exporting records to Elasticsearch then the failure is not reported and no retry happens. As a result, records may not be exported.
**To Reproduce**
Not sure how to trigger a failure expect using mocks.
Found while fixing #4640.
**Expected behavior**
The failure is reported to the log and the exporter try again.
**Log/Stacktrace**
Nope.
**Environment:**
- OS: [e.g. Linux]
- Zeebe Version: 0.24.0-alpha2
- Configuration: elasticsearch-exporter
|
1.0
|
Elasticsearch exporter ignores errors on flush - **Describe the bug**
If a failure occurs while exporting records to Elasticsearch then the failure is not reported and no retry happens. As a result, records may not be exported.
**To Reproduce**
Not sure how to trigger a failure expect using mocks.
Found while fixing #4640.
**Expected behavior**
The failure is reported to the log and the exporter try again.
**Log/Stacktrace**
Nope.
**Environment:**
- OS: [e.g. Linux]
- Zeebe Version: 0.24.0-alpha2
- Configuration: elasticsearch-exporter
|
priority
|
elasticsearch exporter ignores errors on flush describe the bug if a failure occurs while exporting records to elasticsearch then the failure is not reported and no retry happens as a result records may not be exported to reproduce not sure how to trigger a failure expect using mocks found while fixing expected behavior the failure is reported to the log and the exporter try again log stacktrace nope environment os zeebe version configuration elasticsearch exporter
| 1
|
723,642
| 24,904,602,731
|
IssuesEvent
|
2022-10-29 04:32:50
|
AY2223S1-CS2103T-F12-2/tp
|
https://api.github.com/repos/AY2223S1-CS2103T-F12-2/tp
|
closed
|
Possible to save task with illegal values
|
priority.High
|
Entering the command :
"addTask n/homework d/coding assignment pr/high c/backend dl/2022-12-12 pe/johnd@example.com"
will cause the AddressBook to save the task with illegal vales.
This results in the AddressBook starting with empty data the next time it is run
|
1.0
|
Possible to save task with illegal values - Entering the command :
"addTask n/homework d/coding assignment pr/high c/backend dl/2022-12-12 pe/johnd@example.com"
will cause the AddressBook to save the task with illegal vales.
This results in the AddressBook starting with empty data the next time it is run
|
priority
|
possible to save task with illegal values entering the command addtask n homework d coding assignment pr high c backend dl pe johnd example com will cause the addressbook to save the task with illegal vales this results in the addressbook starting with empty data the next time it is run
| 1
|
99,001
| 4,044,937,408
|
IssuesEvent
|
2016-05-21 17:10:02
|
CPLamb/RouteTracker
|
https://api.github.com/repos/CPLamb/RouteTracker
|
opened
|
Rework Home screen to add final elements & fix constraints
|
Priority - High
|
1. As discussed a million times we need to get the constraints fixed. Test for largest iPhone on stimulator.
2. Add paper logo to this page above the filename field
3. Make the NewpaperTaxi logo properly fixed. It disappears for some stupid reason
4. Not sure what else
5. ---
|
1.0
|
Rework Home screen to add final elements & fix constraints - 1. As discussed a million times we need to get the constraints fixed. Test for largest iPhone on stimulator.
2. Add paper logo to this page above the filename field
3. Make the NewpaperTaxi logo properly fixed. It disappears for some stupid reason
4. Not sure what else
5. ---
|
priority
|
rework home screen to add final elements fix constraints as discussed a million times we need to get the constraints fixed test for largest iphone on stimulator add paper logo to this page above the filename field make the newpapertaxi logo properly fixed it disappears for some stupid reason not sure what else
| 1
|
784,348
| 27,567,428,888
|
IssuesEvent
|
2023-03-08 05:49:54
|
curiouslearning/FeedTheMonsterJS
|
https://api.github.com/repos/curiouslearning/FeedTheMonsterJS
|
closed
|
Create manifest file for all apps in container app
|
High Priority
|
File should ideally contain these infos:-
1. App icon
2. Title
3. Url
**Acceptance criteria**: Manifest file should ideally contain the enough information
to render app icons on container app. Clicking on icon should take us to specified link.
|
1.0
|
Create manifest file for all apps in container app - File should ideally contain these infos:-
1. App icon
2. Title
3. Url
**Acceptance criteria**: Manifest file should ideally contain the enough information
to render app icons on container app. Clicking on icon should take us to specified link.
|
priority
|
create manifest file for all apps in container app file should ideally contain these infos app icon title url acceptance criteria manifest file should ideally contain the enough information to render app icons on container app clicking on icon should take us to specified link
| 1
|
720,059
| 24,777,003,526
|
IssuesEvent
|
2022-10-23 21:11:16
|
RoboJackets/urc-drone
|
https://api.github.com/repos/RoboJackets/urc-drone
|
closed
|
Gazebo Sim Drone
|
area ➤ misc priority ➤ high level ➤ hard
|
Create a gazebo model with dynamics for the drone. This is very necessary for testing our flight, land, and control procedures.
|
1.0
|
Gazebo Sim Drone - Create a gazebo model with dynamics for the drone. This is very necessary for testing our flight, land, and control procedures.
|
priority
|
gazebo sim drone create a gazebo model with dynamics for the drone this is very necessary for testing our flight land and control procedures
| 1
|
601,292
| 18,396,894,095
|
IssuesEvent
|
2021-10-12 12:25:36
|
bounswe/2021SpringGroup4
|
https://api.github.com/repos/bounswe/2021SpringGroup4
|
closed
|
Creating wiki page for Fatih Akgöz
|
wiki individual Priority: High
|
1. Personal wiki page will be created and filled.
2. Weekly Effort wiki page will be created and filled.
|
1.0
|
Creating wiki page for Fatih Akgöz - 1. Personal wiki page will be created and filled.
2. Weekly Effort wiki page will be created and filled.
|
priority
|
creating wiki page for fatih akgöz personal wiki page will be created and filled weekly effort wiki page will be created and filled
| 1
|
213,288
| 7,247,673,561
|
IssuesEvent
|
2018-02-15 04:50:58
|
yanis333/SOEN341_Website
|
https://api.github.com/repos/yanis333/SOEN341_Website
|
closed
|
#56 Upvote & DownVote : add the functionality & the permission to Accepte or Remove Reply
|
High value Priority 2 Risk1 feature sprint 2
|
This is one of the feature for issue #56
[sp]=3
Need to add the Upvote and DownVote for the reply
|
1.0
|
#56 Upvote & DownVote : add the functionality & the permission to Accepte or Remove Reply - This is one of the feature for issue #56
[sp]=3
Need to add the Upvote and DownVote for the reply
|
priority
|
upvote downvote add the functionality the permission to accepte or remove reply this is one of the feature for issue need to add the upvote and downvote for the reply
| 1
|
210,660
| 7,192,110,948
|
IssuesEvent
|
2018-02-03 00:21:59
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
USER ISSUE: Webserver fails to start for solo game in 6.4.2
|
High Priority
|
**Version:** 0.6.4.2 alpha
**Steps to Reproduce:**
Start any solo game with a pre existing game save
**Expected behavior:**
webserver should start so that a webpage will generate
**Actual behavior:**
webserver fails to initialize. This error in console: Failed to start Webserver: Exception has been thrown by the target of an invocation.
|
1.0
|
USER ISSUE: Webserver fails to start for solo game in 6.4.2 - **Version:** 0.6.4.2 alpha
**Steps to Reproduce:**
Start any solo game with a pre existing game save
**Expected behavior:**
webserver should start so that a webpage will generate
**Actual behavior:**
webserver fails to initialize. This error in console: Failed to start Webserver: Exception has been thrown by the target of an invocation.
|
priority
|
user issue webserver fails to start for solo game in version alpha steps to reproduce start any solo game with a pre existing game save expected behavior webserver should start so that a webpage will generate actual behavior webserver fails to initialize this error in console failed to start webserver exception has been thrown by the target of an invocation
| 1
|
802,196
| 28,780,886,697
|
IssuesEvent
|
2023-05-02 00:27:08
|
agency-of-learning/PairApp
|
https://api.github.com/repos/agency-of-learning/PairApp
|
opened
|
Create dev Procfile
|
good first issue high priority solo
|
We should add a `Procfile.dev` that will help with setting up our local host. Params we want are:
1. server
2. worker
|
1.0
|
Create dev Procfile - We should add a `Procfile.dev` that will help with setting up our local host. Params we want are:
1. server
2. worker
|
priority
|
create dev procfile we should add a procfile dev that will help with setting up our local host params we want are server worker
| 1
|
181,259
| 6,657,791,431
|
IssuesEvent
|
2017-09-30 10:38:27
|
status-im/status-react
|
https://api.github.com/repos/status-im/status-react
|
closed
|
Second send transaction can't be sent as send button does nothing and fee value is not shown [feature/upgrade-status-go-g0270590]
|
blocked bug high-priority
|
### Description
*Type*: Bug
*Summary*: User can create 1 send transaction in 1-1 chat. If user does not confirm this transaction and returns to 1-1 chat in order to make 2nd send transaction then it's stuck on Fee step:
1. There is no fee value shown
2. If tap on send button wile input filed contains `/send username 0.2` then nothing happens. Expected: Unsigned screen is shown with 2 transactions.

#### Expected behavior
can send second time send transaction
#### Actual behavior
can't send 2nd transaction as tap on send does nothing
### Reproduction
from 00:12 https://app.testfairy.com/projects/4803590-status/builds/6511547/sessions/5/?accessToken=1xw9BL-p5dEIONbItbxAqqE99xg
- Open Status
- Open 1-1 chat
- tap on /send, select username and type some sum that can be send from your wallet (enough balance), e.g `/send username 0.1`
- on unsigned `transaction screen tap on x in top left corner to close the screen
- in 1-1 chat tap on /send, select username and type some other sum to send. As a result Fee will be shown. Tap on send button
### Additional Information
* Status version: PR `feature/upgrade-status-go-g0270590` 0.9.11d60 (1429.188)
* Operating System:
Real device iPhone 6s, iOS 10.3.3
|
1.0
|
Second send transaction can't be sent as send button does nothing and fee value is not shown [feature/upgrade-status-go-g0270590] - ### Description
*Type*: Bug
*Summary*: User can create 1 send transaction in 1-1 chat. If user does not confirm this transaction and returns to 1-1 chat in order to make 2nd send transaction then it's stuck on Fee step:
1. There is no fee value shown
2. If tap on send button wile input filed contains `/send username 0.2` then nothing happens. Expected: Unsigned screen is shown with 2 transactions.

#### Expected behavior
can send second time send transaction
#### Actual behavior
can't send 2nd transaction as tap on send does nothing
### Reproduction
from 00:12 https://app.testfairy.com/projects/4803590-status/builds/6511547/sessions/5/?accessToken=1xw9BL-p5dEIONbItbxAqqE99xg
- Open Status
- Open 1-1 chat
- tap on /send, select username and type some sum that can be send from your wallet (enough balance), e.g `/send username 0.1`
- on unsigned `transaction screen tap on x in top left corner to close the screen
- in 1-1 chat tap on /send, select username and type some other sum to send. As a result Fee will be shown. Tap on send button
### Additional Information
* Status version: PR `feature/upgrade-status-go-g0270590` 0.9.11d60 (1429.188)
* Operating System:
Real device iPhone 6s, iOS 10.3.3
|
priority
|
second send transaction can t be sent as send button does nothing and fee value is not shown description type bug summary user can create send transaction in chat if user does not confirm this transaction and returns to chat in order to make send transaction then it s stuck on fee step there is no fee value shown if tap on send button wile input filed contains send username then nothing happens expected unsigned screen is shown with transactions expected behavior can send second time send transaction actual behavior can t send transaction as tap on send does nothing reproduction from open status open chat tap on send select username and type some sum that can be send from your wallet enough balance e g send username on unsigned transaction screen tap on x in top left corner to close the screen in chat tap on send select username and type some other sum to send as a result fee will be shown tap on send button additional information status version pr feature upgrade status go operating system real device iphone ios
| 1
|
383,517
| 11,357,067,150
|
IssuesEvent
|
2020-01-25 01:39:28
|
v0idp/Mellow
|
https://api.github.com/repos/v0idp/Mellow
|
closed
|
save password in config file with encryption
|
Hight Priority Rework
|
right now the password is in plaintext which is a no go. needs to be reworked alongside the auth and session code
|
1.0
|
save password in config file with encryption - right now the password is in plaintext which is a no go. needs to be reworked alongside the auth and session code
|
priority
|
save password in config file with encryption right now the password is in plaintext which is a no go needs to be reworked alongside the auth and session code
| 1
|
614,834
| 19,190,541,514
|
IssuesEvent
|
2021-12-05 22:43:07
|
MarkMcGregor-dev/BloodySimulatorGame
|
https://api.github.com/repos/MarkMcGregor-dev/BloodySimulatorGame
|
closed
|
Fix score not being reset after after death
|
bug high priority
|
Score looks like it resets after respawning (goes back to zero) but immediately shows the previous round's score after increasing
|
1.0
|
Fix score not being reset after after death - Score looks like it resets after respawning (goes back to zero) but immediately shows the previous round's score after increasing
|
priority
|
fix score not being reset after after death score looks like it resets after respawning goes back to zero but immediately shows the previous round s score after increasing
| 1
|
491,450
| 14,164,251,297
|
IssuesEvent
|
2020-11-12 04:32:58
|
wso2/docs-ei
|
https://api.github.com/repos/wso2/docs-ei
|
closed
|
Update call template and sequence template docs with new features
|
Priority/High micro-integrator
|
**Description:**
Update documentation with the new features introduced with
https://github.com/wso2/micro-integrator/issues/1673
https://ei.docs.wso2.com/en/latest/micro-integrator/references/mediators/call-Template-Mediator/
|
1.0
|
Update call template and sequence template docs with new features - **Description:**
Update documentation with the new features introduced with
https://github.com/wso2/micro-integrator/issues/1673
https://ei.docs.wso2.com/en/latest/micro-integrator/references/mediators/call-Template-Mediator/
|
priority
|
update call template and sequence template docs with new features description update documentation with the new features introduced with
| 1
|
23,855
| 2,664,564,320
|
IssuesEvent
|
2015-03-20 15:16:52
|
cwrc/CWRC-Writer
|
https://api.github.com/repos/cwrc/CWRC-Writer
|
opened
|
Avoiding tag redundancy when adding annotations to existing tag
|
enhancement high priority
|
http://apps.testing.cwrc.ca/editor/dev/editor_dev.htm#templates/ceww_new_entry_template
Using this template brings home the problems of annotating existing markup: you end up with duplicate tags and it's hard to figure out which one has the entity and which one should be deleted.
James or Andrew, can you think of any creative way of dealing with this? Could we add a right click functionality along the lines of "add entity to this tag"?
|
1.0
|
Avoiding tag redundancy when adding annotations to existing tag - http://apps.testing.cwrc.ca/editor/dev/editor_dev.htm#templates/ceww_new_entry_template
Using this template brings home the problems of annotating existing markup: you end up with duplicate tags and it's hard to figure out which one has the entity and which one should be deleted.
James or Andrew, can you think of any creative way of dealing with this? Could we add a right click functionality along the lines of "add entity to this tag"?
|
priority
|
avoiding tag redundancy when adding annotations to existing tag using this template brings home the problems of annotating existing markup you end up with duplicate tags and it s hard to figure out which one has the entity and which one should be deleted james or andrew can you think of any creative way of dealing with this could we add a right click functionality along the lines of add entity to this tag
| 1
|
731,903
| 25,236,040,321
|
IssuesEvent
|
2022-11-15 01:01:28
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
closed
|
New Request - DOI's for Projects
|
Priority-High (Needed for work) Function-PublicationOrProject Enhancement
|
With journals requiring more publicly available data, we would like to be able to cite Arctos datasets used in original research. Arctos Projects are already set up to associate data loans and other transactions and we can link to journal etc. So being able to then share a DOI for the Project URL would be a huge boon. Currently the project does have a stable URI but it is not recognized as a digital object. (GBIF is not a good option for a variety of reasons...) This would be available across all collection types and has a lot of potential
Can we add a new function button in the edit Project-- generate DOI? This would need to be displayed in the public view when available. This would not be required.
|
1.0
|
New Request - DOI's for Projects - With journals requiring more publicly available data, we would like to be able to cite Arctos datasets used in original research. Arctos Projects are already set up to associate data loans and other transactions and we can link to journal etc. So being able to then share a DOI for the Project URL would be a huge boon. Currently the project does have a stable URI but it is not recognized as a digital object. (GBIF is not a good option for a variety of reasons...) This would be available across all collection types and has a lot of potential
Can we add a new function button in the edit Project-- generate DOI? This would need to be displayed in the public view when available. This would not be required.
|
priority
|
new request doi s for projects with journals requiring more publicly available data we would like to be able to cite arctos datasets used in original research arctos projects are already set up to associate data loans and other transactions and we can link to journal etc so being able to then share a doi for the project url would be a huge boon currently the project does have a stable uri but it is not recognized as a digital object gbif is not a good option for a variety of reasons this would be available across all collection types and has a lot of potential can we add a new function button in the edit project generate doi this would need to be displayed in the public view when available this would not be required
| 1
|
715,206
| 24,590,081,818
|
IssuesEvent
|
2022-10-14 00:44:45
|
OregonDigital/OD2
|
https://api.github.com/repos/OregonDigital/OD2
|
opened
|
Cache cacheable content and investigate additional caching options
|
Enhancement Priority - High MVP Review ruby
|
### Descriptive summary
OD1 is still able to take advantage of Rails 4's Page and Action caching, as you can see with the following command on OD1:
`memcached-tool localhost display`
Rails 5 removed the integrated Page and Action caching, which means OD2 does not yet have the same level of caching that OD1 has.
Without widespread caching, OD2 is going to have a difficult time handling large amounts of traffic. We are able to take advantage of other caches like ActiveFedora's LDP cache for Fedora content. We have Blazegraph for caching triples. We have caching for SQL queries to our primary PostgreSQL database. But to handle larger amounts of traffic and decrease overall response times we need more caching for our Rails frontend application.
Without caching, Rails has to query backend services for every request, which manifests as longer and longer response times as backend services get more busy. Our average response time when the system is relatively idle ranges from 1.5 - 2.5 seconds. While large amounts of Sidekiq jobs are running this can increase to 5 seconds to 20+ seconds.
With more caching we can maintain a more stable User Experience regardless of how busy the overall system is. This will also help to lighten the load on backend services. In addition to the extra resources that background jobs will have available, caching can be added to background jobs where it makes sense to speed them up even further.
Since earlier this week we're now hooked up to `memcached` via `Rails.application.config.cache_store`. We can `write` and `fetch` keys from the `Rails.cache`. Without anything taking advantage of that our `cache_store` is empty, or mostly empty.
```
irb(main):034:0> Rails.cache.stats["memcache.od2-prod.svc.cluster.local:11211"]["total_items"]
=> "1"
```
This was after creating and clearing keys manually, otherwise it would still be returning 0.
Out of the box we can [cache fragments and partials](https://guides.rubyonrails.org/caching_with_rails.html#fragment-caching)
To make use of caching for OD2, we will need to add caching code for content that should be cached.
Additionally, Rack Cache can cache content without changes to code. See the References section for a link.
The Page and Action caching that are missing in Rails 5 are available as gems, which we should test on Staging. See the References section for links.
We can also do low level caching with `Rails.cache.fetch()` and `Rails.cache.write()` for data from backends that won't change often, or are acceptable to be eventually correct. Whether that's SOLR query results, or responses from other external APIs, if the response changes infrequently or eventually correct results are ok we should be caching them. See the low level caching section of the Ruby on Rails Caching Guide for details.
### Expected behavior
Infrequently changed content should be cached wherever possible.
We should be seeing increasing numbers of items in the memcache cache_store.
### References
[https://guides.rubyonrails.org/caching_with_rails.html](https://guides.rubyonrails.org/caching_with_rails.html)
[https://rtomayko.github.io/rack-cache/configuration](https://rtomayko.github.io/rack-cache/configuration)
[https://github.com/rails/actionpack-page_caching](https://github.com/rails/actionpack-page_caching)
|
1.0
|
Cache cacheable content and investigate additional caching options - ### Descriptive summary
OD1 is still able to take advantage of Rails 4's Page and Action caching, as you can see with the following command on OD1:
`memcached-tool localhost display`
Rails 5 removed the integrated Page and Action caching, which means OD2 does not yet have the same level of caching that OD1 has.
Without widespread caching, OD2 is going to have a difficult time handling large amounts of traffic. We are able to take advantage of other caches like ActiveFedora's LDP cache for Fedora content. We have Blazegraph for caching triples. We have caching for SQL queries to our primary PostgreSQL database. But to handle larger amounts of traffic and decrease overall response times we need more caching for our Rails frontend application.
Without caching, Rails has to query backend services for every request, which manifests as longer and longer response times as backend services get more busy. Our average response time when the system is relatively idle ranges from 1.5 - 2.5 seconds. While large amounts of Sidekiq jobs are running this can increase to 5 seconds to 20+ seconds.
With more caching we can maintain a more stable User Experience regardless of how busy the overall system is. This will also help to lighten the load on backend services. In addition to the extra resources that background jobs will have available, caching can be added to background jobs where it makes sense to speed them up even further.
Since earlier this week we're now hooked up to `memcached` via `Rails.application.config.cache_store`. We can `write` and `fetch` keys from the `Rails.cache`. Without anything taking advantage of that our `cache_store` is empty, or mostly empty.
```
irb(main):034:0> Rails.cache.stats["memcache.od2-prod.svc.cluster.local:11211"]["total_items"]
=> "1"
```
This was after creating and clearing keys manually, otherwise it would still be returning 0.
Out of the box we can [cache fragments and partials](https://guides.rubyonrails.org/caching_with_rails.html#fragment-caching)
To make use of caching for OD2, we will need to add caching code for content that should be cached.
Additionally, Rack Cache can cache content without changes to code. See the References section for a link.
The Page and Action caching that are missing in Rails 5 are available as gems, which we should test on Staging. See the References section for links.
We can also do low level caching with `Rails.cache.fetch()` and `Rails.cache.write()` for data from backends that won't change often, or are acceptable to be eventually correct. Whether that's SOLR query results, or responses from other external APIs, if the response changes infrequently or eventually correct results are ok we should be caching them. See the low level caching section of the Ruby on Rails Caching Guide for details.
### Expected behavior
Infrequently changed content should be cached wherever possible.
We should be seeing increasing numbers of items in the memcache cache_store.
### References
[https://guides.rubyonrails.org/caching_with_rails.html](https://guides.rubyonrails.org/caching_with_rails.html)
[https://rtomayko.github.io/rack-cache/configuration](https://rtomayko.github.io/rack-cache/configuration)
[https://github.com/rails/actionpack-page_caching](https://github.com/rails/actionpack-page_caching)
|
priority
|
cache cacheable content and investigate additional caching options descriptive summary is still able to take advantage of rails s page and action caching as you can see with the following command on memcached tool localhost display rails removed the integrated page and action caching which means does not yet have the same level of caching that has without widespread caching is going to have a difficult time handling large amounts of traffic we are able to take advantage of other caches like activefedora s ldp cache for fedora content we have blazegraph for caching triples we have caching for sql queries to our primary postgresql database but to handle larger amounts of traffic and decrease overall response times we need more caching for our rails frontend application without caching rails has to query backend services for every request which manifests as longer and longer response times as backend services get more busy our average response time when the system is relatively idle ranges from seconds while large amounts of sidekiq jobs are running this can increase to seconds to seconds with more caching we can maintain a more stable user experience regardless of how busy the overall system is this will also help to lighten the load on backend services in addition to the extra resources that background jobs will have available caching can be added to background jobs where it makes sense to speed them up even further since earlier this week we re now hooked up to memcached via rails application config cache store we can write and fetch keys from the rails cache without anything taking advantage of that our cache store is empty or mostly empty irb main rails cache stats this was after creating and clearing keys manually otherwise it would still be returning out of the box we can to make use of caching for we will need to add caching code for content that should be cached additionally rack cache can cache content without changes to code see the references section for a link the page and action caching that are missing in rails are available as gems which we should test on staging see the references section for links we can also do low level caching with rails cache fetch and rails cache write for data from backends that won t change often or are acceptable to be eventually correct whether that s solr query results or responses from other external apis if the response changes infrequently or eventually correct results are ok we should be caching them see the low level caching section of the ruby on rails caching guide for details expected behavior infrequently changed content should be cached wherever possible we should be seeing increasing numbers of items in the memcache cache store references
| 1
|
816,803
| 30,612,984,631
|
IssuesEvent
|
2023-07-23 21:02:44
|
silverstripe/addons.silverstripe.org
|
https://api.github.com/repos/silverstripe/addons.silverstripe.org
|
closed
|
Create a generic SS community sites base CSS
|
type/enhancement priority-high
|
To be used with bootstrap-based sites, for now primarily api.ss.org, extensions.ss.org and eventually ss.org. Note that this is different from a SilverStripe theme, since it should be useable in non-SS projects like the staticically generated api.ss.org as well.
|
1.0
|
Create a generic SS community sites base CSS - To be used with bootstrap-based sites, for now primarily api.ss.org, extensions.ss.org and eventually ss.org. Note that this is different from a SilverStripe theme, since it should be useable in non-SS projects like the staticically generated api.ss.org as well.
|
priority
|
create a generic ss community sites base css to be used with bootstrap based sites for now primarily api ss org extensions ss org and eventually ss org note that this is different from a silverstripe theme since it should be useable in non ss projects like the staticically generated api ss org as well
| 1
|
536,568
| 15,710,817,132
|
IssuesEvent
|
2021-03-27 03:49:02
|
Blackoutburst/Wally
|
https://api.github.com/repos/Blackoutburst/Wally
|
closed
|
Search commands does not support player title and custom background
|
bug high priority incorrect
|
**Describe the bug**
When displaying player stats using `!search` their title and background does not display
**To Reproduce**
Steps to reproduce the behavior:
1. use `!search` on any player with a custom title and/or background
**Expected behavior**
It should display everything like we used `!stats` on the player
**Screenshots**
Output

Expected

|
1.0
|
Search commands does not support player title and custom background - **Describe the bug**
When displaying player stats using `!search` their title and background does not display
**To Reproduce**
Steps to reproduce the behavior:
1. use `!search` on any player with a custom title and/or background
**Expected behavior**
It should display everything like we used `!stats` on the player
**Screenshots**
Output

Expected

|
priority
|
search commands does not support player title and custom background describe the bug when displaying player stats using search their title and background does not display to reproduce steps to reproduce the behavior use search on any player with a custom title and or background expected behavior it should display everything like we used stats on the player screenshots output expected
| 1
|
575,977
| 17,067,898,579
|
IssuesEvent
|
2021-07-07 09:36:51
|
geosolutions-it/MapStore2-C040
|
https://api.github.com/repos/geosolutions-it/MapStore2-C040
|
opened
|
Install/Configure Grid Shift Transform files in GeoServer
|
C040-COMUNE_GE-2021-SUPPORT Priority: High Task
|
### Description
Install the Grid Shift Transform files in GeoServer for the CRS 7791 using the transform file provided by the client. Files are available at this path: \\vm-linuxgeofe\MapStore2\grigliati_GE.zip
The documentation to install Grid Shift Transform files is available in the [GeoServer online doc](https://docs.geoserver.org/stable/en/user/configuration/crshandling/coordtransforms.html#add-grid-shift-transform-files).
### Other useful information (optional):
|
1.0
|
Install/Configure Grid Shift Transform files in GeoServer - ### Description
Install the Grid Shift Transform files in GeoServer for the CRS 7791 using the transform file provided by the client. Files are available at this path: \\vm-linuxgeofe\MapStore2\grigliati_GE.zip
The documentation to install Grid Shift Transform files is available in the [GeoServer online doc](https://docs.geoserver.org/stable/en/user/configuration/crshandling/coordtransforms.html#add-grid-shift-transform-files).
### Other useful information (optional):
|
priority
|
install configure grid shift transform files in geoserver description install the grid shift transform files in geoserver for the crs using the transform file provided by the client files are available at this path vm linuxgeofe grigliati ge zip the documentation to install grid shift transform files is available in the other useful information optional
| 1
|
109,456
| 4,387,744,036
|
IssuesEvent
|
2016-08-08 16:41:12
|
danya02/public-bodyguard
|
https://api.github.com/repos/danya02/public-bodyguard
|
closed
|
Presentation
|
meta Very-High-Priority
|
We need a presentation for the project to explain the project, as per the rules of the project for which this one is being developed.
|
1.0
|
Presentation - We need a presentation for the project to explain the project, as per the rules of the project for which this one is being developed.
|
priority
|
presentation we need a presentation for the project to explain the project as per the rules of the project for which this one is being developed
| 1
|
539,619
| 15,792,099,832
|
IssuesEvent
|
2021-04-02 06:20:32
|
kubesphere/kubesphere
|
https://api.github.com/repos/kubesphere/kubesphere
|
closed
|
Alarm policies that do not set thresholds and monitor targets can still generate alarms
|
area/alerting kind/bug kind/need-to-verify priority/high
|
<!--
You don't need to remove this comment section, it's invisible on the issues page.
## General remarks
* Attention, please fill out this issues form using English only!
* 注意!GitHub Issue 仅支持英文,中文 Issue 请在 [论坛](https://kubesphere.com.cn/forum/) 提交。
* This form is to report bugs. For general usage questions you can join our Slack channel
[KubeSphere-users](https://join.slack.com/t/kubesphere/shared_invite/enQtNTE3MDIxNzUxNzQ0LTZkNTdkYWNiYTVkMTM5ZThhODY1MjAyZmVlYWEwZmQ3ODQ1NmM1MGVkNWEzZTRhNzk0MzM5MmY4NDc3ZWVhMjE)
-->
**Describe the Bug**
Alarm policies that do not set thresholds and monitor targets can still generate alarms
**Environment**
http://xxxx:79:30104
**Preset conditions**
1、There is account A, which has cluster management privileges.
2、Use account A to log in to KS, and go to Platform management--》Cluster management--》Alerting Policies.
**How To Reproduce**
Steps to reproduce the behavior:
1. click "create"
2. Enter the name "test"
3. Set the alarm duration to 1 minute
4. Set the alarm rule expression to "node:node_cpu_utilisation:avg1m"
5. save policy
6. Check whether policy “test” produces an alert message
**Expected behavior**
Policy “test” does not produce an alert message


/priority high
/area alerting
/cc @junotx
/kind bug
/milestone 3.1.0
|
1.0
|
Alarm policies that do not set thresholds and monitor targets can still generate alarms - <!--
You don't need to remove this comment section, it's invisible on the issues page.
## General remarks
* Attention, please fill out this issues form using English only!
* 注意!GitHub Issue 仅支持英文,中文 Issue 请在 [论坛](https://kubesphere.com.cn/forum/) 提交。
* This form is to report bugs. For general usage questions you can join our Slack channel
[KubeSphere-users](https://join.slack.com/t/kubesphere/shared_invite/enQtNTE3MDIxNzUxNzQ0LTZkNTdkYWNiYTVkMTM5ZThhODY1MjAyZmVlYWEwZmQ3ODQ1NmM1MGVkNWEzZTRhNzk0MzM5MmY4NDc3ZWVhMjE)
-->
**Describe the Bug**
Alarm policies that do not set thresholds and monitor targets can still generate alarms
**Environment**
http://xxxx:79:30104
**Preset conditions**
1、There is account A, which has cluster management privileges.
2、Use account A to log in to KS, and go to Platform management--》Cluster management--》Alerting Policies.
**How To Reproduce**
Steps to reproduce the behavior:
1. click "create"
2. Enter the name "test"
3. Set the alarm duration to 1 minute
4. Set the alarm rule expression to "node:node_cpu_utilisation:avg1m"
5. save policy
6. Check whether policy “test” produces an alert message
**Expected behavior**
Policy “test” does not produce an alert message


/priority high
/area alerting
/cc @junotx
/kind bug
/milestone 3.1.0
|
priority
|
alarm policies that do not set thresholds and monitor targets can still generate alarms you don t need to remove this comment section it s invisible on the issues page general remarks attention please fill out this issues form using english only 注意!github issue 仅支持英文,中文 issue 请在 提交。 this form is to report bugs for general usage questions you can join our slack channel describe the bug alarm policies that do not set thresholds and monitor targets can still generate alarms environment preset conditions 、there is account a which has cluster management privileges 、use account a to log in to ks and go to platform management 》cluster management 》alerting policies how to reproduce steps to reproduce the behavior click create enter the name test set the alarm duration to minute set the alarm rule expression to node node cpu utilisation save policy check whether policy “test” produces an alert message expected behavior policy “test” does not produce an alert message priority high area alerting cc junotx kind bug milestone
| 1
|
192,443
| 6,849,982,590
|
IssuesEvent
|
2017-11-14 00:37:04
|
eriq-augustine/psl
|
https://api.github.com/repos/eriq-augustine/psl
|
opened
|
More Aggressive Atom Type Conversion
|
Difficulty - Easy Priority - High Type - Bug
|
We validate and do some mild type conversions for atoms (not ground atom) that come from rules.
https://github.com/eriq-augustine/psl/blob/develop/psl-core/src/main/java/org/linqs/psl/model/atom/Atom.java#L113
Like:
```
Foo(+A, 'Dan') = 1 .
```
But we should be doing more aggressive type conversions from strings (since all constants must be quoted).
A string can always be attempted to be converted into whatever type.
|
1.0
|
More Aggressive Atom Type Conversion - We validate and do some mild type conversions for atoms (not ground atom) that come from rules.
https://github.com/eriq-augustine/psl/blob/develop/psl-core/src/main/java/org/linqs/psl/model/atom/Atom.java#L113
Like:
```
Foo(+A, 'Dan') = 1 .
```
But we should be doing more aggressive type conversions from strings (since all constants must be quoted).
A string can always be attempted to be converted into whatever type.
|
priority
|
more aggressive atom type conversion we validate and do some mild type conversions for atoms not ground atom that come from rules like foo a dan but we should be doing more aggressive type conversions from strings since all constants must be quoted a string can always be attempted to be converted into whatever type
| 1
|
183,101
| 6,677,221,544
|
IssuesEvent
|
2017-10-05 09:35:15
|
metasfresh/metasfresh
|
https://api.github.com/repos/metasfresh/metasfresh
|
closed
|
Make letter print preview work with AD_BoilerPlate jasper process
|
branch:master priority:high
|
### Is this a bug or feature request?
Bug
### What is the current behavior?
jasper process is not used
feature was originally implemented here: https://github.com/metasfresh/metasfresh-webui-api/issues/178
#### Which are the steps to reproduce?
1. define a text tempate (boilerplate)
1. select jasper process in template
1. go in existing sales order
1. press letter action
1. select template
1. click print preview
1. OK: letter is opened showing the text of template
1. NOK: jasper process is not used
### What is the expected or desired behavior?
When you print preview form the letter window, the jasper assign to the boiler plate to be included in the preview.
|
1.0
|
Make letter print preview work with AD_BoilerPlate jasper process - ### Is this a bug or feature request?
Bug
### What is the current behavior?
jasper process is not used
feature was originally implemented here: https://github.com/metasfresh/metasfresh-webui-api/issues/178
#### Which are the steps to reproduce?
1. define a text tempate (boilerplate)
1. select jasper process in template
1. go in existing sales order
1. press letter action
1. select template
1. click print preview
1. OK: letter is opened showing the text of template
1. NOK: jasper process is not used
### What is the expected or desired behavior?
When you print preview form the letter window, the jasper assign to the boiler plate to be included in the preview.
|
priority
|
make letter print preview work with ad boilerplate jasper process is this a bug or feature request bug what is the current behavior jasper process is not used feature was originally implemented here which are the steps to reproduce define a text tempate boilerplate select jasper process in template go in existing sales order press letter action select template click print preview ok letter is opened showing the text of template nok jasper process is not used what is the expected or desired behavior when you print preview form the letter window the jasper assign to the boiler plate to be included in the preview
| 1
|
249,999
| 7,966,331,840
|
IssuesEvent
|
2018-07-14 20:38:52
|
City-Bureau/city-scrapers
|
https://api.github.com/repos/City-Bureau/city-scrapers
|
opened
|
Non-coding Contributions: Local Government Spreadsheet (Chicago and Detroit)
|
help wanted non-coding non-coding: agency research non-coding: good first issue priority: high (must have)
|
We're collecting public meeting information from local governing bodies but there's a good deal of supporting information that'll help us increase access to information on those bodies. If you're interested in a non-coding task this one can be done at your own pace and would help us survey the local political landscape.
Information we need help collecting:
- Phone
- Email
- Twitter Handle
- Meeting Agenda URL
- Meeting Minutes URL
- Board/Commission Members
- Annual Budget
- Officer Selection Process
See the Chicago and Detroit tabs of the Public Meetings Agency Table—Columns J-Q (i.e. the blue columns). Contribute info for any department by adding the comments: https://docs.google.com/spreadsheets/d/1Uu9FmcQAy8AXhBR2bj84LbSdlfgMNNroa9P89rsDJSQ/edit#gid=0 (You can request access to the document to contribute information directly).
|
1.0
|
Non-coding Contributions: Local Government Spreadsheet (Chicago and Detroit) - We're collecting public meeting information from local governing bodies but there's a good deal of supporting information that'll help us increase access to information on those bodies. If you're interested in a non-coding task this one can be done at your own pace and would help us survey the local political landscape.
Information we need help collecting:
- Phone
- Email
- Twitter Handle
- Meeting Agenda URL
- Meeting Minutes URL
- Board/Commission Members
- Annual Budget
- Officer Selection Process
See the Chicago and Detroit tabs of the Public Meetings Agency Table—Columns J-Q (i.e. the blue columns). Contribute info for any department by adding the comments: https://docs.google.com/spreadsheets/d/1Uu9FmcQAy8AXhBR2bj84LbSdlfgMNNroa9P89rsDJSQ/edit#gid=0 (You can request access to the document to contribute information directly).
|
priority
|
non coding contributions local government spreadsheet chicago and detroit we re collecting public meeting information from local governing bodies but there s a good deal of supporting information that ll help us increase access to information on those bodies if you re interested in a non coding task this one can be done at your own pace and would help us survey the local political landscape information we need help collecting phone email twitter handle meeting agenda url meeting minutes url board commission members annual budget officer selection process see the chicago and detroit tabs of the public meetings agency table—columns j q i e the blue columns contribute info for any department by adding the comments you can request access to the document to contribute information directly
| 1
|
87,189
| 3,739,015,615
|
IssuesEvent
|
2016-03-09 01:46:41
|
projectcalico/calico-containers
|
https://api.github.com/repos/projectcalico/calico-containers
|
closed
|
Calico Node/Bird routes through default interface(eth0) and not tunnelled interface (tunl0)
|
Priority: High
|
How did i notice the issue -
docker instance on Machine1 was not able to communicate with docker instance on Machine2 inspite of they being part of same profile and the profile had default settings which is allow all communication within the profile.
Background:
I was running a few docker instances on Machine 1(192.168.0.12 , 13 and 14) and they were part of same profile so obviously as expected they were able to communicate with each other.
Now i configured calico on another machine(Machine 2) and created a new docker instance (192.168.0.15) and adding it to same profile as the above mentioned docker instances, But now the docker instances on Machine1 were not able to reach instances on Machine2 and viceversa.
ip route show output from both boxes:
**Machine 1: **
default via 172.31.16.1 dev eth0
10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1
172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1
172.18.0.0/16 dev docker_gwbridge proto kernel scope link src 172.18.0.1
172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.27.45
blackhole 192.168.0.0/26 proto bird
192.168.0.12 dev calie8beb266c79 scope link
192.168.0.13 dev caliefab0610c79 scope link
192.168.0.14 dev calif2c7a52ec79 scope link
192.168.0.15 via 172.31.29.77 dev tunl0 proto bird onlink
192.168.0.64/26 via 172.31.29.77 dev tunl0 proto bird onlink
**Machine 2: **
default via 172.31.16.1 dev eth0
172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1
172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77
192.168.0.0/26 via 172.31.27.45 dev eth0 proto bird
192.168.0.15 dev cali7fcbe484c7a scope link
blackhole 192.168.0.64/26 proto bird
As we can see on machine 2 192.168.0.0/26 is being routed from interface eth0 and with bird proto, so obviously calico must have configured it, which is the bug! it should have configured it on tunl0 interface which was up and running on Machine 2.
tunl0 Link encap:IPIP Tunnel HWaddr
inet addr:192.168.0.64 Mask:255.255.255.255
UP RUNNING NOARP MTU:1440 Metric:1
RX packets:0 errors:0 dropped:0 overruns:0 frame:0
TX packets:4 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:0 (0.0 B) TX bytes:336 (336.0 B)
I tried a few silly things to fix it like removing tunl0 interface and restarting calico node on Machine 2, but none of them helped.So finally i manually deleted the route entry for 192.168.0.0/26 from the route table on Machine2 and bingo! , it learned the new route through tunl0 and the instances were now able to communicate.
ip route show on Machine 2:
default via 172.31.16.1 dev eth0
10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1
172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1
172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77
192.168.0.0/26 via 172.31.27.45 dev tunl0 proto bird onlink
192.168.0.16 dev calie209b94cc7a scope link
blackhole 192.168.0.64/26 proto bird
So shouldnt the bird have configured proper route through tunl0 initially and not through eth0. If the bird dint do it then shouldnt bird be flushing the wrong entries after creating the new interface so the new routes can be learned?
|
1.0
|
Calico Node/Bird routes through default interface(eth0) and not tunnelled interface (tunl0) - How did i notice the issue -
docker instance on Machine1 was not able to communicate with docker instance on Machine2 inspite of they being part of same profile and the profile had default settings which is allow all communication within the profile.
Background:
I was running a few docker instances on Machine 1(192.168.0.12 , 13 and 14) and they were part of same profile so obviously as expected they were able to communicate with each other.
Now i configured calico on another machine(Machine 2) and created a new docker instance (192.168.0.15) and adding it to same profile as the above mentioned docker instances, But now the docker instances on Machine1 were not able to reach instances on Machine2 and viceversa.
ip route show output from both boxes:
**Machine 1: **
default via 172.31.16.1 dev eth0
10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1
172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1
172.18.0.0/16 dev docker_gwbridge proto kernel scope link src 172.18.0.1
172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.27.45
blackhole 192.168.0.0/26 proto bird
192.168.0.12 dev calie8beb266c79 scope link
192.168.0.13 dev caliefab0610c79 scope link
192.168.0.14 dev calif2c7a52ec79 scope link
192.168.0.15 via 172.31.29.77 dev tunl0 proto bird onlink
192.168.0.64/26 via 172.31.29.77 dev tunl0 proto bird onlink
**Machine 2: **
default via 172.31.16.1 dev eth0
172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1
172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77
192.168.0.0/26 via 172.31.27.45 dev eth0 proto bird
192.168.0.15 dev cali7fcbe484c7a scope link
blackhole 192.168.0.64/26 proto bird
As we can see on machine 2 192.168.0.0/26 is being routed from interface eth0 and with bird proto, so obviously calico must have configured it, which is the bug! it should have configured it on tunl0 interface which was up and running on Machine 2.
tunl0 Link encap:IPIP Tunnel HWaddr
inet addr:192.168.0.64 Mask:255.255.255.255
UP RUNNING NOARP MTU:1440 Metric:1
RX packets:0 errors:0 dropped:0 overruns:0 frame:0
TX packets:4 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:0 (0.0 B) TX bytes:336 (336.0 B)
I tried a few silly things to fix it like removing tunl0 interface and restarting calico node on Machine 2, but none of them helped.So finally i manually deleted the route entry for 192.168.0.0/26 from the route table on Machine2 and bingo! , it learned the new route through tunl0 and the instances were now able to communicate.
ip route show on Machine 2:
default via 172.31.16.1 dev eth0
10.0.3.0/24 dev lxcbr0 proto kernel scope link src 10.0.3.1
172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.42.1
172.31.16.0/20 dev eth0 proto kernel scope link src 172.31.29.77
192.168.0.0/26 via 172.31.27.45 dev tunl0 proto bird onlink
192.168.0.16 dev calie209b94cc7a scope link
blackhole 192.168.0.64/26 proto bird
So shouldnt the bird have configured proper route through tunl0 initially and not through eth0. If the bird dint do it then shouldnt bird be flushing the wrong entries after creating the new interface so the new routes can be learned?
|
priority
|
calico node bird routes through default interface and not tunnelled interface how did i notice the issue docker instance on was not able to communicate with docker instance on inspite of they being part of same profile and the profile had default settings which is allow all communication within the profile background i was running a few docker instances on machine and and they were part of same profile so obviously as expected they were able to communicate with each other now i configured calico on another machine machine and created a new docker instance and adding it to same profile as the above mentioned docker instances but now the docker instances on were not able to reach instances on and viceversa ip route show output from both boxes machine default via dev dev proto kernel scope link src dev proto kernel scope link src dev docker gwbridge proto kernel scope link src dev proto kernel scope link src blackhole proto bird dev scope link dev scope link dev scope link via dev proto bird onlink via dev proto bird onlink machine default via dev dev proto kernel scope link src dev proto kernel scope link src via dev proto bird dev scope link blackhole proto bird as we can see on machine is being routed from interface and with bird proto so obviously calico must have configured it which is the bug it should have configured it on interface which was up and running on machine link encap ipip tunnel hwaddr inet addr mask up running noarp mtu metric rx packets errors dropped overruns frame tx packets errors dropped overruns carrier collisions txqueuelen rx bytes b tx bytes b i tried a few silly things to fix it like removing interface and restarting calico node on machine but none of them helped so finally i manually deleted the route entry for from the route table on and bingo it learned the new route through and the instances were now able to communicate ip route show on machine default via dev dev proto kernel scope link src dev proto kernel scope link src dev proto kernel scope link src via dev proto bird onlink dev scope link blackhole proto bird so shouldnt the bird have configured proper route through initially and not through if the bird dint do it then shouldnt bird be flushing the wrong entries after creating the new interface so the new routes can be learned
| 1
|
380,247
| 11,256,105,734
|
IssuesEvent
|
2020-01-12 14:09:37
|
williamjcm/wxMASSManager
|
https://api.github.com/repos/williamjcm/wxMASSManager
|
closed
|
Code reorganisation
|
enhancement high priority
|
When I started the app, I put all code, even the backend stuff, into EvtMainFrame.
Before releasing 1.0 of the app, I should put as much backend stuff as possible into a separate class.
|
1.0
|
Code reorganisation - When I started the app, I put all code, even the backend stuff, into EvtMainFrame.
Before releasing 1.0 of the app, I should put as much backend stuff as possible into a separate class.
|
priority
|
code reorganisation when i started the app i put all code even the backend stuff into evtmainframe before releasing of the app i should put as much backend stuff as possible into a separate class
| 1
|
175,529
| 6,551,786,638
|
IssuesEvent
|
2017-09-05 15:51:17
|
opencurrents/opencurrents
|
https://api.github.com/repos/opencurrents/opencurrents
|
closed
|
time-tracker: Display error if admin is not selected or "Not sure"
|
priority high
|
Display error: "Please select your admin to continue"
|
1.0
|
time-tracker: Display error if admin is not selected or "Not sure" - Display error: "Please select your admin to continue"
|
priority
|
time tracker display error if admin is not selected or not sure display error please select your admin to continue
| 1
|
141,633
| 5,439,239,319
|
IssuesEvent
|
2017-03-06 12:53:34
|
Cadasta/cadasta-platform
|
https://api.github.com/repos/Cadasta/cadasta-platform
|
closed
|
Hardcoded S3 paths
|
bug priority: high security
|
Hardcoded S3 bucket paths appear across multiple files: https://github.com/Cadasta/cadasta-platform/search?utf8=%E2%9C%93&q=platformprod&type=Code
This results in all environments (staging, demo, and production) hitting the production S3 bucket, which violates our isolation requirements. It also exposes the bucket names.
We should remove hardcoded references to bucket names/paths and change bucket names ASAP. We should also avoid commingling static assets like platform icons with project data, to allow for stricter security policies.
|
1.0
|
Hardcoded S3 paths - Hardcoded S3 bucket paths appear across multiple files: https://github.com/Cadasta/cadasta-platform/search?utf8=%E2%9C%93&q=platformprod&type=Code
This results in all environments (staging, demo, and production) hitting the production S3 bucket, which violates our isolation requirements. It also exposes the bucket names.
We should remove hardcoded references to bucket names/paths and change bucket names ASAP. We should also avoid commingling static assets like platform icons with project data, to allow for stricter security policies.
|
priority
|
hardcoded paths hardcoded bucket paths appear across multiple files this results in all environments staging demo and production hitting the production bucket which violates our isolation requirements it also exposes the bucket names we should remove hardcoded references to bucket names paths and change bucket names asap we should also avoid commingling static assets like platform icons with project data to allow for stricter security policies
| 1
|
768,875
| 26,985,024,425
|
IssuesEvent
|
2023-02-09 15:35:16
|
ballerina-platform/ballerina-standard-library
|
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
|
closed
|
NATS module build is failing more frequently
|
Points/2 Priority/High Type/Task module/nats Team/PCM
|
**Description:**
NATS build is intermittently failing for quite some time and now is more frequently failing. Trying the approach of moving most of the tests into a separate test module.
|
1.0
|
NATS module build is failing more frequently - **Description:**
NATS build is intermittently failing for quite some time and now is more frequently failing. Trying the approach of moving most of the tests into a separate test module.
|
priority
|
nats module build is failing more frequently description nats build is intermittently failing for quite some time and now is more frequently failing trying the approach of moving most of the tests into a separate test module
| 1
|
252,080
| 8,031,547,601
|
IssuesEvent
|
2018-07-28 03:04:17
|
crslab/cloud-vision-explorer
|
https://api.github.com/repos/crslab/cloud-vision-explorer
|
closed
|
[HIGH] Update build webpack build process to build quicker
|
enhancement high priority
|
Build process takes forever right now. Rewrite the webpack config, remove gulp from the process, and build for dev not for prod. Hopefully this will make things faster.
|
1.0
|
[HIGH] Update build webpack build process to build quicker - Build process takes forever right now. Rewrite the webpack config, remove gulp from the process, and build for dev not for prod. Hopefully this will make things faster.
|
priority
|
update build webpack build process to build quicker build process takes forever right now rewrite the webpack config remove gulp from the process and build for dev not for prod hopefully this will make things faster
| 1
|
47,327
| 2,974,767,124
|
IssuesEvent
|
2015-07-15 04:16:19
|
OpenSourceMalaria/OSM_To_Do_List
|
https://api.github.com/repos/OpenSourceMalaria/OSM_To_Do_List
|
reopened
|
Synthesis of 6-chloropyrazin-2-amine
|
Being Synthesised Now High Priority Series 4 Synthetic Chemistry Needed
|

InChI=1S/C4H4ClN3/c5-3-1-7-2-4(6)8-3/h1-2H,(H2,6,8)
ClC1=CN=CC(N)=N1
Hi, I am Tianyi. I am a student on exchange from Nanjing University and I am working with Alice on the lab. We try to make 6-chloropyrazin-2-amine under the same conditions used to make 2-Chloro-6-hydrazinylpyrazine (AEW 85-7), but replacing hydrazine hydrate with ammonia (http://malaria.ourexperiment.org/uri/810). The TLC so far suggests only starting material but we will confirm today. I did literature search and found a patent describing synthesis of 6-chloropyrazin-2-amine using the same starting material but with aqueous solvent at 100 degrees in a sealed tube. We will try that today.http://worldwide.espacenet.com/publicationDetails/originalDocument?CC=WO&NR=2013068755A1&KC=A1&FT=D&ND=&date=20130516&DB=&&locale=en_EP
Ref https://github.com/tyzhengsydney
|
1.0
|
Synthesis of 6-chloropyrazin-2-amine - 
InChI=1S/C4H4ClN3/c5-3-1-7-2-4(6)8-3/h1-2H,(H2,6,8)
ClC1=CN=CC(N)=N1
Hi, I am Tianyi. I am a student on exchange from Nanjing University and I am working with Alice on the lab. We try to make 6-chloropyrazin-2-amine under the same conditions used to make 2-Chloro-6-hydrazinylpyrazine (AEW 85-7), but replacing hydrazine hydrate with ammonia (http://malaria.ourexperiment.org/uri/810). The TLC so far suggests only starting material but we will confirm today. I did literature search and found a patent describing synthesis of 6-chloropyrazin-2-amine using the same starting material but with aqueous solvent at 100 degrees in a sealed tube. We will try that today.http://worldwide.espacenet.com/publicationDetails/originalDocument?CC=WO&NR=2013068755A1&KC=A1&FT=D&ND=&date=20130516&DB=&&locale=en_EP
Ref https://github.com/tyzhengsydney
|
priority
|
synthesis of chloropyrazin amine inchi cn cc n hi i am tianyi i am a student on exchange from nanjing university and i am working with alice on the lab we try to make chloropyrazin amine under the same conditions used to make chloro hydrazinylpyrazine aew but replacing hydrazine hydrate with ammonia the tlc so far suggests only starting material but we will confirm today i did literature search and found a patent describing synthesis of chloropyrazin amine using the same starting material but with aqueous solvent at degrees in a sealed tube we will try that today ref
| 1
|
684,831
| 23,434,381,534
|
IssuesEvent
|
2022-08-15 08:08:28
|
BenjamooseCalto/SlasherBot
|
https://api.github.com/repos/BenjamooseCalto/SlasherBot
|
opened
|
BUG: The Steam game update check occasionally hangs, causing the check to fail.
|
bug high priority
|
This is sort of high prio, prevents the game update notifications from working.
|
1.0
|
BUG: The Steam game update check occasionally hangs, causing the check to fail. - This is sort of high prio, prevents the game update notifications from working.
|
priority
|
bug the steam game update check occasionally hangs causing the check to fail this is sort of high prio prevents the game update notifications from working
| 1
|
281,686
| 8,698,241,283
|
IssuesEvent
|
2018-12-04 22:42:35
|
broadinstitute/cromwell
|
https://api.github.com/repos/broadinstitute/cromwell
|
closed
|
Fix flakey test: KeyValueServiceActorSpec
|
HIGH PRIORITY QA Triage
|
https://broadinstitute.atlassian.net/browse/GAWB-3950
https://fc-jenkins.dsp-techops.broadinstitute.org/view/Testing/view/Test%20Runners/job/cromwell-test-runner/790/
https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/833/
https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/804/
tc: KeyValueServiceActor should insert a key/value
tc: KeyValueServiceActor should return error if key doesn't exist
tc: KeyValueServiceActor should be able to overwrite values
{quote}
org.scalatest.exceptions.TestFailedException: The future returned an exception of type: akka.pattern.AskTimeoutException, with message: Ask timed out on [Actor[akka://KeyValueServiceActorSpec/user/$a#-1019375090]] after [200000 ms]. Sender[null] sent message of type "cromwell.services.keyvalue.KeyValueServiceActor$KvPut".. at org.scalatest.concurrent.Futures$FutureConcept.tryTryAgain$1(Futures.scala:531) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl(Futures.scala:550) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl$(Futures.scala:479) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValueImpl(ScalaFutures.scala:275) at org.scalatest.concurrent.Futures$FutureConcept.futureValue(Futures.scala:476) at org.scalatest.concurrent.Futures$FutureConcept.futureValue$(Futures.scala:475) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValue(ScalaFutures.scala:275) at cromwell.services.keyvalue.impl.KeyValueServiceActorSpec.$anonfun$new$2(KeyValueServiceActorSpec.scala:46) at
{quote}
|
1.0
|
Fix flakey test: KeyValueServiceActorSpec - https://broadinstitute.atlassian.net/browse/GAWB-3950
https://fc-jenkins.dsp-techops.broadinstitute.org/view/Testing/view/Test%20Runners/job/cromwell-test-runner/790/
https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/833/
https://fc-jenkins.dsp-techops.broadinstitute.org/job/cromwell-test-runner/804/
tc: KeyValueServiceActor should insert a key/value
tc: KeyValueServiceActor should return error if key doesn't exist
tc: KeyValueServiceActor should be able to overwrite values
{quote}
org.scalatest.exceptions.TestFailedException: The future returned an exception of type: akka.pattern.AskTimeoutException, with message: Ask timed out on [Actor[akka://KeyValueServiceActorSpec/user/$a#-1019375090]] after [200000 ms]. Sender[null] sent message of type "cromwell.services.keyvalue.KeyValueServiceActor$KvPut".. at org.scalatest.concurrent.Futures$FutureConcept.tryTryAgain$1(Futures.scala:531) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl(Futures.scala:550) at org.scalatest.concurrent.Futures$FutureConcept.futureValueImpl$(Futures.scala:479) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValueImpl(ScalaFutures.scala:275) at org.scalatest.concurrent.Futures$FutureConcept.futureValue(Futures.scala:476) at org.scalatest.concurrent.Futures$FutureConcept.futureValue$(Futures.scala:475) at org.scalatest.concurrent.ScalaFutures$$anon$1.futureValue(ScalaFutures.scala:275) at cromwell.services.keyvalue.impl.KeyValueServiceActorSpec.$anonfun$new$2(KeyValueServiceActorSpec.scala:46) at
{quote}
|
priority
|
fix flakey test keyvalueserviceactorspec tc keyvalueserviceactor should insert a key value tc keyvalueserviceactor should return error if key doesn t exist tc keyvalueserviceactor should be able to overwrite values quote org scalatest exceptions testfailedexception the future returned an exception of type akka pattern asktimeoutexception with message ask timed out on after sender sent message of type cromwell services keyvalue keyvalueserviceactor kvput at org scalatest concurrent futures futureconcept trytryagain futures scala at org scalatest concurrent futures futureconcept futurevalueimpl futures scala at org scalatest concurrent futures futureconcept futurevalueimpl futures scala at org scalatest concurrent scalafutures anon futurevalueimpl scalafutures scala at org scalatest concurrent futures futureconcept futurevalue futures scala at org scalatest concurrent futures futureconcept futurevalue futures scala at org scalatest concurrent scalafutures anon futurevalue scalafutures scala at cromwell services keyvalue impl keyvalueserviceactorspec anonfun new keyvalueserviceactorspec scala at quote
| 1
|
687,598
| 23,532,999,058
|
IssuesEvent
|
2022-08-19 17:17:20
|
canonical/cos-lite-bundle
|
https://api.github.com/repos/canonical/cos-lite-bundle
|
closed
|
Integrate Loki and Alertmanager
|
Type: Enhancement Priority: High
|
Loki has an integration with Alertmanager that is currently not leveraged by the bundle. Besides the one-liner to add the missing relation, we need integration tests where Loki-generated alerts are ported through Alertmanager.
|
1.0
|
Integrate Loki and Alertmanager - Loki has an integration with Alertmanager that is currently not leveraged by the bundle. Besides the one-liner to add the missing relation, we need integration tests where Loki-generated alerts are ported through Alertmanager.
|
priority
|
integrate loki and alertmanager loki has an integration with alertmanager that is currently not leveraged by the bundle besides the one liner to add the missing relation we need integration tests where loki generated alerts are ported through alertmanager
| 1
|
740,308
| 25,743,675,746
|
IssuesEvent
|
2022-12-08 08:20:08
|
HackerN64/HackerSM64
|
https://api.github.com/repos/HackerN64/HackerSM64
|
opened
|
Split `segment2.c` into multiple files
|
high priority
|
The file contains way too many assets that could be split into multiple different categorized files. ASCII and paintings need to be merged before this split can happen.
|
1.0
|
Split `segment2.c` into multiple files - The file contains way too many assets that could be split into multiple different categorized files. ASCII and paintings need to be merged before this split can happen.
|
priority
|
split c into multiple files the file contains way too many assets that could be split into multiple different categorized files ascii and paintings need to be merged before this split can happen
| 1
|
781,006
| 27,418,288,044
|
IssuesEvent
|
2023-03-01 15:07:47
|
jj248/RealmsInExile
|
https://api.github.com/repos/jj248/RealmsInExile
|
closed
|
Evil Isengard event chain rework
|
localization lore priority: high events scripting
|
Revamp event chain to be more robust and to have multiple branching paths. Note that good Saruman altlore will be covered in a separate issue.
|
1.0
|
Evil Isengard event chain rework - Revamp event chain to be more robust and to have multiple branching paths. Note that good Saruman altlore will be covered in a separate issue.
|
priority
|
evil isengard event chain rework revamp event chain to be more robust and to have multiple branching paths note that good saruman altlore will be covered in a separate issue
| 1
|
260,170
| 8,204,713,571
|
IssuesEvent
|
2018-09-03 07:43:47
|
kowala-tech/kcoin
|
https://api.github.com/repos/kowala-tech/kcoin
|
closed
|
Review minimum validator stake
|
High priority blocker bootes-launch question research
|
The minimum stake from the whitepaper is 30,000 mTokens. The amount will likely increase over time naturally as the registry (with its default of 500 slots) fills up.
What should the hard minimum be?
Consideration factors:
- The stake should represent a significant financial interest for anyone, such that it can't be trivially thrown in an attack.
- All kUSD backers and other interested parties should be able to afford to be validators if they want to be. They shouldn't be locked out of validation over time.
- The stake shouldn't be so low that a handful of substantial token holders could occupy all the available slots by splitting their tokens and making many validators.
Possible solutions:
- Increase the number of validation slots. Since the network will slow down with a large number of validators, they can't all be active at the same time. This solution involves somehow changing the validation and consensus systems so that an unpredictable subgroup of all the registered validators in get to participate in the validation at any time. This will constitute a major change and require the delivery of a new consensus protocol on a running network.
- Introduce delegation into validation. That is, allow existing validators to accept tokens staked on their behalf in exchange for some pre-agreed, contractual validation revenue. This requires a much smaller change and no protocol updates.
|
1.0
|
Review minimum validator stake - The minimum stake from the whitepaper is 30,000 mTokens. The amount will likely increase over time naturally as the registry (with its default of 500 slots) fills up.
What should the hard minimum be?
Consideration factors:
- The stake should represent a significant financial interest for anyone, such that it can't be trivially thrown in an attack.
- All kUSD backers and other interested parties should be able to afford to be validators if they want to be. They shouldn't be locked out of validation over time.
- The stake shouldn't be so low that a handful of substantial token holders could occupy all the available slots by splitting their tokens and making many validators.
Possible solutions:
- Increase the number of validation slots. Since the network will slow down with a large number of validators, they can't all be active at the same time. This solution involves somehow changing the validation and consensus systems so that an unpredictable subgroup of all the registered validators in get to participate in the validation at any time. This will constitute a major change and require the delivery of a new consensus protocol on a running network.
- Introduce delegation into validation. That is, allow existing validators to accept tokens staked on their behalf in exchange for some pre-agreed, contractual validation revenue. This requires a much smaller change and no protocol updates.
|
priority
|
review minimum validator stake the minimum stake from the whitepaper is mtokens the amount will likely increase over time naturally as the registry with its default of slots fills up what should the hard minimum be consideration factors the stake should represent a significant financial interest for anyone such that it can t be trivially thrown in an attack all kusd backers and other interested parties should be able to afford to be validators if they want to be they shouldn t be locked out of validation over time the stake shouldn t be so low that a handful of substantial token holders could occupy all the available slots by splitting their tokens and making many validators possible solutions increase the number of validation slots since the network will slow down with a large number of validators they can t all be active at the same time this solution involves somehow changing the validation and consensus systems so that an unpredictable subgroup of all the registered validators in get to participate in the validation at any time this will constitute a major change and require the delivery of a new consensus protocol on a running network introduce delegation into validation that is allow existing validators to accept tokens staked on their behalf in exchange for some pre agreed contractual validation revenue this requires a much smaller change and no protocol updates
| 1
|
318,498
| 9,693,388,643
|
IssuesEvent
|
2019-05-24 15:57:12
|
pilosus/pilosus-pelican-theme
|
https://api.github.com/repos/pilosus/pilosus-pelican-theme
|
closed
|
Pagination for blog entries
|
bug priority: high
|
Pagination does work, but a template doesn't render url for pages:
[prev] [2] [next]
|
1.0
|
Pagination for blog entries - Pagination does work, but a template doesn't render url for pages:
[prev] [2] [next]
|
priority
|
pagination for blog entries pagination does work but a template doesn t render url for pages
| 1
|
546,764
| 16,018,846,397
|
IssuesEvent
|
2021-04-20 19:40:35
|
ORNL-AMO/VERIFI
|
https://api.github.com/repos/ORNL-AMO/VERIFI
|
closed
|
Update Examples
|
Data Input High Priority
|
Need more generic names for the example data.
What we should do is create the example accounts within verifi, including facilities/meters/data etc. Then you can export the accounts and send them to me to use for the examples.
|
1.0
|
Update Examples - Need more generic names for the example data.
What we should do is create the example accounts within verifi, including facilities/meters/data etc. Then you can export the accounts and send them to me to use for the examples.
|
priority
|
update examples need more generic names for the example data what we should do is create the example accounts within verifi including facilities meters data etc then you can export the accounts and send them to me to use for the examples
| 1
|
95,924
| 3,962,285,585
|
IssuesEvent
|
2016-05-02 16:14:01
|
salesagility/SuiteCRM
|
https://api.github.com/repos/salesagility/SuiteCRM
|
closed
|
"SQL Error : Incorrect syntax near 'LIMIT'" when using MS SQL Server 2014
|
bug High Priority
|
"SQL Error : Incorrect syntax near 'LIMIT'. " is thrown when viewing existing report output using SuiteCRM 7.5.1, connecting to MS SQL Server 2014.
Log file: "SELECT id FROM aor_fields WHERE aor_report_id = N'981b609d-47a9-9efd-c193-56bdc9373c10' AND deleted = 0 ORDER BY field_order ASC: SQL Server error: Incorrect syntax near 'LIMIT'"
The LIMIT clause is not supported by MS SQL & the appropriate connector/query needs to be updated
|
1.0
|
"SQL Error : Incorrect syntax near 'LIMIT'" when using MS SQL Server 2014 - "SQL Error : Incorrect syntax near 'LIMIT'. " is thrown when viewing existing report output using SuiteCRM 7.5.1, connecting to MS SQL Server 2014.
Log file: "SELECT id FROM aor_fields WHERE aor_report_id = N'981b609d-47a9-9efd-c193-56bdc9373c10' AND deleted = 0 ORDER BY field_order ASC: SQL Server error: Incorrect syntax near 'LIMIT'"
The LIMIT clause is not supported by MS SQL & the appropriate connector/query needs to be updated
|
priority
|
sql error incorrect syntax near limit when using ms sql server sql error incorrect syntax near limit is thrown when viewing existing report output using suitecrm connecting to ms sql server log file select id from aor fields where aor report id n and deleted order by field order asc sql server error incorrect syntax near limit the limit clause is not supported by ms sql the appropriate connector query needs to be updated
| 1
|
273,461
| 8,530,931,749
|
IssuesEvent
|
2018-11-04 05:34:31
|
cyberperspectives/sagacity
|
https://api.github.com/repos/cyberperspectives/sagacity
|
opened
|
Ingesting eChecklists results in duplicate findings in DB
|
High Priority bug
|
I load SCC results (and can do so repeatedly) and it works fine, but ingesting eChecklists results in duplicate findings in the DB. As a result, I can't export the final .ckl files.
I think there is a problem with the get_Finding() routine in database.inc that doesn't return the info when there is a finding in the database. parse_excel_echecklist.php never goes into the updated_findings routine -only the new_findings section. I put an error_log() at the end of get_Finding() and it doesn't seem to return anything.
This is urgent. We are testing this week and will need to export .ckl files at the end of the week. If we can get a patch we will be able to finish the assessment.
|
1.0
|
Ingesting eChecklists results in duplicate findings in DB - I load SCC results (and can do so repeatedly) and it works fine, but ingesting eChecklists results in duplicate findings in the DB. As a result, I can't export the final .ckl files.
I think there is a problem with the get_Finding() routine in database.inc that doesn't return the info when there is a finding in the database. parse_excel_echecklist.php never goes into the updated_findings routine -only the new_findings section. I put an error_log() at the end of get_Finding() and it doesn't seem to return anything.
This is urgent. We are testing this week and will need to export .ckl files at the end of the week. If we can get a patch we will be able to finish the assessment.
|
priority
|
ingesting echecklists results in duplicate findings in db i load scc results and can do so repeatedly and it works fine but ingesting echecklists results in duplicate findings in the db as a result i can t export the final ckl files i think there is a problem with the get finding routine in database inc that doesn t return the info when there is a finding in the database parse excel echecklist php never goes into the updated findings routine only the new findings section i put an error log at the end of get finding and it doesn t seem to return anything this is urgent we are testing this week and will need to export ckl files at the end of the week if we can get a patch we will be able to finish the assessment
| 1
|
780,835
| 27,410,318,348
|
IssuesEvent
|
2023-03-01 10:04:49
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
Remove adding unnecessary any cast when creating tuple values #32891
|
Type/Improvement Priority/High Team/CompilerFE
|
**Description:**
Every time we create a tuple value, we have added to cast to any. This is not needed for primitive types. This has 10% when we create tuple arrays.
|
1.0
|
Remove adding unnecessary any cast when creating tuple values #32891 - **Description:**
Every time we create a tuple value, we have added to cast to any. This is not needed for primitive types. This has 10% when we create tuple arrays.
|
priority
|
remove adding unnecessary any cast when creating tuple values description every time we create a tuple value we have added to cast to any this is not needed for primitive types this has when we create tuple arrays
| 1
|
344,274
| 10,342,289,457
|
IssuesEvent
|
2019-09-04 05:56:04
|
redeclipse/base
|
https://api.github.com/repos/redeclipse/base
|
closed
|
[mac os][bug] Couldn't launch the game with last update
|
branch: stable difficulty: easy priority: high status: bug
|
MacOS Mojave 14.6
I don't know what libsteam license is, but I thought you can link it statically.
Console output:
```
bins: 2e9ede6132ca3ee7748c6919e7be15db6b638924 is the current version.
dyld: Library not loaded: @rpath/libsteam_api.dylib
Referenced from: /Applications/redeclipse.app/Contents/Resources/bin/redeclipse.app/Contents/MacOS/redeclipse_universal
Reason: image not found
Abort trap: 6
logout
...bla-bla-bla
```
- Version and build: [2e9ede6132ca3ee7748c6919e7be15db6b638924]
|
1.0
|
[mac os][bug] Couldn't launch the game with last update - MacOS Mojave 14.6
I don't know what libsteam license is, but I thought you can link it statically.
Console output:
```
bins: 2e9ede6132ca3ee7748c6919e7be15db6b638924 is the current version.
dyld: Library not loaded: @rpath/libsteam_api.dylib
Referenced from: /Applications/redeclipse.app/Contents/Resources/bin/redeclipse.app/Contents/MacOS/redeclipse_universal
Reason: image not found
Abort trap: 6
logout
...bla-bla-bla
```
- Version and build: [2e9ede6132ca3ee7748c6919e7be15db6b638924]
|
priority
|
couldn t launch the game with last update macos mojave i don t know what libsteam license is but i thought you can link it statically console output bins is the current version dyld library not loaded rpath libsteam api dylib referenced from applications redeclipse app contents resources bin redeclipse app contents macos redeclipse universal reason image not found abort trap logout bla bla bla version and build
| 1
|
463,185
| 13,261,427,118
|
IssuesEvent
|
2020-08-20 19:52:46
|
ChainSafe/forest
|
https://api.github.com/repos/ChainSafe/forest
|
closed
|
Change to using sha2-256 hashing instead of blake2b
|
IPLD Priority: 2 - High Spec Change Status: On Ice
|
This is a change to the spec for performance reasons.
|
1.0
|
Change to using sha2-256 hashing instead of blake2b - This is a change to the spec for performance reasons.
|
priority
|
change to using hashing instead of this is a change to the spec for performance reasons
| 1
|
546,659
| 16,016,965,985
|
IssuesEvent
|
2021-04-20 17:13:31
|
ESCOMP/CTSM
|
https://api.github.com/repos/ESCOMP/CTSM
|
closed
|
Add landunit_mask (formerly l2g_scale_type) metadata to history file
|
priority: high type: enhancement
|
Connected with #1297 - We should add the landunit_mask attribute (currently named l2g_scale_type, but intended to be renamed in #1342) to each field in the history file. Since #1342 suggests changing strings to integer constants, we'll need to have a translation from these integer constants back to human-readable strings when writing out the history metadata.
I'm not sure what to do for fields that don't have a l2g_scale_type / landunit_mask attribute. Long-term, once we have made everything consistent, I think we probably want this attribute on all history fields (it will just be something like 'all' for many fields). But for now, that would be misleading, because many fields currently do masking via the initial setting to spval everywhere. So probably for now we should only add this attribute to fields for which l2g_scale_type / landunit_mask is explicitly defined.
|
1.0
|
Add landunit_mask (formerly l2g_scale_type) metadata to history file - Connected with #1297 - We should add the landunit_mask attribute (currently named l2g_scale_type, but intended to be renamed in #1342) to each field in the history file. Since #1342 suggests changing strings to integer constants, we'll need to have a translation from these integer constants back to human-readable strings when writing out the history metadata.
I'm not sure what to do for fields that don't have a l2g_scale_type / landunit_mask attribute. Long-term, once we have made everything consistent, I think we probably want this attribute on all history fields (it will just be something like 'all' for many fields). But for now, that would be misleading, because many fields currently do masking via the initial setting to spval everywhere. So probably for now we should only add this attribute to fields for which l2g_scale_type / landunit_mask is explicitly defined.
|
priority
|
add landunit mask formerly scale type metadata to history file connected with we should add the landunit mask attribute currently named scale type but intended to be renamed in to each field in the history file since suggests changing strings to integer constants we ll need to have a translation from these integer constants back to human readable strings when writing out the history metadata i m not sure what to do for fields that don t have a scale type landunit mask attribute long term once we have made everything consistent i think we probably want this attribute on all history fields it will just be something like all for many fields but for now that would be misleading because many fields currently do masking via the initial setting to spval everywhere so probably for now we should only add this attribute to fields for which scale type landunit mask is explicitly defined
| 1
|
290,451
| 8,895,386,565
|
IssuesEvent
|
2019-01-16 08:30:37
|
gluster/glusterd2
|
https://api.github.com/repos/gluster/glusterd2
|
closed
|
Volume creation is failing after 59 volumes with 30 node cluster
|
FW: Volume Management GCS/1.0-Blocker priority: high
|
-> Created 30 node cluster
-> Tried to create 200 volumes through script
-> It's distributed-replica volume 10x3 (means each volume will be having 30 bricks )
-> One brick from each node that means 30 bricks form 30 nodes
-> Volume creation failing with below error
Volume creation failed
Error: Request failed with HTTP Status code 500
Transaction step vol-create.StoreVolume failed on peer 5a011251-efb4-4efe-b9dc-b25b345c6634 with error: etcdserver: request is too large
i didn't find any logs for volume creation failed
After that I have tried to create single brick volume manually without script on 30 node cluster setup.
actually volume creation is successful, but got error message like below
./glustercli volume create vol1 10.70.42.199:/bricks/brick0/vol1
error message:
Volume creation failed
Error: Post http://127.0.0.1:24007/v1/volumes: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
Then tried to start the volume, failed with below error
/glustercli volume start vol1
volume start failed
Error: Post http://127.0.0.1:24007/v1/volumes/vol1/start: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
Again tried to start the volume, this time volume started successfully
./glustercli volume start vol1
Volume vol1 started successfully
|
1.0
|
Volume creation is failing after 59 volumes with 30 node cluster - -> Created 30 node cluster
-> Tried to create 200 volumes through script
-> It's distributed-replica volume 10x3 (means each volume will be having 30 bricks )
-> One brick from each node that means 30 bricks form 30 nodes
-> Volume creation failing with below error
Volume creation failed
Error: Request failed with HTTP Status code 500
Transaction step vol-create.StoreVolume failed on peer 5a011251-efb4-4efe-b9dc-b25b345c6634 with error: etcdserver: request is too large
i didn't find any logs for volume creation failed
After that I have tried to create single brick volume manually without script on 30 node cluster setup.
actually volume creation is successful, but got error message like below
./glustercli volume create vol1 10.70.42.199:/bricks/brick0/vol1
error message:
Volume creation failed
Error: Post http://127.0.0.1:24007/v1/volumes: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
Then tried to start the volume, failed with below error
/glustercli volume start vol1
volume start failed
Error: Post http://127.0.0.1:24007/v1/volumes/vol1/start: net/http: request canceled (Client.Timeout exceeded while awaiting headers)
Again tried to start the volume, this time volume started successfully
./glustercli volume start vol1
Volume vol1 started successfully
|
priority
|
volume creation is failing after volumes with node cluster created node cluster tried to create volumes through script it s distributed replica volume means each volume will be having bricks one brick from each node that means bricks form nodes volume creation failing with below error volume creation failed error request failed with http status code transaction step vol create storevolume failed on peer with error etcdserver request is too large i didn t find any logs for volume creation failed after that i have tried to create single brick volume manually without script on node cluster setup actually volume creation is successful but got error message like below glustercli volume create bricks error message volume creation failed error post net http request canceled client timeout exceeded while awaiting headers then tried to start the volume failed with below error glustercli volume start volume start failed error post net http request canceled client timeout exceeded while awaiting headers again tried to start the volume this time volume started successfully glustercli volume start volume started successfully
| 1
|
203,708
| 7,077,876,747
|
IssuesEvent
|
2018-01-10 00:11:21
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
closed
|
Provide EndpointRequest for configuring WebFlux-based Security
|
priority: high theme: security type: enhancement
|
Using Spring Boot, SpringMVC and Spring Security I can configure the security part of my Spring Boot app as follows:
```
...requestMatchers(EndpointRequest.to("status", "info"))...
```
Now, I'm migrating from SpringMVC to Spring WebFlux. I see that I've to use `pathMatchers()` instead of `requestMatchers` from Spring Security. However, `EndpointRequest` is depending on `HttpServletRequest` (and thus on Spring MVC).
In https://stackoverflow.com/questions/47287312/spring-security-with-webflux-how-to-migrate-requestmatchersendpointrequest-to#comment-81531115 @bclozel suggests to open an issue.
|
1.0
|
Provide EndpointRequest for configuring WebFlux-based Security - Using Spring Boot, SpringMVC and Spring Security I can configure the security part of my Spring Boot app as follows:
```
...requestMatchers(EndpointRequest.to("status", "info"))...
```
Now, I'm migrating from SpringMVC to Spring WebFlux. I see that I've to use `pathMatchers()` instead of `requestMatchers` from Spring Security. However, `EndpointRequest` is depending on `HttpServletRequest` (and thus on Spring MVC).
In https://stackoverflow.com/questions/47287312/spring-security-with-webflux-how-to-migrate-requestmatchersendpointrequest-to#comment-81531115 @bclozel suggests to open an issue.
|
priority
|
provide endpointrequest for configuring webflux based security using spring boot springmvc and spring security i can configure the security part of my spring boot app as follows requestmatchers endpointrequest to status info now i m migrating from springmvc to spring webflux i see that i ve to use pathmatchers instead of requestmatchers from spring security however endpointrequest is depending on httpservletrequest and thus on spring mvc in bclozel suggests to open an issue
| 1
|
455,270
| 13,123,276,703
|
IssuesEvent
|
2020-08-06 00:03:02
|
ampproject/amphtml
|
https://api.github.com/repos/ampproject/amphtml
|
opened
|
[amp-story-player] PE: only preload distance = 1 iframes
|
P1: High Priority Type: Feature Request WG: stories
|
Today the Player always preloads 3 stories, when it should only preload N - 1, N, N + 1. This means one less story to load for the initial render.
cc @ampproject/wg-stories
|
1.0
|
[amp-story-player] PE: only preload distance = 1 iframes - Today the Player always preloads 3 stories, when it should only preload N - 1, N, N + 1. This means one less story to load for the initial render.
cc @ampproject/wg-stories
|
priority
|
pe only preload distance iframes today the player always preloads stories when it should only preload n n n this means one less story to load for the initial render cc ampproject wg stories
| 1
|
456,299
| 13,148,898,988
|
IssuesEvent
|
2020-08-09 00:59:07
|
HumanExposure/factotum
|
https://api.github.com/repos/HumanExposure/factotum
|
opened
|
Security Issue with Pillow
|
Priority :: High To point security
|
Out of Bounds read in Pillow. Upgrade Pillow to version 7.1.0 or later. For example:
`Pillow>=7.1.0`
|
1.0
|
Security Issue with Pillow - Out of Bounds read in Pillow. Upgrade Pillow to version 7.1.0 or later. For example:
`Pillow>=7.1.0`
|
priority
|
security issue with pillow out of bounds read in pillow upgrade pillow to version or later for example pillow
| 1
|
369,807
| 10,918,207,852
|
IssuesEvent
|
2019-11-21 16:28:41
|
carbon-design-system/carbon
|
https://api.github.com/repos/carbon-design-system/carbon
|
reopened
|
AVT 1 - Code Snippet -> Inline has a DAP violation
|
package: react priority: high type: a11y ♿
|
## Environment
macOS Mojave version 10.14.6
Chrome Version Version 77.0.3865.90
Carbon v10 - React
DAP IBM Accessibility 1.8.0.0 Sept 2019 Ruleset
## Detailed Description
Go to the React Component
Run DAP and the following violations are present (see screenshot):
<img width="914" alt="Screen Shot 2019-11-01 at 12 20 22 PM" src="https://user-images.githubusercontent.com/21676914/68042886-02991880-fca2-11e9-910a-943cb256dd6f.png">
Note: "node -v" which is the button text does not match the aria-label="copyable code snippet".
|
1.0
|
AVT 1 - Code Snippet -> Inline has a DAP violation - ## Environment
macOS Mojave version 10.14.6
Chrome Version Version 77.0.3865.90
Carbon v10 - React
DAP IBM Accessibility 1.8.0.0 Sept 2019 Ruleset
## Detailed Description
Go to the React Component
Run DAP and the following violations are present (see screenshot):
<img width="914" alt="Screen Shot 2019-11-01 at 12 20 22 PM" src="https://user-images.githubusercontent.com/21676914/68042886-02991880-fca2-11e9-910a-943cb256dd6f.png">
Note: "node -v" which is the button text does not match the aria-label="copyable code snippet".
|
priority
|
avt code snippet inline has a dap violation environment macos mojave version chrome version version carbon react dap ibm accessibility sept ruleset detailed description go to the react component run dap and the following violations are present see screenshot img width alt screen shot at pm src note node v which is the button text does not match the aria label copyable code snippet
| 1
|
719,957
| 24,774,463,145
|
IssuesEvent
|
2022-10-23 15:01:29
|
signum-network/signum-node
|
https://api.github.com/repos/signum-network/signum-node
|
closed
|
New API call getTradeJournal
|
enhancement priority-high API
|
Currently we have single API calls to request orders or trades for account(s) and token(s).
This is technical okay, but those calls are not fitting into the nedd for a DeFi portal.
To handle the requirement we propose to create a new API call : `getTradeJournal`
Parameter to enter :
- Account-ID
- Token-ID
The call is always for a given account + token.
The JSON output looks like this:
`
{
account: "ID",
token:"ID",
tradeJournal:
[
order-id:"tx-id",
timestamp:"tx-timestamp",
status: integer,
tradeType:"buy or sell"
limitNQT: "limit from tx",
quantityQNT:"quantity from tx",
averagePriceOfOneQNT: float,
executedAmountQNT:float,
executionSlides:
[
trade-id:"trade-id",
timestamp:"trade-tx",
height:integer,
quantityQNT:"quantity from trade",
priceOfOneQNT:"price from trade"
],
]
}
`
**Description**
- account: The account from the request
- token: The token from the request
- `tradeJournal`: The section with the list of open/closed and canceled orders for the given account/token
- order-id: The ID from the transaction which creates the buy or sell order
- timestamp: The timestamp from the transaction which creates the buy or sell order
- status : A integer value with the following range:
- (-1) order is caneled
- (0) order is open
- (1 until 99) partial fill ofd the order in %
- (100) order is closed ( 100% filled)
- tradeType: Direction of the order (buy/sell)
- limitNQT: The limit set on the transaction in Planck ( converted)
- quantityQNT: The amount of tokens to buy or sell from the original transaction
- averagePriceOfOneQNT : The calcualted execution price by all `executionSlides`
- executedAmountQNT: The sum of all amount by the `executionSlides`
- `trades` : In this section we show all trades related to the order-id
- trade-id : ID of the trade
- timestamp: Timestamp of the trade
- height : Block-ID
- quantityQNT : Executed amount
- priceOfOneQNT: Excecuted price for the quantityQNT
**Logic**
We should loop over all transaction from the account for Place BID/ASK orders filtered by asset(token)
In this case we have the main list of all orders from the user.
Now we loop over all trades by the order-ids. We filter on asset-id (token) and check ask_order_id/bid_order_id.
Finally we check all Cancel transaction and set those order-ids to status -1
The NQT fields should be expressed in Planck ( see #654 )
This ticket covers the following requirements:
- #657
- #656
- #655
|
1.0
|
New API call getTradeJournal - Currently we have single API calls to request orders or trades for account(s) and token(s).
This is technical okay, but those calls are not fitting into the nedd for a DeFi portal.
To handle the requirement we propose to create a new API call : `getTradeJournal`
Parameter to enter :
- Account-ID
- Token-ID
The call is always for a given account + token.
The JSON output looks like this:
`
{
account: "ID",
token:"ID",
tradeJournal:
[
order-id:"tx-id",
timestamp:"tx-timestamp",
status: integer,
tradeType:"buy or sell"
limitNQT: "limit from tx",
quantityQNT:"quantity from tx",
averagePriceOfOneQNT: float,
executedAmountQNT:float,
executionSlides:
[
trade-id:"trade-id",
timestamp:"trade-tx",
height:integer,
quantityQNT:"quantity from trade",
priceOfOneQNT:"price from trade"
],
]
}
`
**Description**
- account: The account from the request
- token: The token from the request
- `tradeJournal`: The section with the list of open/closed and canceled orders for the given account/token
- order-id: The ID from the transaction which creates the buy or sell order
- timestamp: The timestamp from the transaction which creates the buy or sell order
- status : A integer value with the following range:
- (-1) order is caneled
- (0) order is open
- (1 until 99) partial fill ofd the order in %
- (100) order is closed ( 100% filled)
- tradeType: Direction of the order (buy/sell)
- limitNQT: The limit set on the transaction in Planck ( converted)
- quantityQNT: The amount of tokens to buy or sell from the original transaction
- averagePriceOfOneQNT : The calcualted execution price by all `executionSlides`
- executedAmountQNT: The sum of all amount by the `executionSlides`
- `trades` : In this section we show all trades related to the order-id
- trade-id : ID of the trade
- timestamp: Timestamp of the trade
- height : Block-ID
- quantityQNT : Executed amount
- priceOfOneQNT: Excecuted price for the quantityQNT
**Logic**
We should loop over all transaction from the account for Place BID/ASK orders filtered by asset(token)
In this case we have the main list of all orders from the user.
Now we loop over all trades by the order-ids. We filter on asset-id (token) and check ask_order_id/bid_order_id.
Finally we check all Cancel transaction and set those order-ids to status -1
The NQT fields should be expressed in Planck ( see #654 )
This ticket covers the following requirements:
- #657
- #656
- #655
|
priority
|
new api call gettradejournal currently we have single api calls to request orders or trades for account s and token s this is technical okay but those calls are not fitting into the nedd for a defi portal to handle the requirement we propose to create a new api call gettradejournal parameter to enter account id token id the call is always for a given account token the json output looks like this account id token id tradejournal order id tx id timestamp tx timestamp status integer tradetype buy or sell limitnqt limit from tx quantityqnt quantity from tx averagepriceofoneqnt float executedamountqnt float executionslides trade id trade id timestamp trade tx height integer quantityqnt quantity from trade priceofoneqnt price from trade description account the account from the request token the token from the request tradejournal the section with the list of open closed and canceled orders for the given account token order id the id from the transaction which creates the buy or sell order timestamp the timestamp from the transaction which creates the buy or sell order status a integer value with the following range order is caneled order is open until partial fill ofd the order in order is closed filled tradetype direction of the order buy sell limitnqt the limit set on the transaction in planck converted quantityqnt the amount of tokens to buy or sell from the original transaction averagepriceofoneqnt the calcualted execution price by all executionslides executedamountqnt the sum of all amount by the executionslides trades in this section we show all trades related to the order id trade id id of the trade timestamp timestamp of the trade height block id quantityqnt executed amount priceofoneqnt excecuted price for the quantityqnt logic we should loop over all transaction from the account for place bid ask orders filtered by asset token in this case we have the main list of all orders from the user now we loop over all trades by the order ids we filter on asset id token and check ask order id bid order id finally we check all cancel transaction and set those order ids to status the nqt fields should be expressed in planck see this ticket covers the following requirements
| 1
|
577,163
| 17,104,515,216
|
IssuesEvent
|
2021-07-09 15:40:13
|
huridocs/uwazi
|
https://api.github.com/repos/huridocs/uwazi
|
closed
|
Filter and Aggregate by Inherited metadata (30pts)
|
Priority: High
|
Expected features:
- Filter by inherited metadata instead of the title of the inherited entity
- Show appropriate aggregations for each value.
- Inherited filters should match with common filters.
Implementation decisions
- ~Template properties will have type (text, relationship...) and filterType (relationships will use inherited type)~
- ~The inheritedValue will be moved from inheritedValue to value~
- ~The value will be moved to ref {value: label:} or ref / refLabel~
- ~The label should be replaced with the inherited property label~
- Do not allow inheriting from inherited
Tips
- Make sure copy from copies everything and not just the value
- check permissions ??
- Check if inherited properties are exported with the correct value and just report it but do not develop
Example of inherited values
```
{
sharedId: 456,
title: 'Konz the famous magician',
template: A,
metadata: {
favourite-colors: [{value: 28s7d3, label: 'green'}, {value: 234js9d3, label: 'blue'}],
},
}
{
sharedId: 859,
title: 'Rafa the copy cat',
template: B,
metadata: {
inherited-color: [{value: [28s7d3, 234js9d3], label: ['green', 'blue'], ref: {value: '456', label: 'Konz the famous magician'}}],
},
}
```
|
1.0
|
Filter and Aggregate by Inherited metadata (30pts) - Expected features:
- Filter by inherited metadata instead of the title of the inherited entity
- Show appropriate aggregations for each value.
- Inherited filters should match with common filters.
Implementation decisions
- ~Template properties will have type (text, relationship...) and filterType (relationships will use inherited type)~
- ~The inheritedValue will be moved from inheritedValue to value~
- ~The value will be moved to ref {value: label:} or ref / refLabel~
- ~The label should be replaced with the inherited property label~
- Do not allow inheriting from inherited
Tips
- Make sure copy from copies everything and not just the value
- check permissions ??
- Check if inherited properties are exported with the correct value and just report it but do not develop
Example of inherited values
```
{
sharedId: 456,
title: 'Konz the famous magician',
template: A,
metadata: {
favourite-colors: [{value: 28s7d3, label: 'green'}, {value: 234js9d3, label: 'blue'}],
},
}
{
sharedId: 859,
title: 'Rafa the copy cat',
template: B,
metadata: {
inherited-color: [{value: [28s7d3, 234js9d3], label: ['green', 'blue'], ref: {value: '456', label: 'Konz the famous magician'}}],
},
}
```
|
priority
|
filter and aggregate by inherited metadata expected features filter by inherited metadata instead of the title of the inherited entity show appropriate aggregations for each value inherited filters should match with common filters implementation decisions template properties will have type text relationship and filtertype relationships will use inherited type the inheritedvalue will be moved from inheritedvalue to value the value will be moved to ref value label or ref reflabel the label should be replaced with the inherited property label do not allow inheriting from inherited tips make sure copy from copies everything and not just the value check permissions check if inherited properties are exported with the correct value and just report it but do not develop example of inherited values sharedid title konz the famous magician template a metadata favourite colors sharedid title rafa the copy cat template b metadata inherited color label ref value label konz the famous magician
| 1
|
819,507
| 30,739,158,946
|
IssuesEvent
|
2023-07-28 10:01:56
|
MaFi-Games/Captain-of-Industry-issues
|
https://api.github.com/repos/MaFi-Games/Captain-of-Industry-issues
|
closed
|
Its dark i cant see.
|
could not reproduce high priority
|
### Issue description
So with the newest update i found a bug that crashes my game every time. So i unlocked the transformer for my ppl but every time i place it i get the white boxs that has the Captain of Industry logo on it and my game crashes. thankfully autosave is on but it dark out here my ppl cant see when they wake up at 4 am to go to work and no coffee OMG there so slow now haha.
### Expected behavior
just expected it to place. i did this 2 times till i figured out what it was.
### Reproduction steps
1. click on transformer
2. place transformer
3. disapointment has arrived
4. game crashed
### Game version
v0.4.13
### Attachments
https://1drv.ms/u/s!ArMq3s42zGYsgQUB0hvAyNvBGQXi?e=rXHZpr
|
1.0
|
Its dark i cant see. - ### Issue description
So with the newest update i found a bug that crashes my game every time. So i unlocked the transformer for my ppl but every time i place it i get the white boxs that has the Captain of Industry logo on it and my game crashes. thankfully autosave is on but it dark out here my ppl cant see when they wake up at 4 am to go to work and no coffee OMG there so slow now haha.
### Expected behavior
just expected it to place. i did this 2 times till i figured out what it was.
### Reproduction steps
1. click on transformer
2. place transformer
3. disapointment has arrived
4. game crashed
### Game version
v0.4.13
### Attachments
https://1drv.ms/u/s!ArMq3s42zGYsgQUB0hvAyNvBGQXi?e=rXHZpr
|
priority
|
its dark i cant see issue description so with the newest update i found a bug that crashes my game every time so i unlocked the transformer for my ppl but every time i place it i get the white boxs that has the captain of industry logo on it and my game crashes thankfully autosave is on but it dark out here my ppl cant see when they wake up at am to go to work and no coffee omg there so slow now haha expected behavior just expected it to place i did this times till i figured out what it was reproduction steps click on transformer place transformer disapointment has arrived game crashed game version attachments
| 1
|
172,827
| 6,516,704,776
|
IssuesEvent
|
2017-08-27 13:16:54
|
fossasia/open-event
|
https://api.github.com/repos/fossasia/open-event
|
closed
|
Update all the samples according to the new API specifications [ On Hold ]
|
has-PR Priority HIGH
|
The samples in the repo have become inconsistent. [Mozilla All Hands 2017](https://github.com/fossasia/open-event/pull/191/files) and [FOSSASIA 2016](https://github.com/fossasia/open-event/commit/77783374aca8b9cc2849484564dbd9cab54aff37) are using the new attributes as defined by the new API specifications while the other events are using the old attributes. Since we are removing support for the old attributes from the web and Android generators too, it would be good to convert all the remaining samples too according to the new API specs. Otherwise, a part of the samples (either old or new) won't be working anymore. Please suggest. Thanks :)
@mariobehling @niranjan94 @aayusharora @Shailesh351
|
1.0
|
Update all the samples according to the new API specifications [ On Hold ] - The samples in the repo have become inconsistent. [Mozilla All Hands 2017](https://github.com/fossasia/open-event/pull/191/files) and [FOSSASIA 2016](https://github.com/fossasia/open-event/commit/77783374aca8b9cc2849484564dbd9cab54aff37) are using the new attributes as defined by the new API specifications while the other events are using the old attributes. Since we are removing support for the old attributes from the web and Android generators too, it would be good to convert all the remaining samples too according to the new API specs. Otherwise, a part of the samples (either old or new) won't be working anymore. Please suggest. Thanks :)
@mariobehling @niranjan94 @aayusharora @Shailesh351
|
priority
|
update all the samples according to the new api specifications the samples in the repo have become inconsistent and are using the new attributes as defined by the new api specifications while the other events are using the old attributes since we are removing support for the old attributes from the web and android generators too it would be good to convert all the remaining samples too according to the new api specs otherwise a part of the samples either old or new won t be working anymore please suggest thanks mariobehling aayusharora
| 1
|
354,736
| 10,571,552,139
|
IssuesEvent
|
2019-10-07 07:27:50
|
hotosm/tasking-manager
|
https://api.github.com/repos/hotosm/tasking-manager
|
closed
|
Cannot execute migrations for staging deployment
|
Component: Backend Difficulty: Medium Priority: High Status: In Progress Type: Bug
|
Per chat with @dakotabenjamin and @thadk. Migration script returns the following error within staging server:
```ascii' codec can't decode byte 0xc3 in position 5067619: ordinal not in range(128)```
This error is present within migration file `migrations/versions/0eee8c1abd3a_.py", line 60`
|
1.0
|
Cannot execute migrations for staging deployment - Per chat with @dakotabenjamin and @thadk. Migration script returns the following error within staging server:
```ascii' codec can't decode byte 0xc3 in position 5067619: ordinal not in range(128)```
This error is present within migration file `migrations/versions/0eee8c1abd3a_.py", line 60`
|
priority
|
cannot execute migrations for staging deployment per chat with dakotabenjamin and thadk migration script returns the following error within staging server ascii codec can t decode byte in position ordinal not in range this error is present within migration file migrations versions py line
| 1
|
342,471
| 10,317,599,834
|
IssuesEvent
|
2019-08-30 13:07:13
|
pmem/issues
|
https://api.github.com/repos/pmem/issues
|
closed
|
Test: pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck)
|
Exposure: Medium OS: Linux Priority: 2 high Type: Bug
|
<!--
Before creating new issue, ensure that similar issue wasn't already created
* Search: https://github.com/pmem/issues/issues
Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report.
Remember this is just a minimal template. You can extend it with data you think may be useful.
-->
# ISSUE: <!-- fill the title of issue -->
## Environment Information
- PMDK package version(s): 1.4.3-rc1
- OS(es) version(s): SLES 12.4
- ndctl version(s): 61.2
- kernel version(s): 4.12.14-95.29-default
## Please provide a reproduction of the bug:
```
./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all
```
## How often bug is revealed: (always, often, rare): always
<!-- describe special circumstances in section above -->
```
./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all
pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck)
pmempool_transform/TEST18 failed with Valgrind. See memcheck18.log. First 20 lines below.
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x2f5c00000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1bae00000, 0x1fae00000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x33ae00000, 0x37adff000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132==
pmempool_transform/TEST18 memcheck18.log ==31132== HEAP SUMMARY:
pmempool_transform/TEST18 memcheck18.log ==31132== in use at exit: 0 bytes in 0 blocks
pmempool_transform/TEST18 memcheck18.log ==31132== total heap usage: 1,148 allocs, 1,148 frees, 3,619,892 bytes allocated
pmempool_transform/TEST18 memcheck18.log ==31132==
pmempool_transform/TEST18 memcheck18.log ==31132== All heap blocks were freed -- no leaks are possible
pmempool_transform/TEST18 memcheck18.log ==31132==
pmempool_transform/TEST18 memcheck18.log ==31132== For counts of detected and suppressed errors, rerun with: -v
pmempool_transform/TEST18 memcheck18.log ==31132== ERROR SUMMARY: 6 errors from 3 contexts (suppressed: 0 from 0)
RUNTESTS: stopping: pmempool_transform/TEST18 failed, TEST=all FS=any BUILD=debug
```
<!-- fill this out -->
## Expected behavior:
Test should pass.
## Details
<!-- fill this out -->
## Additional information about Priority and Help Requested:
Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible -->
Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
|
1.0
|
Test: pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck) - <!--
Before creating new issue, ensure that similar issue wasn't already created
* Search: https://github.com/pmem/issues/issues
Note that if you do not provide enough information to reproduce the issue, we may not be able to take action on your report.
Remember this is just a minimal template. You can extend it with data you think may be useful.
-->
# ISSUE: <!-- fill the title of issue -->
## Environment Information
- PMDK package version(s): 1.4.3-rc1
- OS(es) version(s): SLES 12.4
- ndctl version(s): 61.2
- kernel version(s): 4.12.14-95.29-default
## Please provide a reproduction of the bug:
```
./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all
```
## How often bug is revealed: (always, often, rare): always
<!-- describe special circumstances in section above -->
```
./RUNTESTS pmempool_transform -s TEST18 -m force-enable -t all
pmempool_transform/TEST18: SETUP (all/pmem/debug/memcheck)
pmempool_transform/TEST18 failed with Valgrind. See memcheck18.log. First 20 lines below.
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x2f5c00000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x59e3e000, 0x214c3e000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1bae00000, 0x1fae00000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x1fae00000, 0x3b5bff000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x33ae00000, 0x37adff000) (defined)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x200000000, 0x37adff000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132== Warning: set address range perms: large range [0x80000000, 0x1fae00000) (noaccess)
pmempool_transform/TEST18 memcheck18.log ==31132==
pmempool_transform/TEST18 memcheck18.log ==31132== HEAP SUMMARY:
pmempool_transform/TEST18 memcheck18.log ==31132== in use at exit: 0 bytes in 0 blocks
pmempool_transform/TEST18 memcheck18.log ==31132== total heap usage: 1,148 allocs, 1,148 frees, 3,619,892 bytes allocated
pmempool_transform/TEST18 memcheck18.log ==31132==
pmempool_transform/TEST18 memcheck18.log ==31132== All heap blocks were freed -- no leaks are possible
pmempool_transform/TEST18 memcheck18.log ==31132==
pmempool_transform/TEST18 memcheck18.log ==31132== For counts of detected and suppressed errors, rerun with: -v
pmempool_transform/TEST18 memcheck18.log ==31132== ERROR SUMMARY: 6 errors from 3 contexts (suppressed: 0 from 0)
RUNTESTS: stopping: pmempool_transform/TEST18 failed, TEST=all FS=any BUILD=debug
```
<!-- fill this out -->
## Expected behavior:
Test should pass.
## Details
<!-- fill this out -->
## Additional information about Priority and Help Requested:
Are you willing to submit a pull request with a proposed change? (Yes, No) <!-- check one if possible -->
Requested priority: (Showstopper, High, Medium, Low) <!-- check one if possible -->
|
priority
|
test pmempool transform setup all pmem debug memcheck before creating new issue ensure that similar issue wasn t already created search note that if you do not provide enough information to reproduce the issue we may not be able to take action on your report remember this is just a minimal template you can extend it with data you think may be useful issue environment information pmdk package version s os es version s sles ndctl version s kernel version s default please provide a reproduction of the bug runtests pmempool transform s m force enable t all how often bug is revealed always often rare always runtests pmempool transform s m force enable t all pmempool transform setup all pmem debug memcheck pmempool transform failed with valgrind see log first lines below pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range defined pmempool transform log warning set address range perms large range noaccess pmempool transform log warning set address range perms large range noaccess pmempool transform log pmempool transform log heap summary pmempool transform log in use at exit bytes in blocks pmempool transform log total heap usage allocs frees bytes allocated pmempool transform log pmempool transform log all heap blocks were freed no leaks are possible pmempool transform log pmempool transform log for counts of detected and suppressed errors rerun with v pmempool transform log error summary errors from contexts suppressed from runtests stopping pmempool transform failed test all fs any build debug expected behavior test should pass details additional information about priority and help requested are you willing to submit a pull request with a proposed change yes no requested priority showstopper high medium low
| 1
|
215,543
| 7,295,073,962
|
IssuesEvent
|
2018-02-26 04:29:18
|
utopian-io/utopian.io
|
https://api.github.com/repos/utopian-io/utopian.io
|
closed
|
Improve GitHub Integration
|
feature high-priority review
|
Show the synced github account on the contribution details and only allow to add a contribution on the development category if github is synced. If a user has no github account connected show a notice if he tries to create a development contribution
|
1.0
|
Improve GitHub Integration - Show the synced github account on the contribution details and only allow to add a contribution on the development category if github is synced. If a user has no github account connected show a notice if he tries to create a development contribution
|
priority
|
improve github integration show the synced github account on the contribution details and only allow to add a contribution on the development category if github is synced if a user has no github account connected show a notice if he tries to create a development contribution
| 1
|
205,760
| 7,105,799,394
|
IssuesEvent
|
2018-01-16 14:49:37
|
StratoDem/sd-material-ui
|
https://api.github.com/repos/StratoDem/sd-material-ui
|
opened
|
Buttons width is not adjusting correctly
|
Priority: High Tech: JS Type: Bug
|
The SDRaisedButton does not respond properly to width adjustment. The button itself can be reduced in width, but its containing Div does not shrink, and can disrupt the page's styling when it leaves a blank space.
Change the SDRaisedButton (and SDFlatButton if necessary) to allow it to change style properly with the button.

|
1.0
|
Buttons width is not adjusting correctly - The SDRaisedButton does not respond properly to width adjustment. The button itself can be reduced in width, but its containing Div does not shrink, and can disrupt the page's styling when it leaves a blank space.
Change the SDRaisedButton (and SDFlatButton if necessary) to allow it to change style properly with the button.

|
priority
|
buttons width is not adjusting correctly the sdraisedbutton does not respond properly to width adjustment the button itself can be reduced in width but its containing div does not shrink and can disrupt the page s styling when it leaves a blank space change the sdraisedbutton and sdflatbutton if necessary to allow it to change style properly with the button
| 1
|
659,634
| 21,935,585,039
|
IssuesEvent
|
2022-05-23 13:32:37
|
CarnegieLearningWeb/UpGrade
|
https://api.github.com/repos/CarnegieLearningWeb/UpGrade
|
reopened
|
Patch code for XPRIZE and Demonstration phase
|
priority: high
|
We need to decide what to do with the current xprize patch code and how it should be maintained if we're advanced to the demonstration phase (probably starting in Sept 22). Do we need to create a more permanent solution to the issue of experiment point/id conflicts?
|
1.0
|
Patch code for XPRIZE and Demonstration phase - We need to decide what to do with the current xprize patch code and how it should be maintained if we're advanced to the demonstration phase (probably starting in Sept 22). Do we need to create a more permanent solution to the issue of experiment point/id conflicts?
|
priority
|
patch code for xprize and demonstration phase we need to decide what to do with the current xprize patch code and how it should be maintained if we re advanced to the demonstration phase probably starting in sept do we need to create a more permanent solution to the issue of experiment point id conflicts
| 1
|
723,549
| 24,901,315,059
|
IssuesEvent
|
2022-10-28 21:15:13
|
virtualcell/vcell
|
https://api.github.com/repos/virtualcell/vcell
|
closed
|
Orphan overrides due to init/runtime naming issues
|
bug High Priority VCell-7.5.0
|
This issue affects clamped variables. Let's assume species EGF is clamped with some number or expression.
Previously math generation would generate two constants, named EGF and EGF_init, both of which would be listed in the math and in the table for overrides. But only one of them (the first one) would actually be used in the equations.
If the user would override EGF_init, nothing would actually happen (which in itself is a serious issue). But in most cases both of them would be overridden, usually with the same value/expression (if different, the first would be actually applied).
Currently math generation will generate only one constant, named EGF_init_uM (whatever the units may be are suffixed). We are applying the pre-existing math override for EGF_init to the constant EGF_init_uM.
This is wrong for two reasons:
- first, whatever the override previously was on EGF_init was being ignored anyway; we should in fact apply the pre-existing math override for EGF to the constant EGF_init_uM in order to get the same results
- second, the previous override on EGF would now be "oprhaned" and an error generated which would block any save/run/export until the user fixes it; however, the user does not have any way of understanding this issue and apply the correct solution manually
|
1.0
|
Orphan overrides due to init/runtime naming issues - This issue affects clamped variables. Let's assume species EGF is clamped with some number or expression.
Previously math generation would generate two constants, named EGF and EGF_init, both of which would be listed in the math and in the table for overrides. But only one of them (the first one) would actually be used in the equations.
If the user would override EGF_init, nothing would actually happen (which in itself is a serious issue). But in most cases both of them would be overridden, usually with the same value/expression (if different, the first would be actually applied).
Currently math generation will generate only one constant, named EGF_init_uM (whatever the units may be are suffixed). We are applying the pre-existing math override for EGF_init to the constant EGF_init_uM.
This is wrong for two reasons:
- first, whatever the override previously was on EGF_init was being ignored anyway; we should in fact apply the pre-existing math override for EGF to the constant EGF_init_uM in order to get the same results
- second, the previous override on EGF would now be "oprhaned" and an error generated which would block any save/run/export until the user fixes it; however, the user does not have any way of understanding this issue and apply the correct solution manually
|
priority
|
orphan overrides due to init runtime naming issues this issue affects clamped variables let s assume species egf is clamped with some number or expression previously math generation would generate two constants named egf and egf init both of which would be listed in the math and in the table for overrides but only one of them the first one would actually be used in the equations if the user would override egf init nothing would actually happen which in itself is a serious issue but in most cases both of them would be overridden usually with the same value expression if different the first would be actually applied currently math generation will generate only one constant named egf init um whatever the units may be are suffixed we are applying the pre existing math override for egf init to the constant egf init um this is wrong for two reasons first whatever the override previously was on egf init was being ignored anyway we should in fact apply the pre existing math override for egf to the constant egf init um in order to get the same results second the previous override on egf would now be oprhaned and an error generated which would block any save run export until the user fixes it however the user does not have any way of understanding this issue and apply the correct solution manually
| 1
|
461,670
| 13,234,182,442
|
IssuesEvent
|
2020-08-18 15:53:18
|
sugarlabs-appstore/aslo-v4
|
https://api.github.com/repos/sugarlabs-appstore/aslo-v4
|
opened
|
Rewrite static resources
|
high-priority
|
The static (HTML, CSS, JS) libraries, following #37 , #35 , #28, #34, and other performance issues, it is important to rewrite the Appstore again in a more performance optimized Javascript framework.
Possibly, by using the powerful, but new CSS library Bulma, replacing the Bootstrap 4 library.
|
1.0
|
Rewrite static resources - The static (HTML, CSS, JS) libraries, following #37 , #35 , #28, #34, and other performance issues, it is important to rewrite the Appstore again in a more performance optimized Javascript framework.
Possibly, by using the powerful, but new CSS library Bulma, replacing the Bootstrap 4 library.
|
priority
|
rewrite static resources the static html css js libraries following and other performance issues it is important to rewrite the appstore again in a more performance optimized javascript framework possibly by using the powerful but new css library bulma replacing the bootstrap library
| 1
|
138,090
| 5,327,864,311
|
IssuesEvent
|
2017-02-15 10:22:51
|
odalic/sti
|
https://api.github.com/repos/odalic/sti
|
closed
|
Cannot process one of the testing files - it was working before!
|
bug priority: High
|
Please set up knowledge base as defined at:
https://grips.semantic-web.at/display/ADEQ/Odalic+-+Connection+to+the+Knowledge+Base
Use file:
[STMK_2012_LIVEBIRTHS-10lines.csv.zip](https://github.com/odalic/sti/files/776542/STMK_2012_LIVEBIRTHS-10lines.csv.zip)
Cannot further process that file: but it was working before (approx 2 weeks back)
The problematic query
```
SELECT DISTINCT ?subject
WHERE
{ ?subject <http://www.w3.org/2004/02/skos/core#prefLabel> "Gro� Sankt Florian"@de ;
a <http://www.w3.org/2004/02/skos/core#Concept>
}
```
I guess that there may be an issue with that special char?
The log:
```
java.util.concurrent.ExecutionException: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at java.util.concurrent.FutureTask.report(FutureTask.java:122) at java.util.concurrent.FutureTask.get(FutureTask.java:192) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.getResultForTaskId(DbCachedFutureBasedExecutionService.java:227) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:63) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:78) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.glassfish.jersey.server.model.internal.ResourceMethodInvocationHandlerFactory$1.invoke(ResourceMethodInvocationHandlerFactory.java:81) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher$1.run(AbstractJavaResourceMethodDispatcher.java:144) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.invoke(AbstractJavaResourceMethodDispatcher.java:161) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$ResponseOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:160) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:99) at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:389) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:347) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:102) at org.glassfish.jersey.server.ServerRuntime$2.run(ServerRuntime.java:326) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at org.glassfish.jersey.internal.Errors.process(Errors.java:315) at org.glassfish.jersey.internal.Errors.process(Errors.java:297) at org.glassfish.jersey.internal.Errors.process(Errors.java:267) at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317) at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:305) at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:1154) at org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:473) at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:427) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:388) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:341) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:228) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:292) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:240) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212) at org.apache.catalina.core.StandardContextValve.__invoke(StandardContextValve.java:106) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:528) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1099) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:670) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1520) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1476) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:745) Caused by: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:370) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:313) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:227) at javax.xml.stream.XMLInputFactory.newInstance(XMLInputFactory.java:154) at org.apache.jena.sparql.resultset.XMLInputStAX.<init>(XMLInputStAX.java:101) at org.apache.jena.sparql.resultset.XMLInput.make(XMLInput.java:74) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:38) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:34) at org.apache.jena.query.ResultSetFactory.fromXML(ResultSetFactory.java:308) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execResultSetInner(QueryEngineHTTP.java:365) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execSelect(QueryEngineHTTP.java:334) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:228) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:219) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryEntityCandidates(SPARQLProxy.java:631) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.findEntityCandidatesInternal(SPARQLProxy.java:364) at uk.ac.shef.dcs.kbproxy.KBProxy.lambda$findEntityCandidates$0(KBProxy.java:144) at uk.ac.shef.dcs.kbproxy.KBProxy.Do(KBProxy.java:387) at uk.ac.shef.dcs.kbproxy.KBProxy.findEntityCandidates(KBProxy.java:144) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNINGPreliminaryColumnClassifier.runPreliminaryColumnClassifier(LEARNINGPreliminaryColumnClassifier.java:120) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNING.learn(LEARNING.java:37) at uk.ac.shef.dcs.sti.core.algorithm.tmp.TMPOdalicInterpreter.start(TMPOdalicInterpreter.java:114) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.lambda$submitForTaskId$29(DbCachedFutureBasedExecutionService.java:160) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ... 1 more Caused by: java.lang.RuntimeException: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:367) ... 25 more Caused by: java.util.ServiceConfigurationError: javax.xml.stream.XMLInputFactory: Error reading configuration file at java.util.ServiceLoader.fail(ServiceLoader.java:232) at java.util.ServiceLoader.parse(ServiceLoader.java:309) at java.util.ServiceLoader.access$200(ServiceLoader.java:185) at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) at javax.xml.stream.FactoryFinder$1.run(FactoryFinder.java:352) at java.security.AccessController.doPrivileged(Native Method) at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:341) ... 25 more Caused by: java.io.FileNotFoundException: /usr/local/apache-tomcat-8.0.36/webapps/odalic/WEB-INF/lib/woodstox-core-asl-4.4.1.jar (No such file or directory) at java.util.zip.ZipFile.open(Native Method) at java.util.zip.ZipFile.<init>(ZipFile.java:219) at java.util.zip.ZipFile.<init>(ZipFile.java:149) at java.util.jar.JarFile.<init>(JarFile.java:166) at java.util.jar.JarFile.<init>(JarFile.java:103) at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) at java.net.URL.openStream(URL.java:1038) at java.util.ServiceLoader.parse(ServiceLoader.java:304) ... 32 more
```
This also means that #166 does not work properly??
|
1.0
|
Cannot process one of the testing files - it was working before! - Please set up knowledge base as defined at:
https://grips.semantic-web.at/display/ADEQ/Odalic+-+Connection+to+the+Knowledge+Base
Use file:
[STMK_2012_LIVEBIRTHS-10lines.csv.zip](https://github.com/odalic/sti/files/776542/STMK_2012_LIVEBIRTHS-10lines.csv.zip)
Cannot further process that file: but it was working before (approx 2 weeks back)
The problematic query
```
SELECT DISTINCT ?subject
WHERE
{ ?subject <http://www.w3.org/2004/02/skos/core#prefLabel> "Gro� Sankt Florian"@de ;
a <http://www.w3.org/2004/02/skos/core#Concept>
}
```
I guess that there may be an issue with that special char?
The log:
```
java.util.concurrent.ExecutionException: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at java.util.concurrent.FutureTask.report(FutureTask.java:122) at java.util.concurrent.FutureTask.get(FutureTask.java:192) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.getResultForTaskId(DbCachedFutureBasedExecutionService.java:227) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:63) at cz.cuni.mff.xrg.odalic.api.rest.resources.ResultResource.getResult(ResultResource.java:78) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:497) at org.glassfish.jersey.server.model.internal.ResourceMethodInvocationHandlerFactory$1.invoke(ResourceMethodInvocationHandlerFactory.java:81) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher$1.run(AbstractJavaResourceMethodDispatcher.java:144) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.invoke(AbstractJavaResourceMethodDispatcher.java:161) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$ResponseOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:160) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:99) at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:389) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:347) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:102) at org.glassfish.jersey.server.ServerRuntime$2.run(ServerRuntime.java:326) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at org.glassfish.jersey.internal.Errors.process(Errors.java:315) at org.glassfish.jersey.internal.Errors.process(Errors.java:297) at org.glassfish.jersey.internal.Errors.process(Errors.java:267) at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317) at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:305) at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:1154) at org.glassfish.jersey.servlet.WebComponent.serviceImpl(WebComponent.java:473) at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:427) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:388) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:341) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:228) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:292) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:240) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:207) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:212) at org.apache.catalina.core.StandardContextValve.__invoke(StandardContextValve.java:106) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:502) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:141) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:79) at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:616) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:88) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:528) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1099) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:670) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1520) at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1476) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:745) Caused by: javax.xml.stream.FactoryConfigurationError: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:370) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:313) at javax.xml.stream.FactoryFinder.find(FactoryFinder.java:227) at javax.xml.stream.XMLInputFactory.newInstance(XMLInputFactory.java:154) at org.apache.jena.sparql.resultset.XMLInputStAX.<init>(XMLInputStAX.java:101) at org.apache.jena.sparql.resultset.XMLInput.make(XMLInput.java:74) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:38) at org.apache.jena.sparql.resultset.XMLInput.fromXML(XMLInput.java:34) at org.apache.jena.query.ResultSetFactory.fromXML(ResultSetFactory.java:308) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execResultSetInner(QueryEngineHTTP.java:365) at org.apache.jena.sparql.engine.http.QueryEngineHTTP.execSelect(QueryEngineHTTP.java:334) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:228) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryReturnSingleValues(SPARQLProxy.java:219) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.queryEntityCandidates(SPARQLProxy.java:631) at uk.ac.shef.dcs.kbproxy.sparql.SPARQLProxy.findEntityCandidatesInternal(SPARQLProxy.java:364) at uk.ac.shef.dcs.kbproxy.KBProxy.lambda$findEntityCandidates$0(KBProxy.java:144) at uk.ac.shef.dcs.kbproxy.KBProxy.Do(KBProxy.java:387) at uk.ac.shef.dcs.kbproxy.KBProxy.findEntityCandidates(KBProxy.java:144) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNINGPreliminaryColumnClassifier.runPreliminaryColumnClassifier(LEARNINGPreliminaryColumnClassifier.java:120) at uk.ac.shef.dcs.sti.core.algorithm.tmp.LEARNING.learn(LEARNING.java:37) at uk.ac.shef.dcs.sti.core.algorithm.tmp.TMPOdalicInterpreter.start(TMPOdalicInterpreter.java:114) at cz.cuni.mff.xrg.odalic.tasks.executions.DbCachedFutureBasedExecutionService.lambda$submitForTaskId$29(DbCachedFutureBasedExecutionService.java:160) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) ... 1 more Caused by: java.lang.RuntimeException: Provider for class javax.xml.stream.XMLInputFactory cannot be created at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:367) ... 25 more Caused by: java.util.ServiceConfigurationError: javax.xml.stream.XMLInputFactory: Error reading configuration file at java.util.ServiceLoader.fail(ServiceLoader.java:232) at java.util.ServiceLoader.parse(ServiceLoader.java:309) at java.util.ServiceLoader.access$200(ServiceLoader.java:185) at java.util.ServiceLoader$LazyIterator.hasNextService(ServiceLoader.java:357) at java.util.ServiceLoader$LazyIterator.hasNext(ServiceLoader.java:393) at java.util.ServiceLoader$1.hasNext(ServiceLoader.java:474) at javax.xml.stream.FactoryFinder$1.run(FactoryFinder.java:352) at java.security.AccessController.doPrivileged(Native Method) at javax.xml.stream.FactoryFinder.findServiceProvider(FactoryFinder.java:341) ... 25 more Caused by: java.io.FileNotFoundException: /usr/local/apache-tomcat-8.0.36/webapps/odalic/WEB-INF/lib/woodstox-core-asl-4.4.1.jar (No such file or directory) at java.util.zip.ZipFile.open(Native Method) at java.util.zip.ZipFile.<init>(ZipFile.java:219) at java.util.zip.ZipFile.<init>(ZipFile.java:149) at java.util.jar.JarFile.<init>(JarFile.java:166) at java.util.jar.JarFile.<init>(JarFile.java:103) at sun.net.www.protocol.jar.URLJarFile.<init>(URLJarFile.java:93) at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) at sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) at sun.net.www.protocol.jar.JarURLConnection.getInputStream(JarURLConnection.java:150) at java.net.URL.openStream(URL.java:1038) at java.util.ServiceLoader.parse(ServiceLoader.java:304) ... 32 more
```
This also means that #166 does not work properly??
|
priority
|
cannot process one of the testing files it was working before please set up knowledge base as defined at use file cannot further process that file but it was working before approx weeks back the problematic query select distinct subject where subject gro� sankt florian de a i guess that there may be an issue with that special char the log java util concurrent executionexception javax xml stream factoryconfigurationerror provider for class javax xml stream xmlinputfactory cannot be created at java util concurrent futuretask report futuretask java at java util concurrent futuretask get futuretask java at cz cuni mff xrg odalic tasks executions dbcachedfuturebasedexecutionservice getresultfortaskid dbcachedfuturebasedexecutionservice java at cz cuni mff xrg odalic api rest resources resultresource getresult resultresource java at cz cuni mff xrg odalic api rest resources resultresource getresult resultresource java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org glassfish jersey server model internal resourcemethodinvocationhandlerfactory invoke resourcemethodinvocationhandlerfactory java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher run abstractjavaresourcemethoddispatcher java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher invoke abstractjavaresourcemethoddispatcher java at org glassfish jersey server model internal javaresourcemethoddispatcherprovider responseoutinvoker dodispatch javaresourcemethoddispatcherprovider java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher dispatch abstractjavaresourcemethoddispatcher java at org glassfish jersey server model resourcemethodinvoker invoke resourcemethodinvoker java at org glassfish jersey server model resourcemethodinvoker apply resourcemethodinvoker java at org glassfish jersey server model resourcemethodinvoker apply resourcemethodinvoker java at org glassfish jersey server serverruntime run serverruntime java at org glassfish jersey internal errors call errors java at org glassfish jersey internal errors call errors java at org glassfish jersey internal errors process errors java at org glassfish jersey internal errors process errors java at org glassfish jersey internal errors process errors java at org glassfish jersey process internal requestscope runinscope requestscope java at org glassfish jersey server serverruntime process serverruntime java at org glassfish jersey server applicationhandler handle applicationhandler java at org glassfish jersey servlet webcomponent serviceimpl webcomponent java at org glassfish jersey servlet webcomponent service webcomponent java at org glassfish jersey servlet servletcontainer service servletcontainer java at org glassfish jersey servlet servletcontainer service servletcontainer java at org glassfish jersey servlet servletcontainer service servletcontainer java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote process java at org apache coyote abstractprotocol abstractconnectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net nioendpoint socketprocessor run nioendpoint java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java caused by javax xml stream factoryconfigurationerror provider for class javax xml stream xmlinputfactory cannot be created at javax xml stream factoryfinder findserviceprovider factoryfinder java at javax xml stream factoryfinder find factoryfinder java at javax xml stream factoryfinder find factoryfinder java at javax xml stream xmlinputfactory newinstance xmlinputfactory java at org apache jena sparql resultset xmlinputstax xmlinputstax java at org apache jena sparql resultset xmlinput make xmlinput java at org apache jena sparql resultset xmlinput fromxml xmlinput java at org apache jena sparql resultset xmlinput fromxml xmlinput java at org apache jena query resultsetfactory fromxml resultsetfactory java at org apache jena sparql engine http queryenginehttp execresultsetinner queryenginehttp java at org apache jena sparql engine http queryenginehttp execselect queryenginehttp java at uk ac shef dcs kbproxy sparql sparqlproxy queryreturnsinglevalues sparqlproxy java at uk ac shef dcs kbproxy sparql sparqlproxy queryreturnsinglevalues sparqlproxy java at uk ac shef dcs kbproxy sparql sparqlproxy queryentitycandidates sparqlproxy java at uk ac shef dcs kbproxy sparql sparqlproxy findentitycandidatesinternal sparqlproxy java at uk ac shef dcs kbproxy kbproxy lambda findentitycandidates kbproxy java at uk ac shef dcs kbproxy kbproxy do kbproxy java at uk ac shef dcs kbproxy kbproxy findentitycandidates kbproxy java at uk ac shef dcs sti core algorithm tmp learningpreliminarycolumnclassifier runpreliminarycolumnclassifier learningpreliminarycolumnclassifier java at uk ac shef dcs sti core algorithm tmp learning learn learning java at uk ac shef dcs sti core algorithm tmp tmpodalicinterpreter start tmpodalicinterpreter java at cz cuni mff xrg odalic tasks executions dbcachedfuturebasedexecutionservice lambda submitfortaskid dbcachedfuturebasedexecutionservice java at java util concurrent futuretask run futuretask java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java more caused by java lang runtimeexception provider for class javax xml stream xmlinputfactory cannot be created at javax xml stream factoryfinder findserviceprovider factoryfinder java more caused by java util serviceconfigurationerror javax xml stream xmlinputfactory error reading configuration file at java util serviceloader fail serviceloader java at java util serviceloader parse serviceloader java at java util serviceloader access serviceloader java at java util serviceloader lazyiterator hasnextservice serviceloader java at java util serviceloader lazyiterator hasnext serviceloader java at java util serviceloader hasnext serviceloader java at javax xml stream factoryfinder run factoryfinder java at java security accesscontroller doprivileged native method at javax xml stream factoryfinder findserviceprovider factoryfinder java more caused by java io filenotfoundexception usr local apache tomcat webapps odalic web inf lib woodstox core asl jar no such file or directory at java util zip zipfile open native method at java util zip zipfile zipfile java at java util zip zipfile zipfile java at java util jar jarfile jarfile java at java util jar jarfile jarfile java at sun net at sun net at sun net at sun net at sun net at java net url openstream url java at java util serviceloader parse serviceloader java more this also means that does not work properly
| 1
|
413,283
| 12,064,394,733
|
IssuesEvent
|
2020-04-16 08:12:46
|
openmsupply/mobile
|
https://api.github.com/repos/openmsupply/mobile
|
closed
|
Read-only constraints not being consistently applied
|
Bug: development Docs: not needed Effort: small Feature Ivory Coast (phase 2) Module: dispensary Priority: high
|
## Is your feature request related to a problem? Please describe.
A few problems/inconsistencies with dispensing for non-local patients:
- the `supplyingStoreId` field is not being correctly set for patients created via the lookup API.
- the `storeId` field is not being correctly set for patients created via the lookup API.
- non-local patients/prescribers cannot be edited, but the edit form can still be accessed and the icon is not disabled as it is for policies.
## Describe the solution you'd like
- correctly set supplying store fields for patients/prescribers.
- disable edit icon in the dispensing window for non-local patients/prescribers
## Implementation
- update `PatientActions.patientUpdate`.
- update `PrescriberActions.updatePrescriber`
- update `PrescriptionInfo` component.
## Describe alternatives you've considered
N/A.
## Additional context
See epic #2446.
|
1.0
|
Read-only constraints not being consistently applied - ## Is your feature request related to a problem? Please describe.
A few problems/inconsistencies with dispensing for non-local patients:
- the `supplyingStoreId` field is not being correctly set for patients created via the lookup API.
- the `storeId` field is not being correctly set for patients created via the lookup API.
- non-local patients/prescribers cannot be edited, but the edit form can still be accessed and the icon is not disabled as it is for policies.
## Describe the solution you'd like
- correctly set supplying store fields for patients/prescribers.
- disable edit icon in the dispensing window for non-local patients/prescribers
## Implementation
- update `PatientActions.patientUpdate`.
- update `PrescriberActions.updatePrescriber`
- update `PrescriptionInfo` component.
## Describe alternatives you've considered
N/A.
## Additional context
See epic #2446.
|
priority
|
read only constraints not being consistently applied is your feature request related to a problem please describe a few problems inconsistencies with dispensing for non local patients the supplyingstoreid field is not being correctly set for patients created via the lookup api the storeid field is not being correctly set for patients created via the lookup api non local patients prescribers cannot be edited but the edit form can still be accessed and the icon is not disabled as it is for policies describe the solution you d like correctly set supplying store fields for patients prescribers disable edit icon in the dispensing window for non local patients prescribers implementation update patientactions patientupdate update prescriberactions updateprescriber update prescriptioninfo component describe alternatives you ve considered n a additional context see epic
| 1
|
745,005
| 25,965,761,941
|
IssuesEvent
|
2022-12-19 06:41:58
|
paperclip-ui/paperclip
|
https://api.github.com/repos/paperclip-ui/paperclip
|
closed
|
Consider dropping brackets
|
priority: low impact: low effort: high
|
E.g:
```javascript
import "./common.pc" as common
import "./theme.pc" as theme
import "./input.pc" as input
component ColorPickerPopup
render div
style extends common.panel
position: fixed
flex-shrink:0
transform: translate(calc(-100% - var(theme.space07)))
z-index: 1024
div header
style
padding: var(theme.space04)
display: flex
justify-content: space-between
font-weight: 600
padding-bottom: var(theme.space04)
div
text "Color"
div
text "×"
div
style
padding: var(theme.space04)
display: flex
flex-direction: column
gap: var(theme.space04)
color-picker(width: "220", height: "130")
div
text "Hue"
div
text "Opacity"
div
text "value"
div
style
border-top: 1px solid var(theme.borderColor)
padding: var(theme.space04)
display: flex
flex-direction: column
gap: var(theme.space03)
div
text "color var"
div
text "color var"
div
text "color var"
```
|
1.0
|
Consider dropping brackets - E.g:
```javascript
import "./common.pc" as common
import "./theme.pc" as theme
import "./input.pc" as input
component ColorPickerPopup
render div
style extends common.panel
position: fixed
flex-shrink:0
transform: translate(calc(-100% - var(theme.space07)))
z-index: 1024
div header
style
padding: var(theme.space04)
display: flex
justify-content: space-between
font-weight: 600
padding-bottom: var(theme.space04)
div
text "Color"
div
text "×"
div
style
padding: var(theme.space04)
display: flex
flex-direction: column
gap: var(theme.space04)
color-picker(width: "220", height: "130")
div
text "Hue"
div
text "Opacity"
div
text "value"
div
style
border-top: 1px solid var(theme.borderColor)
padding: var(theme.space04)
display: flex
flex-direction: column
gap: var(theme.space03)
div
text "color var"
div
text "color var"
div
text "color var"
```
|
priority
|
consider dropping brackets e g javascript import common pc as common import theme pc as theme import input pc as input component colorpickerpopup render div style extends common panel position fixed flex shrink transform translate calc var theme z index div header style padding var theme display flex justify content space between font weight padding bottom var theme div text color div text times div style padding var theme display flex flex direction column gap var theme color picker width height div text hue div text opacity div text value div style border top solid var theme bordercolor padding var theme display flex flex direction column gap var theme div text color var div text color var div text color var
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.