Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
8,362
7,359,749,020
IssuesEvent
2018-03-10 10:50:33
MIPT-ILab/mipt-mips
https://api.github.com/repos/MIPT-ILab/mipt-mips
closed
Generate and test MS Visual Studio solution & project files with CMake
infrastructure mentor
**AMB**: make MIPT-MIPS fully buildable with MSVS. Add a wiki manual if there are some tricks. I've added MSVC command line build to AppVeyor with two noted 1. I have to use patched LibELF build as original source contains a small bug 2. MIPS binutils are built only by MinGW — it is OK as we use only binaries Besides, everything is OK to create *.sln and *.vcxproj files (NMake is very, very bad)
1.0
Generate and test MS Visual Studio solution & project files with CMake - **AMB**: make MIPT-MIPS fully buildable with MSVS. Add a wiki manual if there are some tricks. I've added MSVC command line build to AppVeyor with two noted 1. I have to use patched LibELF build as original source contains a small bug 2. MIPS binutils are built only by MinGW — it is OK as we use only binaries Besides, everything is OK to create *.sln and *.vcxproj files (NMake is very, very bad)
infrastructure
generate and test ms visual studio solution project files with cmake amb make mipt mips fully buildable with msvs add a wiki manual if there are some tricks i ve added msvc command line build to appveyor with two noted i have to use patched libelf build as original source contains a small bug mips binutils are built only by mingw — it is ok as we use only binaries besides everything is ok to create sln and vcxproj files nmake is very very bad
1
11,725
9,395,978,984
IssuesEvent
2019-04-08 05:29:01
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Add a sample-based standard deviation to MathUtilities
interface/infrastructure newfeature
MathUtilities currently contains MathUtilities.StandardDeviation which appears to me a population standard deviation (the divisor is the total count, N). Could there also be a MathUtilities.StandardDeviationSample (or something) where the divisor is (N-1) which is the standard deviation of a sample?
1.0
Add a sample-based standard deviation to MathUtilities - MathUtilities currently contains MathUtilities.StandardDeviation which appears to me a population standard deviation (the divisor is the total count, N). Could there also be a MathUtilities.StandardDeviationSample (or something) where the divisor is (N-1) which is the standard deviation of a sample?
infrastructure
add a sample based standard deviation to mathutilities mathutilities currently contains mathutilities standarddeviation which appears to me a population standard deviation the divisor is the total count n could there also be a mathutilities standarddeviationsample or something where the divisor is n which is the standard deviation of a sample
1
52,821
7,786,287,168
IssuesEvent
2018-06-06 18:27:51
PeterCamilleri/fOOrth
https://api.github.com/repos/PeterCamilleri/fOOrth
opened
Describe why the {{ word is an empty macro
documentation
Trying to understand how this works is not easy, even for me the author. Explain how all word ending in {{ are assumed to contain an embedded procedure literal which is pushed onto the stack. Normally the method then proceeds to do something with that procedure. The simple {{ ... }} is a special case in that it does nothing, leaving the procedure on the stack. Maybe also include a pointer to where procedure literals are handled.
1.0
Describe why the {{ word is an empty macro - Trying to understand how this works is not easy, even for me the author. Explain how all word ending in {{ are assumed to contain an embedded procedure literal which is pushed onto the stack. Normally the method then proceeds to do something with that procedure. The simple {{ ... }} is a special case in that it does nothing, leaving the procedure on the stack. Maybe also include a pointer to where procedure literals are handled.
non_infrastructure
describe why the word is an empty macro trying to understand how this works is not easy even for me the author explain how all word ending in are assumed to contain an embedded procedure literal which is pushed onto the stack normally the method then proceeds to do something with that procedure the simple is a special case in that it does nothing leaving the procedure on the stack maybe also include a pointer to where procedure literals are handled
0
32,054
26,390,881,882
IssuesEvent
2023-01-12 15:35:55
gap-system/gap
https://api.github.com/repos/gap-system/gap
closed
CI for creating Windows .exe for release tags is broken
os: windows topic: infrastructure topic: ci
I tagged v4.12.0 and our CI workflow for creating .exe files failed [(log)](https://github.com/gap-system/gap/runs/7907771045?check_suite_focus=true) with this error: ``` tar: gap-4.12.0/pkg/ace/htm/CHAP007.htm: Cannot create symlink to 'CHAP00A.htm': No such file or directory tar: gap-4.12.0/pkg/ace/htm/CHAP008.htm: Cannot create symlink to 'CHAP00B.htm': No such file or directory tar: gap-4.12.0/pkg/ace/htm/CHAP009.htm: Cannot create symlink to 'CHAP00C.htm': No such file or directory tar: Exiting with failure status due to previous errors Error: Process completed with exit code 2. ``` Indeed, the latest ACE release has some symlinks here, to resolve issue #4430. When I made that change, I did not have in mind that Windows doesn't have true symlinks and might not like it. We need to fix the immediate problem; but I also think there is a clear case that we have a long-term issue here, too: while we test the "release CI job" daily, the test was insufficient to catch this issue; because it occurred in one of the parts of the CI job that is really only executed for releases. Bad! We need to resolve both. *Short term*: I will cook up a workaround so the 4.12.0 release can be completed. I may just be evil and modify that gap-4.12.0 tarball by replacing (?) those symlinks *Mid term*: A proper fix for the above issue needs to be found. Perhaps ACE should just not create a symlink there, but instead just copy the HTML file? Then again... why does it only fail when it extracts `https://github.com/gap-system/gap/releases/download/v4.12.0/gap-4.12.0.tar.gz` but does not fail when it extracts https://github.com/gap-system/PackageDistro/releases/download/latest/packages.tar.gz ?? My only theory is that the order in which the files occurs in the tarballs is different; in the "good" one the symlinks come after the real files, while in the "bad" one the symlinks come before, resulting in the observed error. However, this theory is backed by nothing other than my free interpretation of the error message, so it might be utter nonsense. *Long term*: We need to enhance the CI test to find this earlier! To this end, I'd like to minimize the difference between when it runs on a release tag vs. the "usual" runs. One way would be to modify the offending step `Download the appropriate GAP release tarball` as follows: instead of downloading the release tarball, we download an artifact created by the previous job in the workflow. However, to conserve storage, we only upload those artifacts for scheduled runs; well, let's upload them also for release tags; and then modify the `Download the appropriate GAP release tarball`` to use these artifacts, but also run it on the scheduled runs... Of course I am also open for other ideas! CC @wilfwilson @ChrisJefferson @FriedrichRober
1.0
CI for creating Windows .exe for release tags is broken - I tagged v4.12.0 and our CI workflow for creating .exe files failed [(log)](https://github.com/gap-system/gap/runs/7907771045?check_suite_focus=true) with this error: ``` tar: gap-4.12.0/pkg/ace/htm/CHAP007.htm: Cannot create symlink to 'CHAP00A.htm': No such file or directory tar: gap-4.12.0/pkg/ace/htm/CHAP008.htm: Cannot create symlink to 'CHAP00B.htm': No such file or directory tar: gap-4.12.0/pkg/ace/htm/CHAP009.htm: Cannot create symlink to 'CHAP00C.htm': No such file or directory tar: Exiting with failure status due to previous errors Error: Process completed with exit code 2. ``` Indeed, the latest ACE release has some symlinks here, to resolve issue #4430. When I made that change, I did not have in mind that Windows doesn't have true symlinks and might not like it. We need to fix the immediate problem; but I also think there is a clear case that we have a long-term issue here, too: while we test the "release CI job" daily, the test was insufficient to catch this issue; because it occurred in one of the parts of the CI job that is really only executed for releases. Bad! We need to resolve both. *Short term*: I will cook up a workaround so the 4.12.0 release can be completed. I may just be evil and modify that gap-4.12.0 tarball by replacing (?) those symlinks *Mid term*: A proper fix for the above issue needs to be found. Perhaps ACE should just not create a symlink there, but instead just copy the HTML file? Then again... why does it only fail when it extracts `https://github.com/gap-system/gap/releases/download/v4.12.0/gap-4.12.0.tar.gz` but does not fail when it extracts https://github.com/gap-system/PackageDistro/releases/download/latest/packages.tar.gz ?? My only theory is that the order in which the files occurs in the tarballs is different; in the "good" one the symlinks come after the real files, while in the "bad" one the symlinks come before, resulting in the observed error. However, this theory is backed by nothing other than my free interpretation of the error message, so it might be utter nonsense. *Long term*: We need to enhance the CI test to find this earlier! To this end, I'd like to minimize the difference between when it runs on a release tag vs. the "usual" runs. One way would be to modify the offending step `Download the appropriate GAP release tarball` as follows: instead of downloading the release tarball, we download an artifact created by the previous job in the workflow. However, to conserve storage, we only upload those artifacts for scheduled runs; well, let's upload them also for release tags; and then modify the `Download the appropriate GAP release tarball`` to use these artifacts, but also run it on the scheduled runs... Of course I am also open for other ideas! CC @wilfwilson @ChrisJefferson @FriedrichRober
infrastructure
ci for creating windows exe for release tags is broken i tagged and our ci workflow for creating exe files failed with this error tar gap pkg ace htm htm cannot create symlink to htm no such file or directory tar gap pkg ace htm htm cannot create symlink to htm no such file or directory tar gap pkg ace htm htm cannot create symlink to htm no such file or directory tar exiting with failure status due to previous errors error process completed with exit code indeed the latest ace release has some symlinks here to resolve issue when i made that change i did not have in mind that windows doesn t have true symlinks and might not like it we need to fix the immediate problem but i also think there is a clear case that we have a long term issue here too while we test the release ci job daily the test was insufficient to catch this issue because it occurred in one of the parts of the ci job that is really only executed for releases bad we need to resolve both short term i will cook up a workaround so the release can be completed i may just be evil and modify that gap tarball by replacing those symlinks mid term a proper fix for the above issue needs to be found perhaps ace should just not create a symlink there but instead just copy the html file then again why does it only fail when it extracts but does not fail when it extracts my only theory is that the order in which the files occurs in the tarballs is different in the good one the symlinks come after the real files while in the bad one the symlinks come before resulting in the observed error however this theory is backed by nothing other than my free interpretation of the error message so it might be utter nonsense long term we need to enhance the ci test to find this earlier to this end i d like to minimize the difference between when it runs on a release tag vs the usual runs one way would be to modify the offending step download the appropriate gap release tarball as follows instead of downloading the release tarball we download an artifact created by the previous job in the workflow however to conserve storage we only upload those artifacts for scheduled runs well let s upload them also for release tags and then modify the download the appropriate gap release tarball to use these artifacts but also run it on the scheduled runs of course i am also open for other ideas cc wilfwilson chrisjefferson friedrichrober
1
156,356
5,967,912,706
IssuesEvent
2017-05-30 16:57:39
minio/minio
https://api.github.com/repos/minio/minio
closed
Investigate if profile id needs to be supported in gcs .
priority: high
<!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. 2. 3. 4. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio version`): * Environment name and version (e.g. nginx 1.9.1): * Server type and version: * Operating System and version (`uname -a`): * Link to your project:
1.0
Investigate if profile id needs to be supported in gcs . - <!--- Provide a general summary of the issue in the Title above --> ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. 2. 3. 4. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used (`minio version`): * Environment name and version (e.g. nginx 1.9.1): * Server type and version: * Operating System and version (`uname -a`): * Link to your project:
non_infrastructure
investigate if profile id needs to be supported in gcs expected behavior current behavior possible solution steps to reproduce for bugs context your environment version used minio version environment name and version e g nginx server type and version operating system and version uname a link to your project
0
578,379
17,147,168,538
IssuesEvent
2021-07-13 15:46:04
googleapis/python-automl
https://api.github.com/repos/googleapis/python-automl
closed
samples.tables.dataset_test: test_dataset_update failed
api: automl flakybot: issue priority: p1 samples type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 788a1b3502266916a619b0284f62cefa1cb10ca2 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7b204985-ebdd-4a3d-ae14-821ee6364dd9), [Sponge](http://sponge2/7b204985-ebdd-4a3d-ae14-821ee6364dd9) status: failed <details><summary>Test output</summary><br><pre>args = (parent: "projects/1012616486416/locations/us-central1/datasets/TBL8127498693055086592/tableSpecs/5798740762006913024" ,) kwargs = {'metadata': [('x-goog-request-params', 'parent=projects/1012616486416/locations/us-central1/datasets/TBL8127498693055...006913024'), ('x-goog-api-client', 'automl-tables-wrapper/2.4.0 gl-python/3.7.10 grpc/1.38.1 gax/1.31.0 gapic/2.4.0')]} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7fede4371990> request = parent: "projects/1012616486416/locations/us-central1/datasets/TBL8127498693055086592/tableSpecs/5798740762006913024" timeout = None metadata = [('x-goog-request-params', 'parent=projects/1012616486416/locations/us-central1/datasets/TBL8127498693055086592/tableS...2006913024'), ('x-goog-api-client', 'automl-tables-wrapper/2.4.0 gl-python/3.7.10 grpc/1.38.1 gax/1.31.0 gapic/2.4.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7fede5cc1d50> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7fede5c9ecd0> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.UNAUTHENTICATED E details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." E debug_error_string = "{"created":"@1626170935.431998858","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" E > .nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: capsys = <_pytest.capture.CaptureFixture object at 0x7fede43055d0> def test_dataset_update(capsys): dataset = ensure_dataset_ready() automl_tables_dataset.update_dataset( PROJECT, REGION, dataset.display_name, target_column_spec_name="Deposit", > weight_column_spec_name="Balance", ) dataset_test.py:92: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ automl_tables_dataset.py:204: in update_dataset column_spec_display_name=weight_column_spec_name, ../../google/cloud/automl_v1beta1/services/tables/tables_client.py:1786: in set_weight_column region=region, ../../google/cloud/automl_v1beta1/services/tables/tables_client.py:425: in __column_spec_name_from_args region=region, ../../google/cloud/automl_v1beta1/services/tables/tables_client.py:1234: in list_column_specs return self.auto_ml_client.list_column_specs(request=request, **method_kwargs) ../../google/cloud/automl_v1beta1/services/auto_ml/client.py:1560: in list_column_specs response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) .nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py:290: in retry_wrapped_func on_error=on_error, .nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py:188: in retry_target return target() .nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid a...entication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" > > ??? E google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. <string>:3: Unauthenticated</pre></details>
1.0
samples.tables.dataset_test: test_dataset_update failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 788a1b3502266916a619b0284f62cefa1cb10ca2 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/7b204985-ebdd-4a3d-ae14-821ee6364dd9), [Sponge](http://sponge2/7b204985-ebdd-4a3d-ae14-821ee6364dd9) status: failed <details><summary>Test output</summary><br><pre>args = (parent: "projects/1012616486416/locations/us-central1/datasets/TBL8127498693055086592/tableSpecs/5798740762006913024" ,) kwargs = {'metadata': [('x-goog-request-params', 'parent=projects/1012616486416/locations/us-central1/datasets/TBL8127498693055...006913024'), ('x-goog-api-client', 'automl-tables-wrapper/2.4.0 gl-python/3.7.10 grpc/1.38.1 gax/1.31.0 gapic/2.4.0')]} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7fede4371990> request = parent: "projects/1012616486416/locations/us-central1/datasets/TBL8127498693055086592/tableSpecs/5798740762006913024" timeout = None metadata = [('x-goog-request-params', 'parent=projects/1012616486416/locations/us-central1/datasets/TBL8127498693055086592/tableS...2006913024'), ('x-goog-api-client', 'automl-tables-wrapper/2.4.0 gl-python/3.7.10 grpc/1.38.1 gax/1.31.0 gapic/2.4.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7fede5cc1d50> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7fede5c9ecd0> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.UNAUTHENTICATED E details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." E debug_error_string = "{"created":"@1626170935.431998858","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" E > .nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: capsys = <_pytest.capture.CaptureFixture object at 0x7fede43055d0> def test_dataset_update(capsys): dataset = ensure_dataset_ready() automl_tables_dataset.update_dataset( PROJECT, REGION, dataset.display_name, target_column_spec_name="Deposit", > weight_column_spec_name="Balance", ) dataset_test.py:92: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ automl_tables_dataset.py:204: in update_dataset column_spec_display_name=weight_column_spec_name, ../../google/cloud/automl_v1beta1/services/tables/tables_client.py:1786: in set_weight_column region=region, ../../google/cloud/automl_v1beta1/services/tables/tables_client.py:425: in __column_spec_name_from_args region=region, ../../google/cloud/automl_v1beta1/services/tables/tables_client.py:1234: in list_column_specs return self.auto_ml_client.list_column_specs(request=request, **method_kwargs) ../../google/cloud/automl_v1beta1/services/auto_ml/client.py:1560: in list_column_specs response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) .nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py:290: in retry_wrapped_func on_error=on_error, .nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py:188: in retry_target return target() .nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid a...entication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" > > ??? E google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. <string>:3: Unauthenticated</pre></details>
non_infrastructure
samples tables dataset test test dataset update failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output args parent projects locations us datasets tablespecs kwargs metadata six wraps callable def error remapped callable args kwargs try return callable args kwargs nox py lib site packages google api core grpc helpers py self request parent projects locations us datasets tablespecs timeout none metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox py lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode unauthenticated e details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see e debug error string created description error received from peer file src core lib surface call cc file line grpc message request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see e nox py lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception capsys def test dataset update capsys dataset ensure dataset ready automl tables dataset update dataset project region dataset display name target column spec name deposit weight column spec name balance dataset test py automl tables dataset py in update dataset column spec display name weight column spec name google cloud automl services tables tables client py in set weight column region region google cloud automl services tables tables client py in column spec name from args region region google cloud automl services tables tables client py in list column specs return self auto ml client list column specs request request method kwargs google cloud automl services auto ml client py in list column specs response rpc request retry retry timeout timeout metadata metadata nox py lib site packages google api core gapic method py in call return wrapped func args kwargs nox py lib site packages google api core retry py in retry wrapped func on error on error nox py lib site packages google api core retry py in retry target return target nox py lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode unauthenticated details request had invalid a entication credential see e google api core exceptions unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see unauthenticated
0
34,399
29,801,221,549
IssuesEvent
2023-06-16 08:15:28
Jannyboy11/InvSee-plus-plus
https://api.github.com/repos/Jannyboy11/InvSee-plus-plus
closed
CI builds are broken
wontfix bug in other software infrastructure
The modules for CraftBukkit-1.20 currently don't build in CI because the CraftBukkit-1.20 server dependency could not be found: https://github.com/Jannyboy11/InvSee-plus-plus/actions/runs/5249196414. This is because running BuildTools with `--rev 1.20` actually builds CraftBukkit 1.20.1: https://hub.spigotmc.org/jira/browse/BUILDTOOLS-658. md_5 closed this as a wontfix, so until support for CraftBukkit 1.20 will be dropped, this remains an issue. One way to remedy this would be to build CraftBukkit without BuildTools: just clone the repo, revert to the latest 1.20 commit, apply the nms patches and compile and install. Another option is to just wait and hope that most of my users will migrate to 1.20.1 soon so that I don't have to maintain 1.20 support anymore. Addressing this as wontfix for now.
1.0
CI builds are broken - The modules for CraftBukkit-1.20 currently don't build in CI because the CraftBukkit-1.20 server dependency could not be found: https://github.com/Jannyboy11/InvSee-plus-plus/actions/runs/5249196414. This is because running BuildTools with `--rev 1.20` actually builds CraftBukkit 1.20.1: https://hub.spigotmc.org/jira/browse/BUILDTOOLS-658. md_5 closed this as a wontfix, so until support for CraftBukkit 1.20 will be dropped, this remains an issue. One way to remedy this would be to build CraftBukkit without BuildTools: just clone the repo, revert to the latest 1.20 commit, apply the nms patches and compile and install. Another option is to just wait and hope that most of my users will migrate to 1.20.1 soon so that I don't have to maintain 1.20 support anymore. Addressing this as wontfix for now.
infrastructure
ci builds are broken the modules for craftbukkit currently don t build in ci because the craftbukkit server dependency could not be found this is because running buildtools with rev actually builds craftbukkit md closed this as a wontfix so until support for craftbukkit will be dropped this remains an issue one way to remedy this would be to build craftbukkit without buildtools just clone the repo revert to the latest commit apply the nms patches and compile and install another option is to just wait and hope that most of my users will migrate to soon so that i don t have to maintain support anymore addressing this as wontfix for now
1
4,552
5,166,246,106
IssuesEvent
2017-01-17 15:47:21
elmsln/elmsln
https://api.github.com/repos/elmsln/elmsln
closed
drush 7 ugrid plugin incompatibility
bug drush related infrastructure
Drush 7 (standard version we run now) doesn't have some of the functions used in the ugrid plugin. This means roles arne't granted when doing backend jobs (rare that it happens).
1.0
drush 7 ugrid plugin incompatibility - Drush 7 (standard version we run now) doesn't have some of the functions used in the ugrid plugin. This means roles arne't granted when doing backend jobs (rare that it happens).
infrastructure
drush ugrid plugin incompatibility drush standard version we run now doesn t have some of the functions used in the ugrid plugin this means roles arne t granted when doing backend jobs rare that it happens
1
1,165
3,060,371,694
IssuesEvent
2015-08-14 20:17:19
catapult-project/catapult
https://api.github.com/repos/catapult-project/catapult
closed
Initial dev-server unification steps
Infrastructure
0) In each project's x_build folder, (bikeshed0 1) Move dev_Server and its tests up to build/ 2) Introduce PerProjectDevServer class, that provides - AddOptionsToGroup(argparse_group): adds cmdline options to provided argparse group (https://docs.python.org/3/library/argparse.html#argparse.ArgumentParser.add_argument_group) - GetHandlers(options) -> [wsgi route tuples] , where options is all results of argparse - GetSourcePaths() -> [] 3) dev_server has hard coded list of project dev servers I suggest option 2 of a devserver class because the GetHandlers and AddOptionsToGroup are specific to the devserver use case. We could add them to the project, but then the: 1) the project would take on a dependency on wsgi/webapp2 for all uses of the project 2) the project file would get big... it is not a module, but a toplevel file, x_project.py, whereas if we put the per-project devserver stuffs in x_build, we can have class-per-handler were needed I suggest we keep the per-project tests.html for v1, which means that the test handlers are in the perproject parts of the devservers. And, the / redirects would point to some index.html that would let you pick which tests.html to jump to. We might do differetnly down the road , but this'd be a baby step.
1.0
Initial dev-server unification steps - 0) In each project's x_build folder, (bikeshed0 1) Move dev_Server and its tests up to build/ 2) Introduce PerProjectDevServer class, that provides - AddOptionsToGroup(argparse_group): adds cmdline options to provided argparse group (https://docs.python.org/3/library/argparse.html#argparse.ArgumentParser.add_argument_group) - GetHandlers(options) -> [wsgi route tuples] , where options is all results of argparse - GetSourcePaths() -> [] 3) dev_server has hard coded list of project dev servers I suggest option 2 of a devserver class because the GetHandlers and AddOptionsToGroup are specific to the devserver use case. We could add them to the project, but then the: 1) the project would take on a dependency on wsgi/webapp2 for all uses of the project 2) the project file would get big... it is not a module, but a toplevel file, x_project.py, whereas if we put the per-project devserver stuffs in x_build, we can have class-per-handler were needed I suggest we keep the per-project tests.html for v1, which means that the test handlers are in the perproject parts of the devservers. And, the / redirects would point to some index.html that would let you pick which tests.html to jump to. We might do differetnly down the road , but this'd be a baby step.
infrastructure
initial dev server unification steps in each project s x build folder move dev server and its tests up to build introduce perprojectdevserver class that provides addoptionstogroup argparse group adds cmdline options to provided argparse group gethandlers options where options is all results of argparse getsourcepaths dev server has hard coded list of project dev servers i suggest option of a devserver class because the gethandlers and addoptionstogroup are specific to the devserver use case we could add them to the project but then the the project would take on a dependency on wsgi for all uses of the project the project file would get big it is not a module but a toplevel file x project py whereas if we put the per project devserver stuffs in x build we can have class per handler were needed i suggest we keep the per project tests html for which means that the test handlers are in the perproject parts of the devservers and the redirects would point to some index html that would let you pick which tests html to jump to we might do differetnly down the road but this d be a baby step
1
272,617
29,795,056,914
IssuesEvent
2023-06-16 01:07:31
billmcchesney1/hadoop
https://api.github.com/repos/billmcchesney1/hadoop
closed
CVE-2020-9547 (Critical) detected in jackson-databind-2.9.10.1.jar - autoclosed
Mend: dependency security vulnerability
## CVE-2020-9547 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9547>CVE-2020-9547</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: 2.9.10.4</p> </p> </details> <p></p>
True
CVE-2020-9547 (Critical) detected in jackson-databind-2.9.10.1.jar - autoclosed - ## CVE-2020-9547 - Critical Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9547>CVE-2020-9547</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: 2.9.10.4</p> </p> </details> <p></p>
non_infrastructure
cve critical detected in jackson databind jar autoclosed cve critical severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice documentstore target lib jackson databind jar hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase common target lib jackson databind jar hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase client target lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch trunk vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
2,952
3,974,578,382
IssuesEvent
2016-05-04 22:50:31
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Some csproj files still use DNXCore,Version=5.0 as the moniker
2 - In Progress Infrastructure
The document for .NET Platform Standard says that `dnxcore` is deprecated, however it's still being used in a couple of `.csproj` files: https://github.com/dotnet/corefx/search?l=xml&q=DNXCore&utf8=%E2%9C%93 Not sure if this is intentional or not. cc @ericstj
1.0
Some csproj files still use DNXCore,Version=5.0 as the moniker - The document for .NET Platform Standard says that `dnxcore` is deprecated, however it's still being used in a couple of `.csproj` files: https://github.com/dotnet/corefx/search?l=xml&q=DNXCore&utf8=%E2%9C%93 Not sure if this is intentional or not. cc @ericstj
infrastructure
some csproj files still use dnxcore version as the moniker the document for net platform standard says that dnxcore is deprecated however it s still being used in a couple of csproj files not sure if this is intentional or not cc ericstj
1
31,799
26,132,130,779
IssuesEvent
2022-12-29 07:05:18
iree-org/iree
https://api.github.com/repos/iree-org/iree
closed
Migrate compilation statistics pipeline to the e2e test framework
infrastructure infrastructure/benchmark
e2e test framework provides a clear separation between compilation and benchmark run. Compilation statistics pipeline tools can get the dedicated compilation flags and tags from it.
2.0
Migrate compilation statistics pipeline to the e2e test framework - e2e test framework provides a clear separation between compilation and benchmark run. Compilation statistics pipeline tools can get the dedicated compilation flags and tags from it.
infrastructure
migrate compilation statistics pipeline to the test framework test framework provides a clear separation between compilation and benchmark run compilation statistics pipeline tools can get the dedicated compilation flags and tags from it
1
14,657
8,657,635,435
IssuesEvent
2018-11-27 21:55:10
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Lag when typing long text blocks with a lot of complex emojis.
a: text input a: typography severe: performance
Type many complex emojis in a textfield such as the one in flutter gallery. After typing a large amount of text, further input becomes very slow and laggy/sluggish (on debug mode).
True
Lag when typing long text blocks with a lot of complex emojis. - Type many complex emojis in a textfield such as the one in flutter gallery. After typing a large amount of text, further input becomes very slow and laggy/sluggish (on debug mode).
non_infrastructure
lag when typing long text blocks with a lot of complex emojis type many complex emojis in a textfield such as the one in flutter gallery after typing a large amount of text further input becomes very slow and laggy sluggish on debug mode
0
10,682
8,678,909,250
IssuesEvent
2018-11-30 21:35:18
NCarranza-87/OnlineComputerStore
https://api.github.com/repos/NCarranza-87/OnlineComputerStore
closed
Update database
Infrastructure Priority: Moderate Work Estimate: >30mins.
Update database using code first migration for Members attributes in the database.
1.0
Update database - Update database using code first migration for Members attributes in the database.
infrastructure
update database update database using code first migration for members attributes in the database
1
25,476
18,777,995,374
IssuesEvent
2021-11-08 00:02:37
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
opened
Changes to user settings/configuration in GUI should be saved immediately
bug interface/infrastructure
Currently these changes are only saved when the user closes the GUI.
1.0
Changes to user settings/configuration in GUI should be saved immediately - Currently these changes are only saved when the user closes the GUI.
infrastructure
changes to user settings configuration in gui should be saved immediately currently these changes are only saved when the user closes the gui
1
24,433
17,263,536,068
IssuesEvent
2021-07-22 10:55:05
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
closed
Duplicate detection during region import does not work
backend bug import & export infrastructure major vaadin-app
<!-- If you've never submitted an issue to the SORMAS repository before or this is your first time using this template, please read the Contributing guidelines (https://github.com/hzi-braunschweig/SORMAS-Project/blob/development/docs/CONTRIBUTING.md) for an explanation of the information we need you to provide. You don't have to remove this comment or any other comment from this issue as they will automatically be hidden. --> ### Bug Description <!-- Mandatory --> #1621 added duplicate detection for infrastructure data. Importing infrastructure results in duplicate data in the system. ### Steps to Reproduce <!-- Optional; please add more steps if necessary --> 1. export region data on a server 2. re-import the exported regions but do not choose `override existing entries with imported data`. This create duplicate entries even if the name and epid code are the same. 3. Import a region with the same name as a region in the system but with another country. You get the error message `The database already contains a region with this name.` ### Expected Behavior <!-- Optional --> - If `override existing entries with imported data` == t and duplicate are identified, merge the two records using the one being imported as the reference - If `override existing entries with imported data` == f and duplicate are identified, do not import and add the entry to the import error report. - Two regions with the same name but different countries, continents, etc should not be considered as duplicates. Duplicate should consider the parent/ higher administrative unit. ### Screenshots <!-- Optional --> ![image](https://user-images.githubusercontent.com/12996724/124833035-d9a04f00-df7d-11eb-8887-33829a4d3833.png) ![image](https://user-images.githubusercontent.com/12996724/124833104-f472c380-df7d-11eb-8778-bbdfbec5e207.png) ### System Details <!-- Mandatory; you only have to specify the Server URL if the error appeared on a publicly available test server --> * Device: * SORMAS version: 1.61 * Android version/Browser: * Server URL: https://release-international.sormas.netzlink.com/sormas-ui/ * User Role: admin ### Additional Information <!-- Optional -->
1.0
Duplicate detection during region import does not work - <!-- If you've never submitted an issue to the SORMAS repository before or this is your first time using this template, please read the Contributing guidelines (https://github.com/hzi-braunschweig/SORMAS-Project/blob/development/docs/CONTRIBUTING.md) for an explanation of the information we need you to provide. You don't have to remove this comment or any other comment from this issue as they will automatically be hidden. --> ### Bug Description <!-- Mandatory --> #1621 added duplicate detection for infrastructure data. Importing infrastructure results in duplicate data in the system. ### Steps to Reproduce <!-- Optional; please add more steps if necessary --> 1. export region data on a server 2. re-import the exported regions but do not choose `override existing entries with imported data`. This create duplicate entries even if the name and epid code are the same. 3. Import a region with the same name as a region in the system but with another country. You get the error message `The database already contains a region with this name.` ### Expected Behavior <!-- Optional --> - If `override existing entries with imported data` == t and duplicate are identified, merge the two records using the one being imported as the reference - If `override existing entries with imported data` == f and duplicate are identified, do not import and add the entry to the import error report. - Two regions with the same name but different countries, continents, etc should not be considered as duplicates. Duplicate should consider the parent/ higher administrative unit. ### Screenshots <!-- Optional --> ![image](https://user-images.githubusercontent.com/12996724/124833035-d9a04f00-df7d-11eb-8887-33829a4d3833.png) ![image](https://user-images.githubusercontent.com/12996724/124833104-f472c380-df7d-11eb-8778-bbdfbec5e207.png) ### System Details <!-- Mandatory; you only have to specify the Server URL if the error appeared on a publicly available test server --> * Device: * SORMAS version: 1.61 * Android version/Browser: * Server URL: https://release-international.sormas.netzlink.com/sormas-ui/ * User Role: admin ### Additional Information <!-- Optional -->
infrastructure
duplicate detection during region import does not work if you ve never submitted an issue to the sormas repository before or this is your first time using this template please read the contributing guidelines for an explanation of the information we need you to provide you don t have to remove this comment or any other comment from this issue as they will automatically be hidden bug description added duplicate detection for infrastructure data importing infrastructure results in duplicate data in the system steps to reproduce export region data on a server re import the exported regions but do not choose override existing entries with imported data this create duplicate entries even if the name and epid code are the same import a region with the same name as a region in the system but with another country you get the error message the database already contains a region with this name expected behavior if override existing entries with imported data t and duplicate are identified merge the two records using the one being imported as the reference if override existing entries with imported data f and duplicate are identified do not import and add the entry to the import error report two regions with the same name but different countries continents etc should not be considered as duplicates duplicate should consider the parent higher administrative unit screenshots system details device sormas version android version browser server url user role admin additional information
1
50,673
6,106,608,009
IssuesEvent
2017-06-21 05:05:01
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Test: System.Net.Security.Tests.LoggingTest/EventSource_EventsRaisedAsExpected failed with "Xunit.Sdk.EqualException"
area-System.Net.Security test-run-uwp-ilc
Opened on behalf of @Jiayili1 The test `System.Net.Security.Tests.LoggingTest/EventSource_EventsRaisedAsExpected` has failed. Assert.Equal() Failure\r Expected: 42\r Actual: -323331298 Stack Trace: at xunit.console.netcore!<BaseAddress>+0x8615a8 at System.Net.Security.Tests.LoggingTest.EventSource_EventsRaisedAsExpected() at xunit.console.netcore!<BaseAddress>+0x932f2e at xunit.console.netcore!<BaseAddress>+0x80751c at xunit.console.netcore!<BaseAddress>+0x807374 Build : Master - 20170619.01 (UWP ILC Tests) Failing configurations: - Windows.10.Amd64-x86 - Debug - Release - Windows.10.Amd64-x64 - Debug - Release Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Filc~2F/build/20170619.01/workItem/System.Net.Security.Tests/analysis/xunit/System.Net.Security.Tests.LoggingTest~2FEventSource_EventsRaisedAsExpected
1.0
Test: System.Net.Security.Tests.LoggingTest/EventSource_EventsRaisedAsExpected failed with "Xunit.Sdk.EqualException" - Opened on behalf of @Jiayili1 The test `System.Net.Security.Tests.LoggingTest/EventSource_EventsRaisedAsExpected` has failed. Assert.Equal() Failure\r Expected: 42\r Actual: -323331298 Stack Trace: at xunit.console.netcore!<BaseAddress>+0x8615a8 at System.Net.Security.Tests.LoggingTest.EventSource_EventsRaisedAsExpected() at xunit.console.netcore!<BaseAddress>+0x932f2e at xunit.console.netcore!<BaseAddress>+0x80751c at xunit.console.netcore!<BaseAddress>+0x807374 Build : Master - 20170619.01 (UWP ILC Tests) Failing configurations: - Windows.10.Amd64-x86 - Debug - Release - Windows.10.Amd64-x64 - Debug - Release Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Filc~2F/build/20170619.01/workItem/System.Net.Security.Tests/analysis/xunit/System.Net.Security.Tests.LoggingTest~2FEventSource_EventsRaisedAsExpected
non_infrastructure
test system net security tests loggingtest eventsource eventsraisedasexpected failed with xunit sdk equalexception opened on behalf of the test system net security tests loggingtest eventsource eventsraisedasexpected has failed assert equal failure r expected r actual stack trace at xunit console netcore at system net security tests loggingtest eventsource eventsraisedasexpected at xunit console netcore at xunit console netcore at xunit console netcore build master uwp ilc tests failing configurations windows debug release windows debug release detail
0
9,300
6,203,916,669
IssuesEvent
2017-07-06 13:07:36
gbif/portal16
https://api.github.com/repos/gbif/portal16
closed
maps - zoom buttons are occasionally hidden in mobile
browser specific bug maps usability
zoom buttons are occasionally hidden in mobile. This goes also for openlayers own controls. Seemingly the canvas z-index blocks. Or perhaps it is a ios rendering error. Either way it is a shame that the controls disappear.
True
maps - zoom buttons are occasionally hidden in mobile - zoom buttons are occasionally hidden in mobile. This goes also for openlayers own controls. Seemingly the canvas z-index blocks. Or perhaps it is a ios rendering error. Either way it is a shame that the controls disappear.
non_infrastructure
maps zoom buttons are occasionally hidden in mobile zoom buttons are occasionally hidden in mobile this goes also for openlayers own controls seemingly the canvas z index blocks or perhaps it is a ios rendering error either way it is a shame that the controls disappear
0
794,642
28,042,819,677
IssuesEvent
2023-03-28 19:57:46
Satellite-im/Uplink
https://api.github.com/repos/Satellite-im/Uplink
closed
Chat - Unable to send attachment on replies
Chat High Priority
**Issue:** When we are replying to someone in chat and we try to send a file, the message will not go through **Screen recording:** https://user-images.githubusercontent.com/93608357/227601205-d99e9150-8c16-4035-a923-7ec0438c9b04.mov
1.0
Chat - Unable to send attachment on replies - **Issue:** When we are replying to someone in chat and we try to send a file, the message will not go through **Screen recording:** https://user-images.githubusercontent.com/93608357/227601205-d99e9150-8c16-4035-a923-7ec0438c9b04.mov
non_infrastructure
chat unable to send attachment on replies issue when we are replying to someone in chat and we try to send a file the message will not go through screen recording
0
29,494
24,044,066,211
IssuesEvent
2022-09-16 06:31:41
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Add automatic flake detection into CI [Blocked: #59]
Type: Improvement Priority: Nice-to-have issue_type_infrastructure dev_team issue_user_developer
This is a bit more of a suggested solution, but I was thinking that we might be able to add automatic test flake detection for new PRs by: 1. Detecting new tests added in a PR (maybe just test suites to start) 2. Running these tests a certain # of times in parallel (which will be a lot easier to do with Bazel, so marking this as blocked on #59). This will help protect against accidentally introducing newly flaky tests. It doesn't help against new code causing existing tests to become flaky (that will probably require detection against the develop branch rather than per PR branch).
1.0
Add automatic flake detection into CI [Blocked: #59] - This is a bit more of a suggested solution, but I was thinking that we might be able to add automatic test flake detection for new PRs by: 1. Detecting new tests added in a PR (maybe just test suites to start) 2. Running these tests a certain # of times in parallel (which will be a lot easier to do with Bazel, so marking this as blocked on #59). This will help protect against accidentally introducing newly flaky tests. It doesn't help against new code causing existing tests to become flaky (that will probably require detection against the develop branch rather than per PR branch).
infrastructure
add automatic flake detection into ci this is a bit more of a suggested solution but i was thinking that we might be able to add automatic test flake detection for new prs by detecting new tests added in a pr maybe just test suites to start running these tests a certain of times in parallel which will be a lot easier to do with bazel so marking this as blocked on this will help protect against accidentally introducing newly flaky tests it doesn t help against new code causing existing tests to become flaky that will probably require detection against the develop branch rather than per pr branch
1
22,019
14,965,811,228
IssuesEvent
2021-01-27 13:51:35
SDU-eScience/UCloud
https://api.github.com/repos/SDU-eScience/UCloud
closed
Websocket Timeout on job start
backend infrastructure more info required
We recieved a support ticket saying that 5 jobs had started when attempting to start a single. Found timeouts on the websockets in attached logs. Guess is that they resulted in retires but all requests was infact recieved by Kubernetes resulting in 5 jobs. [https://app.zenhub.com/files/102488820/dfa9989a-6fac-4c97-b7cb-5268bd492a1e/download](https://app.zenhub.com/files/102488820/dfa9989a-6fac-4c97-b7cb-5268bd492a1e/download)
1.0
Websocket Timeout on job start - We recieved a support ticket saying that 5 jobs had started when attempting to start a single. Found timeouts on the websockets in attached logs. Guess is that they resulted in retires but all requests was infact recieved by Kubernetes resulting in 5 jobs. [https://app.zenhub.com/files/102488820/dfa9989a-6fac-4c97-b7cb-5268bd492a1e/download](https://app.zenhub.com/files/102488820/dfa9989a-6fac-4c97-b7cb-5268bd492a1e/download)
infrastructure
websocket timeout on job start we recieved a support ticket saying that jobs had started when attempting to start a single found timeouts on the websockets in attached logs guess is that they resulted in retires but all requests was infact recieved by kubernetes resulting in jobs
1
33,793
9,205,463,233
IssuesEvent
2019-03-08 10:38:38
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
Cant easily change locale in win or mac and cant override sys locale
Category: Build/Install Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report
--- Author Name: **Tim Sutton** (Tim Sutton) Original Redmine Issue: 798, https://issues.qgis.org/issues/798 Original Assignee: Tim Sutton --- In windows and mac, its not easy for user to set the locale. Users are submitting questions to the mailing lists because they are not able to set their preferred locale.
1.0
Cant easily change locale in win or mac and cant override sys locale - --- Author Name: **Tim Sutton** (Tim Sutton) Original Redmine Issue: 798, https://issues.qgis.org/issues/798 Original Assignee: Tim Sutton --- In windows and mac, its not easy for user to set the locale. Users are submitting questions to the mailing lists because they are not able to set their preferred locale.
non_infrastructure
cant easily change locale in win or mac and cant override sys locale author name tim sutton tim sutton original redmine issue original assignee tim sutton in windows and mac its not easy for user to set the locale users are submitting questions to the mailing lists because they are not able to set their preferred locale
0
9,021
7,778,329,347
IssuesEvent
2018-06-05 13:55:15
coq/coq
https://api.github.com/repos/coq/coq
closed
Linter reports whitespace issues not fixed by `git rebase --whitespace=fix`
kind: infrastructure
Previous discussions on the linter has concluded that whitespace linting was mainly about aesthetics, IIUC. I think it is fine to care about aesthetics, as long as the cost is low or non-existent. Running `git rebase --whitespace=fix` as part of my PR workflow is fine IMHO. However, the linter currently reports "whitespace errors" which are not fixed by that `git` command, like a blank line at the end of a file. So please let me know if a command fixing all issues reported by the linter exists, and if not, what checks we should disable to make it true.
1.0
Linter reports whitespace issues not fixed by `git rebase --whitespace=fix` - Previous discussions on the linter has concluded that whitespace linting was mainly about aesthetics, IIUC. I think it is fine to care about aesthetics, as long as the cost is low or non-existent. Running `git rebase --whitespace=fix` as part of my PR workflow is fine IMHO. However, the linter currently reports "whitespace errors" which are not fixed by that `git` command, like a blank line at the end of a file. So please let me know if a command fixing all issues reported by the linter exists, and if not, what checks we should disable to make it true.
infrastructure
linter reports whitespace issues not fixed by git rebase whitespace fix previous discussions on the linter has concluded that whitespace linting was mainly about aesthetics iiuc i think it is fine to care about aesthetics as long as the cost is low or non existent running git rebase whitespace fix as part of my pr workflow is fine imho however the linter currently reports whitespace errors which are not fixed by that git command like a blank line at the end of a file so please let me know if a command fixing all issues reported by the linter exists and if not what checks we should disable to make it true
1
3,780
4,561,612,234
IssuesEvent
2016-09-14 12:24:43
lampepfl/dotty
https://api.github.com/repos/lampepfl/dotty
closed
Errors not reported by Jenkins
Infrastructure
See #794. There should be an error reported in this branch, on an assertion violation in expandPrivate in t2183.scala. Locally it always fails for me. But the Jenkins test suite succeeds.
1.0
Errors not reported by Jenkins - See #794. There should be an error reported in this branch, on an assertion violation in expandPrivate in t2183.scala. Locally it always fails for me. But the Jenkins test suite succeeds.
infrastructure
errors not reported by jenkins see there should be an error reported in this branch on an assertion violation in expandprivate in scala locally it always fails for me but the jenkins test suite succeeds
1
670,801
22,705,196,193
IssuesEvent
2022-07-05 14:06:23
Quansight/Quansight-website
https://api.github.com/repos/Quansight/Quansight-website
closed
[BUG] - External URLs in 'Footer Navigation > Footer Link' components are treated as internal navigation, leading to broken links
type: bug 🐛 priority: high 🌋 LLC 🤝 area: react owner: Naturaily
### What site is this for? Quansight LLC ### Expected behavior When an external URL is applied to a Footer Link component with a surrounding Footer Navigation, that URL should be used as-is. ![image](https://user-images.githubusercontent.com/11325439/177198392-b4822c51-85aa-4416-9ba8-95d14e8f31ea.png) ### Actual behavior Instead, the URL is treated as if it were an internal slug: ![image](https://user-images.githubusercontent.com/11325439/177198657-2299c010-2474-435f-8152-e9fe38d545e7.png) ### How to Reproduce the problem? _No response_ ### Anything else? The Footer is a shared component, so this likely affects any external URLs in the Labs site footer, also.
1.0
[BUG] - External URLs in 'Footer Navigation > Footer Link' components are treated as internal navigation, leading to broken links - ### What site is this for? Quansight LLC ### Expected behavior When an external URL is applied to a Footer Link component with a surrounding Footer Navigation, that URL should be used as-is. ![image](https://user-images.githubusercontent.com/11325439/177198392-b4822c51-85aa-4416-9ba8-95d14e8f31ea.png) ### Actual behavior Instead, the URL is treated as if it were an internal slug: ![image](https://user-images.githubusercontent.com/11325439/177198657-2299c010-2474-435f-8152-e9fe38d545e7.png) ### How to Reproduce the problem? _No response_ ### Anything else? The Footer is a shared component, so this likely affects any external URLs in the Labs site footer, also.
non_infrastructure
external urls in footer navigation footer link components are treated as internal navigation leading to broken links what site is this for quansight llc expected behavior when an external url is applied to a footer link component with a surrounding footer navigation that url should be used as is actual behavior instead the url is treated as if it were an internal slug how to reproduce the problem no response anything else the footer is a shared component so this likely affects any external urls in the labs site footer also
0
107,589
4,311,378,473
IssuesEvent
2016-07-21 22:30:01
ngageoint/hootenanny-ui
https://api.github.com/repos/ngageoint/hootenanny-ui
reopened
Disappearing Highlight
Category: UI Priority: Medium Status: Accepted/Closed
Original Ticket: https://github.com/DigitalGlobe/VGI-team-repo/issues/183 1. Open a dataset in review mode. 2. Turn on Highlight Edited Features. 3. Make a change to a feature. 4. Pan AWAY from the changed & highlighted feature. 5. Pan BACK to the changed & highlighted feature. 6. Highlight is GONE. 7. Zoom out a little bit at a time. 8. Highlight is BACK!!! Alternatively, if you make a change & zoom out substantially and then zoom back in, the change is reset (same thing as previously reported, when you draw a feature (point, line, or poly)) AND if I zoom out far enough, when the highlighting disappears the map goes into refresh mode and says "Please wait while loading missing features" This is occuring in stable AWS build and the nightly build.
1.0
Disappearing Highlight - Original Ticket: https://github.com/DigitalGlobe/VGI-team-repo/issues/183 1. Open a dataset in review mode. 2. Turn on Highlight Edited Features. 3. Make a change to a feature. 4. Pan AWAY from the changed & highlighted feature. 5. Pan BACK to the changed & highlighted feature. 6. Highlight is GONE. 7. Zoom out a little bit at a time. 8. Highlight is BACK!!! Alternatively, if you make a change & zoom out substantially and then zoom back in, the change is reset (same thing as previously reported, when you draw a feature (point, line, or poly)) AND if I zoom out far enough, when the highlighting disappears the map goes into refresh mode and says "Please wait while loading missing features" This is occuring in stable AWS build and the nightly build.
non_infrastructure
disappearing highlight original ticket open a dataset in review mode turn on highlight edited features make a change to a feature pan away from the changed highlighted feature pan back to the changed highlighted feature highlight is gone zoom out a little bit at a time highlight is back alternatively if you make a change zoom out substantially and then zoom back in the change is reset same thing as previously reported when you draw a feature point line or poly and if i zoom out far enough when the highlighting disappears the map goes into refresh mode and says please wait while loading missing features this is occuring in stable aws build and the nightly build
0
173,487
6,525,601,342
IssuesEvent
2017-08-29 16:26:53
pravega/pravega
https://api.github.com/repos/pravega/pravega
closed
Concat operation in HDFS fails with "file is under construction"
area/tier2 exp/expert kind/bug priority/P0 status/in-progress
**Problem description** Concat operation fails with above HDFS error. Here is the sequence of operations executed: MultiSegmentStore test starts with 1 instances of SSS service -> Tests create stream and performs a read write test (This was successful) ->Scales SSS to 3 instances -> Tests create stream and performs read write test (This was successful) -> Scales the SSS service to 2 instances -> Tests create stream and perform read write test (This failed). ``` 2017-05-18 10:44:16,879 902995 [segment-store-20] ERROR i.p.s.server.writer.StorageWriter - StorageWriter[1]: Iteration[1].Error. org.apache.hadoop.ipc.RemoteException: concat: target file /_system/_requeststream/0_0_1 is under construction at org.apache.hadoop.hdfs.server.namenode.FSDirConcatOp.verifyTargetFile(FSDirConcatOp.java:99) at org.apache.hadoop.hdfs.server.namenode.FSDirConcatOp.concat(FSDirConcatOp.java:69) ``` **Problem location** HDFSStorage.java concat and fencing. **Suggestions for an improvement** It is observed that these tests were run with HDFS 2.7.2. Recommended version is 2.7.3.
1.0
Concat operation in HDFS fails with "file is under construction" - **Problem description** Concat operation fails with above HDFS error. Here is the sequence of operations executed: MultiSegmentStore test starts with 1 instances of SSS service -> Tests create stream and performs a read write test (This was successful) ->Scales SSS to 3 instances -> Tests create stream and performs read write test (This was successful) -> Scales the SSS service to 2 instances -> Tests create stream and perform read write test (This failed). ``` 2017-05-18 10:44:16,879 902995 [segment-store-20] ERROR i.p.s.server.writer.StorageWriter - StorageWriter[1]: Iteration[1].Error. org.apache.hadoop.ipc.RemoteException: concat: target file /_system/_requeststream/0_0_1 is under construction at org.apache.hadoop.hdfs.server.namenode.FSDirConcatOp.verifyTargetFile(FSDirConcatOp.java:99) at org.apache.hadoop.hdfs.server.namenode.FSDirConcatOp.concat(FSDirConcatOp.java:69) ``` **Problem location** HDFSStorage.java concat and fencing. **Suggestions for an improvement** It is observed that these tests were run with HDFS 2.7.2. Recommended version is 2.7.3.
non_infrastructure
concat operation in hdfs fails with file is under construction problem description concat operation fails with above hdfs error here is the sequence of operations executed multisegmentstore test starts with instances of sss service tests create stream and performs a read write test this was successful scales sss to instances tests create stream and performs read write test this was successful scales the sss service to instances tests create stream and perform read write test this failed error i p s server writer storagewriter storagewriter iteration error org apache hadoop ipc remoteexception concat target file system requeststream is under construction at org apache hadoop hdfs server namenode fsdirconcatop verifytargetfile fsdirconcatop java at org apache hadoop hdfs server namenode fsdirconcatop concat fsdirconcatop java problem location hdfsstorage java concat and fencing suggestions for an improvement it is observed that these tests were run with hdfs recommended version is
0
603,384
18,544,974,339
IssuesEvent
2021-10-21 20:47:32
JustArchiNET/ASF-ui
https://api.github.com/repos/JustArchiNET/ASF-ui
closed
Add support for ASF 2FA management endpoints
✨ Enhancement 🟢 Low priority
### Checklist - [X] I read and understood ASF-ui's **[Contributing Guidelines](https://github.com/JustArchiNET/ASF-ui/blob/main/.github/CONTRIBUTING.md)** - [X] I also read **[FAQ](https://github.com/JustArchiNET/ASF-ui/wiki/FAQ)** - [X] My idea doesn't duplicate existing ASF-ui functionality described on the **[wiki](https://github.com/JustArchiNET/ASF-ui/wiki)** - [X] I believe that my idea falls into ASF-ui's scope and should be offered as part of ASF-ui built-in functionality - [X] This is not **[ASF suggestion](https://github.com/JustArchiNET/ArchiSteamFarm/issues/new/choose)** ### Enhancement purpose ASF recently got new endpoints to support management of ASF 2FA. These should be supported in the official GUI. ### Solution Add a "Remove Authenticator from ASF" button to a bots 2FA modal. If the 2FA modal on a bot without authenticator is opened, there should be a button with a text like "Import authenticator" opening a file-selection menu and using the selected file for import. ### Why currently available solutions are not sufficient? Currently we have to either use swagger, `curl` or copy the file to our ASF instances config folder manually. ### Can you help us with this enhancement idea? Somehow, I can test and offer feedback, but can't code ### Additional info [related merge request](https://github.com/JustArchiNET/ArchiSteamFarm/pull/2426)
1.0
Add support for ASF 2FA management endpoints - ### Checklist - [X] I read and understood ASF-ui's **[Contributing Guidelines](https://github.com/JustArchiNET/ASF-ui/blob/main/.github/CONTRIBUTING.md)** - [X] I also read **[FAQ](https://github.com/JustArchiNET/ASF-ui/wiki/FAQ)** - [X] My idea doesn't duplicate existing ASF-ui functionality described on the **[wiki](https://github.com/JustArchiNET/ASF-ui/wiki)** - [X] I believe that my idea falls into ASF-ui's scope and should be offered as part of ASF-ui built-in functionality - [X] This is not **[ASF suggestion](https://github.com/JustArchiNET/ArchiSteamFarm/issues/new/choose)** ### Enhancement purpose ASF recently got new endpoints to support management of ASF 2FA. These should be supported in the official GUI. ### Solution Add a "Remove Authenticator from ASF" button to a bots 2FA modal. If the 2FA modal on a bot without authenticator is opened, there should be a button with a text like "Import authenticator" opening a file-selection menu and using the selected file for import. ### Why currently available solutions are not sufficient? Currently we have to either use swagger, `curl` or copy the file to our ASF instances config folder manually. ### Can you help us with this enhancement idea? Somehow, I can test and offer feedback, but can't code ### Additional info [related merge request](https://github.com/JustArchiNET/ArchiSteamFarm/pull/2426)
non_infrastructure
add support for asf management endpoints checklist i read and understood asf ui s i also read my idea doesn t duplicate existing asf ui functionality described on the i believe that my idea falls into asf ui s scope and should be offered as part of asf ui built in functionality this is not enhancement purpose asf recently got new endpoints to support management of asf these should be supported in the official gui solution add a remove authenticator from asf button to a bots modal if the modal on a bot without authenticator is opened there should be a button with a text like import authenticator opening a file selection menu and using the selected file for import why currently available solutions are not sufficient currently we have to either use swagger curl or copy the file to our asf instances config folder manually can you help us with this enhancement idea somehow i can test and offer feedback but can t code additional info
0
235,573
25,955,213,095
IssuesEvent
2022-12-18 05:33:55
Dima2022/JS-Demo
https://api.github.com/repos/Dima2022/JS-Demo
closed
CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz, ansi-regex-5.0.0.tgz - autoclosed
security vulnerability
## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-5.0.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ansi-align/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - nodemon-2.0.7.tgz (Root Library) - update-notifier-4.1.3.tgz - boxen-4.2.0.tgz - ansi-align-3.0.0.tgz - string-width-3.1.0.tgz - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-5.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/mocha/node_modules/ansi-regex/package.json,/node_modules/widest-line/node_modules/ansi-regex/package.json,/node_modules/boxen/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - mocha-8.3.2.tgz (Root Library) - yargs-16.2.0.tgz - string-width-4.2.2.tgz - strip-ansi-6.0.0.tgz - :x: **ansi-regex-5.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/4edf46ace164b01728ef7066c6a8e7464b89143a">4edf46ace164b01728ef7066c6a8e7464b89143a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution (ansi-regex): 4.1.1</p> <p>Direct dependency fix Resolution (nodemon): 2.0.8</p><p>Fix Resolution (ansi-regex): 5.0.1</p> <p>Direct dependency fix Resolution (mocha): 8.4.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz, ansi-regex-5.0.0.tgz - autoclosed - ## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-5.0.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ansi-align/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - nodemon-2.0.7.tgz (Root Library) - update-notifier-4.1.3.tgz - boxen-4.2.0.tgz - ansi-align-3.0.0.tgz - string-width-3.1.0.tgz - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-5.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/mocha/node_modules/ansi-regex/package.json,/node_modules/widest-line/node_modules/ansi-regex/package.json,/node_modules/boxen/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - mocha-8.3.2.tgz (Root Library) - yargs-16.2.0.tgz - string-width-4.2.2.tgz - strip-ansi-6.0.0.tgz - :x: **ansi-regex-5.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Dima2022/JS-Demo/commit/4edf46ace164b01728ef7066c6a8e7464b89143a">4edf46ace164b01728ef7066c6a8e7464b89143a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution (ansi-regex): 4.1.1</p> <p>Direct dependency fix Resolution (nodemon): 2.0.8</p><p>Fix Resolution (ansi-regex): 5.0.1</p> <p>Direct dependency fix Resolution (mocha): 8.4.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_infrastructure
cve high detected in ansi regex tgz ansi regex tgz autoclosed cve high severity vulnerability vulnerable libraries ansi regex tgz ansi regex tgz ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file package json path to vulnerable library node modules ansi align node modules ansi regex package json dependency hierarchy nodemon tgz root library update notifier tgz boxen tgz ansi align tgz string width tgz strip ansi tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file package json path to vulnerable library node modules mocha node modules ansi regex package json node modules widest line node modules ansi regex package json node modules boxen node modules ansi regex package json dependency hierarchy mocha tgz root library yargs tgz string width tgz strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href found in base branch master vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex direct dependency fix resolution nodemon fix resolution ansi regex direct dependency fix resolution mocha rescue worker helmet automatic remediation is available for this issue
0
32,950
27,104,469,316
IssuesEvent
2023-02-15 11:02:40
PostHog/posthog
https://api.github.com/repos/PostHog/posthog
closed
Plugin server doesn't URL-encode parameters when building database connection URL
bug infrastructure
## Bug description Upon updating PostHog to v1.41.4 and v1.42.0, the new plugin server does not start and repeatedly crash-loops with the error `getaddrinfo ENOTFOUND posthog`. Unfortunately this error doesn't come with a stack trace, so I cannot identify what lines attempted to resolve `posthog`, and my attempts at searching the code base have been unsuccessful as well. ## How to reproduce Unknown—this issue did not occur when running version 1.40.0 and only presented itself upon upgrading to >=1.41.4. ## Environment - [ ] PostHog Cloud - [X] self-hosted PostHog, version/commit: v1.41.4, v1.42.0, v1.43.0; Helm chart 30.0.0 ## Additional context Logs from the crash-looping pod: ``` ▶️ Starting plugin server... {"level":"info","time":1670009189771,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ⚡ @posthog/plugin-server v1.10.5"} {"level":"info","time":1670009189780,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ℹ️ 4 workers, 10 tasks per worker"} {"level":"info","time":1670009189781,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ℹ️ Connecting to all services:"} {"level":"info","time":1670009189809,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to ClickHouse..."} {"level":"info","time":1670009189875,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 ClickHouse ready"} {"level":"info","time":1670009189875,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to Kafka..."} {"level":"info","time":1670009189957,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Kafka ready"} {"level":"info","time":1670009189957,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to Postgresql..."} {"level":"info","time":1670009189958,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Postgresql ready"} {"level":"info","time":1670009189958,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to Redis..."} {"level":"info","time":1670009189965,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Redis ready"} {"level":"info","time":1670009189965,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to object storage..."} {"level":"info","time":1670009189965,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Object storage ready"} {"level":"info","time":1670009189983,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ✅ Connected to Redis!"} {"level":"error","time":1670009190029,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","error":"Error: getaddrinfo ENOTFOUND posthog\n at GetAddrInfoReqWrap.onlookup [as oncomplete] (node:dns:111:26)","msg":"[MAIN] 💥 Launchpad failure!"} {"level":"info","time":1670009190030,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 💤 Shutting down gracefully..."} {"level":"info","time":1670009190030,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👋 Over and out!"} ``` JSON payload of Sentry event: ``` { "event_id": "eb8c5c56880f4c5796be089ee0c1d2f6", "project": [REDACTED], "release": null, "dist": null, "platform": "node", "message": "", "datetime": "2022-12-02T19:03:18.393000Z", "tags": [ ["environment", "production"], ["handled", "yes"], ["level", "error"], ["mechanism", "generic"], ["runtime", "node v18.9.1"], ["runtime.name", "node"], ["server_name", "posthog-plugins-794d9fc977-xvt94"] ], "_metrics": { "bytes.ingested.event": 1259, "bytes.stored.event": 2450, "sample_rates": [{}] }, "breadcrumbs": { "values": [{ "timestamp": 1670007798.07, "type": "http", "category": "http", "level": "info", "data": { "method": "POST", "status_code": 200, "url": "https://[REDACTED]/?database=posthog&output_format_json_quote_64bit_integers=false" } }] }, "contexts": { "runtime": { "name": "node", "version": "v18.9.1", "type": "runtime" } }, "culprit": "GetAddrInfoReqWrap.onlookup [as oncomplete](node:dns)", "environment": "production", "exception": { "values": [{ "type": "Error", "value": "getaddrinfo ENOTFOUND posthog", "stacktrace": { "frames": [{ "function": "GetAddrInfoReqWrap.onlookup [as oncomplete]", "module": "node:dns", "filename": "node:dns", "abs_path": "node:dns", "lineno": 111, "colno": 26, "in_app": false }] }, "mechanism": { "type": "generic", "handled": true } }] }, "fingerprint": ["{{ default }}"], "grouping_config": { "enhancements": "eJybzDRxY3J-bm5-npWRgaGlroGxrpHxBABcYgcZ", "id": "newstyle:2019-10-29" }, "hashes": ["cb142f47bc36a31540a7101553ebbcd3"], "ingest_path": [{ "version": "22.11.0", "public_key": "[REDACTED]" }], "key_id": "2173646", "level": "error", "location": "node:dns", "logger": "", "metadata": { "display_title_with_tree_label": false, "filename": "node:dns", "function": "GetAddrInfoReqWrap.onlookup [as oncomplete]", "type": "Error", "value": "getaddrinfo ENOTFOUND posthog" }, "nodestore_insert": 1670007802.112358, "received": 1670007798.438835, "sdk": { "name": "sentry.javascript.node", "version": "7.10.0", "integrations": ["InboundFilters", "FunctionToString", "ContextLines", "Console", "Http", "OnUncaughtException", "OnUnhandledRejection", "LinkedErrors", "Postgres"], "packages": [{ "name": "npm:@sentry/node", "version": "7.10.0" }] }, "timestamp": 1670007798.393, "title": "Error: getaddrinfo ENOTFOUND posthog", "type": "error", "version": "7" } ```
1.0
Plugin server doesn't URL-encode parameters when building database connection URL - ## Bug description Upon updating PostHog to v1.41.4 and v1.42.0, the new plugin server does not start and repeatedly crash-loops with the error `getaddrinfo ENOTFOUND posthog`. Unfortunately this error doesn't come with a stack trace, so I cannot identify what lines attempted to resolve `posthog`, and my attempts at searching the code base have been unsuccessful as well. ## How to reproduce Unknown—this issue did not occur when running version 1.40.0 and only presented itself upon upgrading to >=1.41.4. ## Environment - [ ] PostHog Cloud - [X] self-hosted PostHog, version/commit: v1.41.4, v1.42.0, v1.43.0; Helm chart 30.0.0 ## Additional context Logs from the crash-looping pod: ``` ▶️ Starting plugin server... {"level":"info","time":1670009189771,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ⚡ @posthog/plugin-server v1.10.5"} {"level":"info","time":1670009189780,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ℹ️ 4 workers, 10 tasks per worker"} {"level":"info","time":1670009189781,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ℹ️ Connecting to all services:"} {"level":"info","time":1670009189809,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to ClickHouse..."} {"level":"info","time":1670009189875,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 ClickHouse ready"} {"level":"info","time":1670009189875,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to Kafka..."} {"level":"info","time":1670009189957,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Kafka ready"} {"level":"info","time":1670009189957,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to Postgresql..."} {"level":"info","time":1670009189958,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Postgresql ready"} {"level":"info","time":1670009189958,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to Redis..."} {"level":"info","time":1670009189965,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Redis ready"} {"level":"info","time":1670009189965,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 🤔 Connecting to object storage..."} {"level":"info","time":1670009189965,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👍 Object storage ready"} {"level":"info","time":1670009189983,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] ✅ Connected to Redis!"} {"level":"error","time":1670009190029,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","error":"Error: getaddrinfo ENOTFOUND posthog\n at GetAddrInfoReqWrap.onlookup [as oncomplete] (node:dns:111:26)","msg":"[MAIN] 💥 Launchpad failure!"} {"level":"info","time":1670009190030,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 💤 Shutting down gracefully..."} {"level":"info","time":1670009190030,"pid":22,"hostname":"posthog-plugins-5496f989cf-l8hjw","msg":"[MAIN] 👋 Over and out!"} ``` JSON payload of Sentry event: ``` { "event_id": "eb8c5c56880f4c5796be089ee0c1d2f6", "project": [REDACTED], "release": null, "dist": null, "platform": "node", "message": "", "datetime": "2022-12-02T19:03:18.393000Z", "tags": [ ["environment", "production"], ["handled", "yes"], ["level", "error"], ["mechanism", "generic"], ["runtime", "node v18.9.1"], ["runtime.name", "node"], ["server_name", "posthog-plugins-794d9fc977-xvt94"] ], "_metrics": { "bytes.ingested.event": 1259, "bytes.stored.event": 2450, "sample_rates": [{}] }, "breadcrumbs": { "values": [{ "timestamp": 1670007798.07, "type": "http", "category": "http", "level": "info", "data": { "method": "POST", "status_code": 200, "url": "https://[REDACTED]/?database=posthog&output_format_json_quote_64bit_integers=false" } }] }, "contexts": { "runtime": { "name": "node", "version": "v18.9.1", "type": "runtime" } }, "culprit": "GetAddrInfoReqWrap.onlookup [as oncomplete](node:dns)", "environment": "production", "exception": { "values": [{ "type": "Error", "value": "getaddrinfo ENOTFOUND posthog", "stacktrace": { "frames": [{ "function": "GetAddrInfoReqWrap.onlookup [as oncomplete]", "module": "node:dns", "filename": "node:dns", "abs_path": "node:dns", "lineno": 111, "colno": 26, "in_app": false }] }, "mechanism": { "type": "generic", "handled": true } }] }, "fingerprint": ["{{ default }}"], "grouping_config": { "enhancements": "eJybzDRxY3J-bm5-npWRgaGlroGxrpHxBABcYgcZ", "id": "newstyle:2019-10-29" }, "hashes": ["cb142f47bc36a31540a7101553ebbcd3"], "ingest_path": [{ "version": "22.11.0", "public_key": "[REDACTED]" }], "key_id": "2173646", "level": "error", "location": "node:dns", "logger": "", "metadata": { "display_title_with_tree_label": false, "filename": "node:dns", "function": "GetAddrInfoReqWrap.onlookup [as oncomplete]", "type": "Error", "value": "getaddrinfo ENOTFOUND posthog" }, "nodestore_insert": 1670007802.112358, "received": 1670007798.438835, "sdk": { "name": "sentry.javascript.node", "version": "7.10.0", "integrations": ["InboundFilters", "FunctionToString", "ContextLines", "Console", "Http", "OnUncaughtException", "OnUnhandledRejection", "LinkedErrors", "Postgres"], "packages": [{ "name": "npm:@sentry/node", "version": "7.10.0" }] }, "timestamp": 1670007798.393, "title": "Error: getaddrinfo ENOTFOUND posthog", "type": "error", "version": "7" } ```
infrastructure
plugin server doesn t url encode parameters when building database connection url bug description upon updating posthog to and the new plugin server does not start and repeatedly crash loops with the error getaddrinfo enotfound posthog unfortunately this error doesn t come with a stack trace so i cannot identify what lines attempted to resolve posthog and my attempts at searching the code base have been unsuccessful as well how to reproduce unknown—this issue did not occur when running version and only presented itself upon upgrading to environment posthog cloud self hosted posthog version commit helm chart additional context logs from the crash looping pod ▶️ starting plugin server level info time pid hostname posthog plugins msg ⚡ posthog plugin server level info time pid hostname posthog plugins msg ℹ️ workers tasks per worker level info time pid hostname posthog plugins msg ℹ️ connecting to all services level info time pid hostname posthog plugins msg 🤔 connecting to clickhouse level info time pid hostname posthog plugins msg 👍 clickhouse ready level info time pid hostname posthog plugins msg 🤔 connecting to kafka level info time pid hostname posthog plugins msg 👍 kafka ready level info time pid hostname posthog plugins msg 🤔 connecting to postgresql level info time pid hostname posthog plugins msg 👍 postgresql ready level info time pid hostname posthog plugins msg 🤔 connecting to redis level info time pid hostname posthog plugins msg 👍 redis ready level info time pid hostname posthog plugins msg 🤔 connecting to object storage level info time pid hostname posthog plugins msg 👍 object storage ready level info time pid hostname posthog plugins msg ✅ connected to redis level error time pid hostname posthog plugins error error getaddrinfo enotfound posthog n at getaddrinforeqwrap onlookup node dns msg 💥 launchpad failure level info time pid hostname posthog plugins msg 💤 shutting down gracefully level info time pid hostname posthog plugins msg 👋 over and out json payload of sentry event event id project release null dist null platform node message datetime tags metrics bytes ingested event bytes stored event sample rates breadcrumbs values timestamp type http category http level info data method post status code url https database posthog output format json quote integers false contexts runtime name node version type runtime culprit getaddrinforeqwrap onlookup node dns environment production exception values type error value getaddrinfo enotfound posthog stacktrace frames function getaddrinforeqwrap onlookup module node dns filename node dns abs path node dns lineno colno in app false mechanism type generic handled true fingerprint grouping config enhancements npwrgaglrogxrphxbabcygcz id newstyle hashes ingest path version public key key id level error location node dns logger metadata display title with tree label false filename node dns function getaddrinforeqwrap onlookup type error value getaddrinfo enotfound posthog nodestore insert received sdk name sentry javascript node version integrations packages name npm sentry node version timestamp title error getaddrinfo enotfound posthog type error version
1
136,105
12,700,574,636
IssuesEvent
2020-06-22 16:33:39
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
The plugin doc for UserSession plugin is missing
Accepted AppContext Documentation Priority: High bug
## Description <!-- Add here a few sentences describing the bug. --> There should be a related section in plugi documentation for this: https://devmapstore.geo-solutions.it/mapstore/docs/api/plugins#plugins.UserSession This section should be introduced [here](https://devmapstore.geo-solutions.it/mapstore/docs/api/plugins), at list containing a link to the developer guide [here](https://mapstore.readthedocs.io/en/latest/developer-guide/user-sessions/) if the plugin per se doesn't have any relevant option. ## Other useful information <!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
1.0
The plugin doc for UserSession plugin is missing - ## Description <!-- Add here a few sentences describing the bug. --> There should be a related section in plugi documentation for this: https://devmapstore.geo-solutions.it/mapstore/docs/api/plugins#plugins.UserSession This section should be introduced [here](https://devmapstore.geo-solutions.it/mapstore/docs/api/plugins), at list containing a link to the developer guide [here](https://mapstore.readthedocs.io/en/latest/developer-guide/user-sessions/) if the plugin per se doesn't have any relevant option. ## Other useful information <!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
non_infrastructure
the plugin doc for usersession plugin is missing description there should be a related section in plugi documentation for this this section should be introduced at list containing a link to the developer guide if the plugin per se doesn t have any relevant option other useful information
0
22,486
31,395,279,836
IssuesEvent
2023-08-26 21:22:57
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Joan is Awful from "Black Mirror" (Screenshots, Thumbnail, and Title Card added)
suggested title in process
Please add as much of the following info as you can: Title: Joan is Awful Type (film/tv show): TV show - drama Film or show in which it appears: Black Mirror Is the parent film/show streaming anywhere? Yes - Netflix About when in the parent film/show does it appear? Episode 6x01 - "Joan is Awful" Actual footage of the film/show can be seen (yes/no)? Yes Timestamps (interspersed with people watching it): - 12:50 - 13:39 - 14:35 - 14:47 - 15:00 - 19:22 - 20:00 - 20:50 - 21:59 - 22:17 - 25:20 - 25:40 - 30:00 - 30:40 - 35:50 - 36:20 Starring: Salma Hayek, Himesh Patel, & Jaboukie Young-White, Camirin Farmer, Ben Marnes, & Cate Blanchett Network: Streamberry (Streamberry CEO: Mona Javadi) Number of seasons: 1 Note: This episode of _Black Mirror_ is crazy meta. There's a version of _Joan is Awful_ within the show of _Joan is Awful_, and the entirity of the episode is revealed to be the real _Joan is Awful_ show. I'm just putting the primary _Joan is Awful_ show here in order to not be too confusing. ![Joan Thumbnail](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/16b534d8-b935-4595-be0f-f027ec715b3f) ![Joan Title](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/436ef050-2006-4cee-85b4-d21b5e903ab5) ![Joan1](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/bfa18db7-be73-4b0a-8006-31e6a7fd961d) ![Joan2](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/47a33c95-9c4d-421c-bf2c-41e1be55068f) ![Joan3](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7b5dfe5f-9749-45b5-992e-cd26eed91388) ![Joan4](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/c1e4c125-bf0e-4c60-a662-681eb93427ec) ![Joan5](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7370a8d2-6bb2-4314-81b2-56e6a52cc08e) ![Joan6](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0cfb1b8f-0151-4f2c-b0da-c145a3acad22) ![Joan7](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7d659805-d69b-4f69-b3e2-4905d4729fb6) ![Joan9](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2a08ea44-3506-4432-a0e1-dc9d374d0154) ![Joan11](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/941e9cc9-4596-46bc-83dc-e62516fae17c) ![Joan12](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0786abd0-8f1d-4b31-a2f2-0ed718aa5497) ![Joan13](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f2cd77c7-87b1-49c8-bd34-89dd4ee649c1) ![Joan14](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/03d45d0d-29d3-4f0c-bb42-79e8cebb0f92) ![Joan16](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/d00bc43a-203a-4acb-85be-212f687f98ef) ![Joan17](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2538cb51-63c4-4134-ae1c-757a9163b1c2) ![Joan19](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/e7ce6be7-e4fe-4570-9ea4-866444551e60) ![Joan20](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/65003790-cfb6-4c7f-9220-c4cdb9014e5f) ![Joan21](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/73551597-0385-460c-a3ff-17f470fd9ef4) ![Joan22](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/8f60e7e8-82dd-480f-9bd2-9d0164089135) ![Joan23](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/5bc36f97-0e3c-41b2-a786-220b42a92d80) ![Joan24](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/b7d56eda-35a0-47cd-864b-582c1e571b03) ![Joan25](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/fa3aa4c0-b381-402c-b509-8065ce99ba3a) ![Joan26](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/1a6a9c61-bf9d-48ab-8a57-c19e04b31ae2) ![Joan27](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/a7fd989c-4913-49c9-909b-2eaf2813be72) ![Joan28](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f88e8775-cb7b-4875-880c-c3d06c07d82e)
1.0
Add Joan is Awful from "Black Mirror" (Screenshots, Thumbnail, and Title Card added) - Please add as much of the following info as you can: Title: Joan is Awful Type (film/tv show): TV show - drama Film or show in which it appears: Black Mirror Is the parent film/show streaming anywhere? Yes - Netflix About when in the parent film/show does it appear? Episode 6x01 - "Joan is Awful" Actual footage of the film/show can be seen (yes/no)? Yes Timestamps (interspersed with people watching it): - 12:50 - 13:39 - 14:35 - 14:47 - 15:00 - 19:22 - 20:00 - 20:50 - 21:59 - 22:17 - 25:20 - 25:40 - 30:00 - 30:40 - 35:50 - 36:20 Starring: Salma Hayek, Himesh Patel, & Jaboukie Young-White, Camirin Farmer, Ben Marnes, & Cate Blanchett Network: Streamberry (Streamberry CEO: Mona Javadi) Number of seasons: 1 Note: This episode of _Black Mirror_ is crazy meta. There's a version of _Joan is Awful_ within the show of _Joan is Awful_, and the entirity of the episode is revealed to be the real _Joan is Awful_ show. I'm just putting the primary _Joan is Awful_ show here in order to not be too confusing. ![Joan Thumbnail](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/16b534d8-b935-4595-be0f-f027ec715b3f) ![Joan Title](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/436ef050-2006-4cee-85b4-d21b5e903ab5) ![Joan1](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/bfa18db7-be73-4b0a-8006-31e6a7fd961d) ![Joan2](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/47a33c95-9c4d-421c-bf2c-41e1be55068f) ![Joan3](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7b5dfe5f-9749-45b5-992e-cd26eed91388) ![Joan4](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/c1e4c125-bf0e-4c60-a662-681eb93427ec) ![Joan5](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7370a8d2-6bb2-4314-81b2-56e6a52cc08e) ![Joan6](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0cfb1b8f-0151-4f2c-b0da-c145a3acad22) ![Joan7](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/7d659805-d69b-4f69-b3e2-4905d4729fb6) ![Joan9](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2a08ea44-3506-4432-a0e1-dc9d374d0154) ![Joan11](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/941e9cc9-4596-46bc-83dc-e62516fae17c) ![Joan12](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/0786abd0-8f1d-4b31-a2f2-0ed718aa5497) ![Joan13](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f2cd77c7-87b1-49c8-bd34-89dd4ee649c1) ![Joan14](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/03d45d0d-29d3-4f0c-bb42-79e8cebb0f92) ![Joan16](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/d00bc43a-203a-4acb-85be-212f687f98ef) ![Joan17](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/2538cb51-63c4-4134-ae1c-757a9163b1c2) ![Joan19](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/e7ce6be7-e4fe-4570-9ea4-866444551e60) ![Joan20](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/65003790-cfb6-4c7f-9220-c4cdb9014e5f) ![Joan21](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/73551597-0385-460c-a3ff-17f470fd9ef4) ![Joan22](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/8f60e7e8-82dd-480f-9bd2-9d0164089135) ![Joan23](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/5bc36f97-0e3c-41b2-a786-220b42a92d80) ![Joan24](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/b7d56eda-35a0-47cd-864b-582c1e571b03) ![Joan25](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/fa3aa4c0-b381-402c-b509-8065ce99ba3a) ![Joan26](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/1a6a9c61-bf9d-48ab-8a57-c19e04b31ae2) ![Joan27](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/a7fd989c-4913-49c9-909b-2eaf2813be72) ![Joan28](https://github.com/lynnandtonic/nestflix.fun/assets/88982629/f88e8775-cb7b-4875-880c-c3d06c07d82e)
non_infrastructure
add joan is awful from black mirror screenshots thumbnail and title card added please add as much of the following info as you can title joan is awful type film tv show tv show drama film or show in which it appears black mirror is the parent film show streaming anywhere yes netflix about when in the parent film show does it appear episode joan is awful actual footage of the film show can be seen yes no yes timestamps interspersed with people watching it starring salma hayek himesh patel jaboukie young white camirin farmer ben marnes cate blanchett network streamberry streamberry ceo mona javadi number of seasons note this episode of black mirror is crazy meta there s a version of joan is awful within the show of joan is awful and the entirity of the episode is revealed to be the real joan is awful show i m just putting the primary joan is awful show here in order to not be too confusing
0
242,872
26,277,860,301
IssuesEvent
2023-01-07 01:21:05
Hieunc-NT/eShopOnContainers
https://api.github.com/repos/Hieunc-NT/eShopOnContainers
opened
microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg: 1 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</b></p></summary> <p>Targets files to enable the Visual Studio Tools for Containers.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg">https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</a></p> <p>Path to dependency file: /src/Services/Webhooks/Webhooks.API/Webhooks.API.csproj</p> <p>Path to vulnerable library: /ages/microsoft.visualstudio.azure.containers.tools.targets/1.10.8/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Hieunc-NT/eShopOnContainers/commit/58162be7965e66c71394dab67f66ed3d7cfaaef5">58162be7965e66c71394dab67f66ed3d7cfaaef5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2022-0161](https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg | Direct | Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2022-0161</summary> ### Vulnerable Library - <b>microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</b></p> <p>Targets files to enable the Visual Studio Tools for Containers.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg">https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</a></p> <p>Path to dependency file: /src/Services/Webhooks/Webhooks.API/Webhooks.API.csproj</p> <p>Path to vulnerable library: /ages/microsoft.visualstudio.azure.containers.tools.targets/1.10.8/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Hieunc-NT/eShopOnContainers/commit/58162be7965e66c71394dab67f66ed3d7cfaaef5">58162be7965e66c71394dab67f66ed3d7cfaaef5</a></p> <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Improper Handling of Exceptional Conditions in Newtonsoft.Json. Newtonsoft.Json prior to version 13.0.1 is vulnerable to Insecure Defaults due to improper handling of StackOverFlow exception (SOE) whenever nested expressions are being processed. Exploiting this vulnerability results in Denial Of Service (DoS), and it is exploitable when an attacker sends 5 requests that cause SOE in time frame of 5 minutes. This vulnerability affects Internet Information Services (IIS) Applications. <p>Publish Date: 2022-06-22 <p>URL: <a href=https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66>WS-2022-0161</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-22</p> <p>Fix Resolution: Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</b></p></summary> <p>Targets files to enable the Visual Studio Tools for Containers.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg">https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</a></p> <p>Path to dependency file: /src/Services/Webhooks/Webhooks.API/Webhooks.API.csproj</p> <p>Path to vulnerable library: /ages/microsoft.visualstudio.azure.containers.tools.targets/1.10.8/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Hieunc-NT/eShopOnContainers/commit/58162be7965e66c71394dab67f66ed3d7cfaaef5">58162be7965e66c71394dab67f66ed3d7cfaaef5</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [WS-2022-0161](https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg | Direct | Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2022-0161</summary> ### Vulnerable Library - <b>microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</b></p> <p>Targets files to enable the Visual Studio Tools for Containers.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg">https://api.nuget.org/packages/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</a></p> <p>Path to dependency file: /src/Services/Webhooks/Webhooks.API/Webhooks.API.csproj</p> <p>Path to vulnerable library: /ages/microsoft.visualstudio.azure.containers.tools.targets/1.10.8/microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.visualstudio.azure.containers.tools.targets.1.10.8.nupkg** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Hieunc-NT/eShopOnContainers/commit/58162be7965e66c71394dab67f66ed3d7cfaaef5">58162be7965e66c71394dab67f66ed3d7cfaaef5</a></p> <p>Found in base branch: <b>dev</b></p> </p> <p></p> ### Vulnerability Details <p> Improper Handling of Exceptional Conditions in Newtonsoft.Json. Newtonsoft.Json prior to version 13.0.1 is vulnerable to Insecure Defaults due to improper handling of StackOverFlow exception (SOE) whenever nested expressions are being processed. Exploiting this vulnerability results in Denial Of Service (DoS), and it is exploitable when an attacker sends 5 requests that cause SOE in time frame of 5 minutes. This vulnerability affects Internet Information Services (IIS) Applications. <p>Publish Date: 2022-06-22 <p>URL: <a href=https://github.com/JamesNK/Newtonsoft.Json/commit/7e77bbe1beccceac4fc7b174b53abfefac278b66>WS-2022-0161</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-06-22</p> <p>Fix Resolution: Newtonsoft.Json - 13.0.1;Microsoft.Extensions.ApiDescription.Server - 6.0.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_infrastructure
microsoft visualstudio azure containers tools targets nupkg vulnerabilities highest severity is vulnerable library microsoft visualstudio azure containers tools targets nupkg targets files to enable the visual studio tools for containers library home page a href path to dependency file src services webhooks webhooks api webhooks api csproj path to vulnerable library ages microsoft visualstudio azure containers tools targets microsoft visualstudio azure containers tools targets nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in microsoft visualstudio azure containers tools targets nupkg version remediation available high microsoft visualstudio azure containers tools targets nupkg direct newtonsoft json microsoft extensions apidescription server details ws vulnerable library microsoft visualstudio azure containers tools targets nupkg targets files to enable the visual studio tools for containers library home page a href path to dependency file src services webhooks webhooks api webhooks api csproj path to vulnerable library ages microsoft visualstudio azure containers tools targets microsoft visualstudio azure containers tools targets nupkg dependency hierarchy x microsoft visualstudio azure containers tools targets nupkg vulnerable library found in head commit a href found in base branch dev vulnerability details improper handling of exceptional conditions in newtonsoft json newtonsoft json prior to version is vulnerable to insecure defaults due to improper handling of stackoverflow exception soe whenever nested expressions are being processed exploiting this vulnerability results in denial of service dos and it is exploitable when an attacker sends requests that cause soe in time frame of minutes this vulnerability affects internet information services iis applications publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution newtonsoft json microsoft extensions apidescription server step up your open source security game with mend
0
16,396
11,950,969,761
IssuesEvent
2020-04-03 16:03:37
cmu-lib/dhweb_app
https://api.github.com/repos/cmu-lib/dhweb_app
opened
Abstracts list dynamic title
infrastructure
- [ ] dynamically adjust h1 element to contain the selected conference query param - [ ] display the `conference_card` for that conference
1.0
Abstracts list dynamic title - - [ ] dynamically adjust h1 element to contain the selected conference query param - [ ] display the `conference_card` for that conference
infrastructure
abstracts list dynamic title dynamically adjust element to contain the selected conference query param display the conference card for that conference
1
65,434
7,879,409,266
IssuesEvent
2018-06-26 13:20:47
nextcloud/spreed
https://api.github.com/repos/nextcloud/spreed
closed
Active users should always be sorted first, even if not moderators
1. to develop design enhancement papercut
Currently the sorting is: - Moderators (online) - Moderators (off) - Regular users (online) - Regular users (off) Instead it should be: - Moderators (online) - Regular users (online) - Moderators (off) - Regular users (off) (Just talked about with @Ivansss @schiessle )
1.0
Active users should always be sorted first, even if not moderators - Currently the sorting is: - Moderators (online) - Moderators (off) - Regular users (online) - Regular users (off) Instead it should be: - Moderators (online) - Regular users (online) - Moderators (off) - Regular users (off) (Just talked about with @Ivansss @schiessle )
non_infrastructure
active users should always be sorted first even if not moderators currently the sorting is moderators online moderators off regular users online regular users off instead it should be moderators online regular users online moderators off regular users off just talked about with ivansss schiessle
0
35,522
31,788,514,366
IssuesEvent
2023-09-13 00:02:27
praxis-isncsci/ui
https://api.github.com/repos/praxis-isncsci/ui
closed
Set code formatting rules
good first issue infrastructure
## Problem The code formatter is allowing a space curly brackets and declarations on the import statements: ```typescript import type { Meta, StoryObj } from '@storybook/web-components'; ``` We want to enforce no spaces: ```typescript import type { Meta, StoryObj } from '@storybook/web-components'; ``` We also want formatting to take place automatically on save and to enforce the use of single quotes.
1.0
Set code formatting rules - ## Problem The code formatter is allowing a space curly brackets and declarations on the import statements: ```typescript import type { Meta, StoryObj } from '@storybook/web-components'; ``` We want to enforce no spaces: ```typescript import type { Meta, StoryObj } from '@storybook/web-components'; ``` We also want formatting to take place automatically on save and to enforce the use of single quotes.
infrastructure
set code formatting rules problem the code formatter is allowing a space curly brackets and declarations on the import statements typescript import type meta storyobj from storybook web components we want to enforce no spaces typescript import type meta storyobj from storybook web components we also want formatting to take place automatically on save and to enforce the use of single quotes
1
434,054
30,441,594,366
IssuesEvent
2023-07-15 05:48:35
sinProject-Inc/talk
https://api.github.com/repos/sinProject-Inc/talk
closed
Docs: Fix English
documentation
# Tasks - [ ] # Writing the Issue - [x] Write each task for this issue using checkboxes - [x] Write all verbal instructions as tasks # Before Working on the Issue - [x] Assign yourself - [x] Share your screen on Discord - [x] Write which issue you are starting on Slack - [x] Git: Fetch the latest main - [x] Git: Create a branch with the name of the issue # Pre-PR Checks - [x] Insure that changes include only what is necessary for this PR - [x] Implement necessary Unit tests - [x] Implement necessary E2E tests - [x] Perform functionality checks
1.0
Docs: Fix English - # Tasks - [ ] # Writing the Issue - [x] Write each task for this issue using checkboxes - [x] Write all verbal instructions as tasks # Before Working on the Issue - [x] Assign yourself - [x] Share your screen on Discord - [x] Write which issue you are starting on Slack - [x] Git: Fetch the latest main - [x] Git: Create a branch with the name of the issue # Pre-PR Checks - [x] Insure that changes include only what is necessary for this PR - [x] Implement necessary Unit tests - [x] Implement necessary E2E tests - [x] Perform functionality checks
non_infrastructure
docs fix english tasks writing the issue write each task for this issue using checkboxes write all verbal instructions as tasks before working on the issue assign yourself share your screen on discord write which issue you are starting on slack git fetch the latest main git create a branch with the name of the issue pre pr checks insure that changes include only what is necessary for this pr implement necessary unit tests implement necessary tests perform functionality checks
0
18,745
13,090,978,321
IssuesEvent
2020-08-03 05:18:36
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[Linux/arm64] System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs Timeout
arch-arm64 area-Infrastructure-coreclr os-linux test-corefx
`System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs` takes more than 20 minutes ``` Running tests in /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/tests/System.Runtime.Serialization.Formatters.Tests/netcoreapp-Linux-Release-arm64 timeout --kill-after=30s 20m ./RunTests.sh /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/testhost/netcoreapp-Linux-Release-arm64 ./RunTests.sh: line 110: ulimit: core file size: cannot modify limit: Operation not permitted ----- start 11:22:35 =============== To repro directly: ===================================================== pushd . #!/usr/bin/env bash export COMPlus_TieredCompilation=0 chmod +x /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/testhost/netcoreapp-Linux-Release-arm64/dotnet /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/testhost/netcoreapp-Linux-Release-arm64/dotnet xunit.console.dll System.Runtime.Serialization.Formatters.Tests.dll -xml testResults.xml -notrait category=nonnetcoreapptests -notrait category=nonlinuxtests -notrait category=IgnoreForCI -notrait category=OuterLoop -notrait category=failing popd =========================================================================================================== ~/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/tests/System.Runtime.Serialization.Formatters.Tests/netcoreapp-Linux-Release-arm64 ~/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/tests/System.Runtime.Serialization.Formatters.Tests/netcoreapp-Linux-Release-arm64 xUnit.net Console Runner v2.4.1-pre.build.4059 (64-bit .NET Core 4.6.27002.0) Discovering: System.Runtime.Serialization.Formatters.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Runtime.Serialization.Formatters.Tests (found 67 test cases) Starting: System.Runtime.Serialization.Formatters.Tests (parallel test collections = on, max threads = 46) System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:02:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:04:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:06:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:08:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:10:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:12:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:14:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:16:10 error: One or more tests failed while running tests from ''. Exit code 124. ```
1.0
[Linux/arm64] System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs Timeout - `System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs` takes more than 20 minutes ``` Running tests in /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/tests/System.Runtime.Serialization.Formatters.Tests/netcoreapp-Linux-Release-arm64 timeout --kill-after=30s 20m ./RunTests.sh /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/testhost/netcoreapp-Linux-Release-arm64 ./RunTests.sh: line 110: ulimit: core file size: cannot modify limit: Operation not permitted ----- start 11:22:35 =============== To repro directly: ===================================================== pushd . #!/usr/bin/env bash export COMPlus_TieredCompilation=0 chmod +x /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/testhost/netcoreapp-Linux-Release-arm64/dotnet /home/robox/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/testhost/netcoreapp-Linux-Release-arm64/dotnet xunit.console.dll System.Runtime.Serialization.Formatters.Tests.dll -xml testResults.xml -notrait category=nonnetcoreapptests -notrait category=nonlinuxtests -notrait category=IgnoreForCI -notrait category=OuterLoop -notrait category=failing popd =========================================================================================================== ~/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/tests/System.Runtime.Serialization.Formatters.Tests/netcoreapp-Linux-Release-arm64 ~/j/workspace/dotnet_coreclr/master/jitstress/arm64_cross_checked_ubuntu16.04_corefx_baseline_tst_prtest/_/fx/bin/tests/System.Runtime.Serialization.Formatters.Tests/netcoreapp-Linux-Release-arm64 xUnit.net Console Runner v2.4.1-pre.build.4059 (64-bit .NET Core 4.6.27002.0) Discovering: System.Runtime.Serialization.Formatters.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Runtime.Serialization.Formatters.Tests (found 67 test cases) Starting: System.Runtime.Serialization.Formatters.Tests (parallel test collections = on, max threads = 46) System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:02:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:04:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:06:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:08:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:10:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:12:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:14:10 System.Runtime.Serialization.Formatters.Tests: [Long Running Test] 'System.Runtime.Serialization.Formatters.Tests.BinaryFormatterTests.SerializeHugeObjectGraphs', Elapsed: 00:16:10 error: One or more tests failed while running tests from ''. Exit code 124. ```
infrastructure
system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs timeout system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs takes more than minutes running tests in home robox j workspace dotnet coreclr master jitstress cross checked corefx baseline tst prtest fx bin tests system runtime serialization formatters tests netcoreapp linux release timeout kill after runtests sh home robox j workspace dotnet coreclr master jitstress cross checked corefx baseline tst prtest fx bin testhost netcoreapp linux release runtests sh line ulimit core file size cannot modify limit operation not permitted start to repro directly pushd usr bin env bash export complus tieredcompilation chmod x home robox j workspace dotnet coreclr master jitstress cross checked corefx baseline tst prtest fx bin testhost netcoreapp linux release dotnet home robox j workspace dotnet coreclr master jitstress cross checked corefx baseline tst prtest fx bin testhost netcoreapp linux release dotnet xunit console dll system runtime serialization formatters tests dll xml testresults xml notrait category nonnetcoreapptests notrait category nonlinuxtests notrait category ignoreforci notrait category outerloop notrait category failing popd j workspace dotnet coreclr master jitstress cross checked corefx baseline tst prtest fx bin tests system runtime serialization formatters tests netcoreapp linux release j workspace dotnet coreclr master jitstress cross checked corefx baseline tst prtest fx bin tests system runtime serialization formatters tests netcoreapp linux release xunit net console runner pre build bit net core discovering system runtime serialization formatters tests method display classandmethod method display options none discovered system runtime serialization formatters tests found test cases starting system runtime serialization formatters tests parallel test collections on max threads system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed system runtime serialization formatters tests system runtime serialization formatters tests binaryformattertests serializehugeobjectgraphs elapsed error one or more tests failed while running tests from exit code
1
7,089
6,760,144,292
IssuesEvent
2017-10-24 19:34:11
postmarketOS/pmbootstrap
https://api.github.com/repos/postmarketOS/pmbootstrap
closed
Use coveralls.io to display testing coverage?
discussion infrastructure
PR #760 will integrate automatic testing on Travis CI, at least for the fast testcases. I would like to display how much testing coverage we have for pmbootstrap in the README.md (with a badge and with details when you click on it). **Proprietary** service [coveralls.io](https://coveralls.io/) looks like a nice interface and is free to use for open source projects. It is compatible to GitHub and gitlab [in case we decide to move there](https://github.com/postmarketOS/postmarketos.org/issues/37). I have not found a similar service, where the source code is available (is there any?). Worst case is, that for some reason we suddenly can't use that service anymore and lose the coverage reports from the past. We can always generate coverage reports ourselves with [pytest-cov](https://github.com/pytest-dev/pytest-cov) (which is the output that gets used for coveralls.) And of course the upside is, that this motivates (at least me) to make more testcases and have good testing coverage. So what do you think about this proposal, fellow hackers?
1.0
Use coveralls.io to display testing coverage? - PR #760 will integrate automatic testing on Travis CI, at least for the fast testcases. I would like to display how much testing coverage we have for pmbootstrap in the README.md (with a badge and with details when you click on it). **Proprietary** service [coveralls.io](https://coveralls.io/) looks like a nice interface and is free to use for open source projects. It is compatible to GitHub and gitlab [in case we decide to move there](https://github.com/postmarketOS/postmarketos.org/issues/37). I have not found a similar service, where the source code is available (is there any?). Worst case is, that for some reason we suddenly can't use that service anymore and lose the coverage reports from the past. We can always generate coverage reports ourselves with [pytest-cov](https://github.com/pytest-dev/pytest-cov) (which is the output that gets used for coveralls.) And of course the upside is, that this motivates (at least me) to make more testcases and have good testing coverage. So what do you think about this proposal, fellow hackers?
infrastructure
use coveralls io to display testing coverage pr will integrate automatic testing on travis ci at least for the fast testcases i would like to display how much testing coverage we have for pmbootstrap in the readme md with a badge and with details when you click on it proprietary service looks like a nice interface and is free to use for open source projects it is compatible to github and gitlab i have not found a similar service where the source code is available is there any worst case is that for some reason we suddenly can t use that service anymore and lose the coverage reports from the past we can always generate coverage reports ourselves with which is the output that gets used for coveralls and of course the upside is that this motivates at least me to make more testcases and have good testing coverage so what do you think about this proposal fellow hackers
1
280,870
21,315,239,589
IssuesEvent
2022-04-16 06:43:19
AdrianOngJJ/pe
https://api.github.com/repos/AdrianOngJJ/pe
opened
Adding category with similar names
severity.Low type.DocumentationBug
Current Cat list: ![image.png](https://raw.githubusercontent.com/AdrianOngJJ/pe/main/files/7cc7202e-8409-4913-a927-e6f42e1830c3.png) I want to add: "bENs", assuming is a different establishment Outcome: ![image.png](https://raw.githubusercontent.com/AdrianOngJJ/pe/main/files/37a609a4-bd92-472e-9a41-fd1f64fcdd82.png) Not stated in UG about cat with same letters Suggestion: - Add caution in UG to let users know about the case-insensitivity - Use different logic to verify, like `.equals()` <!--session: 1650087851696-1214a6ce-1e0b-4a13-a847-d2b7879ac2ef--> <!--Version: Web v3.4.2-->
1.0
Adding category with similar names - Current Cat list: ![image.png](https://raw.githubusercontent.com/AdrianOngJJ/pe/main/files/7cc7202e-8409-4913-a927-e6f42e1830c3.png) I want to add: "bENs", assuming is a different establishment Outcome: ![image.png](https://raw.githubusercontent.com/AdrianOngJJ/pe/main/files/37a609a4-bd92-472e-9a41-fd1f64fcdd82.png) Not stated in UG about cat with same letters Suggestion: - Add caution in UG to let users know about the case-insensitivity - Use different logic to verify, like `.equals()` <!--session: 1650087851696-1214a6ce-1e0b-4a13-a847-d2b7879ac2ef--> <!--Version: Web v3.4.2-->
non_infrastructure
adding category with similar names current cat list i want to add bens assuming is a different establishment outcome not stated in ug about cat with same letters suggestion add caution in ug to let users know about the case insensitivity use different logic to verify like equals
0
14,625
11,015,385,994
IssuesEvent
2019-12-05 01:25:55
18F/jekyll-get
https://api.github.com/repos/18F/jekyll-get
closed
make this a gem
epic: infrastructure management help wanted
Makes it easier for projects that use this to keep up-to-date. Happy to help if needed.
1.0
make this a gem - Makes it easier for projects that use this to keep up-to-date. Happy to help if needed.
infrastructure
make this a gem makes it easier for projects that use this to keep up to date happy to help if needed
1
32,170
26,491,070,461
IssuesEvent
2023-01-17 22:44:17
ZcashFoundation/zebra
https://api.github.com/repos/ZcashFoundation/zebra
closed
Automate pinning CI dependencies to specific commit hashes
A-dependencies A-infrastructure A-devops S-needs-triage C-security P-Optional :sparkles:
## Motivation <!-- Is your feature request related to a problem? How does this change improve Zebra? --> It'd be nice to be able to manage our CI actions and other dependencies like we do with our cargo etc deps. The newish https://github.com/sethvargo/ratchet may help us do that.
1.0
Automate pinning CI dependencies to specific commit hashes - ## Motivation <!-- Is your feature request related to a problem? How does this change improve Zebra? --> It'd be nice to be able to manage our CI actions and other dependencies like we do with our cargo etc deps. The newish https://github.com/sethvargo/ratchet may help us do that.
infrastructure
automate pinning ci dependencies to specific commit hashes motivation is your feature request related to a problem how does this change improve zebra it d be nice to be able to manage our ci actions and other dependencies like we do with our cargo etc deps the newish may help us do that
1
8,340
3,715,448,082
IssuesEvent
2016-03-03 01:47:24
PolymerElements/iron-fit-behavior
https://api.github.com/repos/PolymerElements/iron-fit-behavior
closed
`position: fixed` problem with `<iron-dropdown>`
bug code-yellow p0 - critical started
_From @David-Mulder on September 23, 2015 18:50_ In certain cases the menu will be `position: fixed` relative to other positioned elements (however this doesn't always happen for reasons that are beyond me) thus causing the menu to be misplaced/not visible at all. I have tried my absolute best to make a minimal reproducible case with no luck, however I have been able to create a set of instructions that reproduce the problem: 1. Install the Polymer Starter Kit 2. Add `<link rel="import" href="../bower_components/paper-menu-button/paper-menu-button.html">` to `elements.html` 3. Replace `#drawerToolbar` in `index.html` with: ``` <paper-toolbar id="drawerToolbar"> <span class="paper-font-title flex">Menu</span> <paper-menu-button horizontal-align="right"> <paper-icon-button icon="subject" class="dropdown-trigger"></paper-icon-button> <div class="dropdown-content">ASDF</div> </paper-menu-button> </paper-toolbar> ``` 4. To see the menu you can make the window narrow enough to be the exact width of the drawer panel. In both Chrome, Edge and Firefox the menu will not show due to incorrect positioning both with shady DOM and shadow DOM. _Copied from original issue: PolymerElements/paper-menu-button#29_
1.0
`position: fixed` problem with `<iron-dropdown>` - _From @David-Mulder on September 23, 2015 18:50_ In certain cases the menu will be `position: fixed` relative to other positioned elements (however this doesn't always happen for reasons that are beyond me) thus causing the menu to be misplaced/not visible at all. I have tried my absolute best to make a minimal reproducible case with no luck, however I have been able to create a set of instructions that reproduce the problem: 1. Install the Polymer Starter Kit 2. Add `<link rel="import" href="../bower_components/paper-menu-button/paper-menu-button.html">` to `elements.html` 3. Replace `#drawerToolbar` in `index.html` with: ``` <paper-toolbar id="drawerToolbar"> <span class="paper-font-title flex">Menu</span> <paper-menu-button horizontal-align="right"> <paper-icon-button icon="subject" class="dropdown-trigger"></paper-icon-button> <div class="dropdown-content">ASDF</div> </paper-menu-button> </paper-toolbar> ``` 4. To see the menu you can make the window narrow enough to be the exact width of the drawer panel. In both Chrome, Edge and Firefox the menu will not show due to incorrect positioning both with shady DOM and shadow DOM. _Copied from original issue: PolymerElements/paper-menu-button#29_
non_infrastructure
position fixed problem with from david mulder on september in certain cases the menu will be position fixed relative to other positioned elements however this doesn t always happen for reasons that are beyond me thus causing the menu to be misplaced not visible at all i have tried my absolute best to make a minimal reproducible case with no luck however i have been able to create a set of instructions that reproduce the problem install the polymer starter kit add to elements html replace drawertoolbar in index html with menu asdf to see the menu you can make the window narrow enough to be the exact width of the drawer panel in both chrome edge and firefox the menu will not show due to incorrect positioning both with shady dom and shadow dom copied from original issue polymerelements paper menu button
0
18,100
12,759,497,648
IssuesEvent
2020-06-29 06:00:31
BlockResearchGroup/compas-RV2
https://api.github.com/repos/BlockResearchGroup/compas-RV2
closed
update setup/installation instructions
Priority: High Type: Infrastructure
under "Setup" [https://app.gitbook.com/invite/blockresearchgroup?invite=-M8uDcBq1Knd3axqdkAC](url) - [x] workshop requirements - [x] updated installation instructions - [x] FAQ (from past, known common problems)
1.0
update setup/installation instructions - under "Setup" [https://app.gitbook.com/invite/blockresearchgroup?invite=-M8uDcBq1Knd3axqdkAC](url) - [x] workshop requirements - [x] updated installation instructions - [x] FAQ (from past, known common problems)
infrastructure
update setup installation instructions under setup url workshop requirements updated installation instructions faq from past known common problems
1
7,104
6,774,036,049
IssuesEvent
2017-10-27 08:52:00
bpmn-io/bpmn-js
https://api.github.com/repos/bpmn-io/bpmn-js
opened
Automate releasing of seed app and bower bundle
infrastructure release
Right now, we spend a lot of time updating the [bundled bpmn-js](https://github.com/bpmn-io/bower-bpmn-js) as well as the [bpmn-js seed project](https://github.com/bpmn-io/bpmn-js-seed) on every library release. __Tasks__ * [ ] Automatically re-build + release bower bundle * [ ] Automatically re-build + release bpmn-js-seed
1.0
Automate releasing of seed app and bower bundle - Right now, we spend a lot of time updating the [bundled bpmn-js](https://github.com/bpmn-io/bower-bpmn-js) as well as the [bpmn-js seed project](https://github.com/bpmn-io/bpmn-js-seed) on every library release. __Tasks__ * [ ] Automatically re-build + release bower bundle * [ ] Automatically re-build + release bpmn-js-seed
infrastructure
automate releasing of seed app and bower bundle right now we spend a lot of time updating the as well as the on every library release tasks automatically re build release bower bundle automatically re build release bpmn js seed
1
569,757
17,016,051,603
IssuesEvent
2021-07-02 12:12:56
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
Link to userpage from user's traces page
Component: website Priority: trivial Type: enhancement
**[Submitted to the original trac issue database at 7.49pm, Monday, 13th August 2012]** Username on user's traces page and user's diary comments page should be a link directing back to userpage. Or if a link in page heading is not desired then an extra line for link to userpage like on user's changesets page.
1.0
Link to userpage from user's traces page - **[Submitted to the original trac issue database at 7.49pm, Monday, 13th August 2012]** Username on user's traces page and user's diary comments page should be a link directing back to userpage. Or if a link in page heading is not desired then an extra line for link to userpage like on user's changesets page.
non_infrastructure
link to userpage from user s traces page username on user s traces page and user s diary comments page should be a link directing back to userpage or if a link in page heading is not desired then an extra line for link to userpage like on user s changesets page
0
26,079
19,649,148,728
IssuesEvent
2022-01-10 03:21:06
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Not easy to add new row to grid in SOBOL
bug interface/infrastructure
The SOBOL presenter has a grid containing the parameter ranges. To add a new row to the table, you presumably click in the empty row at the bottom and start typing. However, you cannot edit the bottom row unless you first click in an existing row, and then, click on the bottom row to start typing.
1.0
Not easy to add new row to grid in SOBOL - The SOBOL presenter has a grid containing the parameter ranges. To add a new row to the table, you presumably click in the empty row at the bottom and start typing. However, you cannot edit the bottom row unless you first click in an existing row, and then, click on the bottom row to start typing.
infrastructure
not easy to add new row to grid in sobol the sobol presenter has a grid containing the parameter ranges to add a new row to the table you presumably click in the empty row at the bottom and start typing however you cannot edit the bottom row unless you first click in an existing row and then click on the bottom row to start typing
1
265,315
23,159,732,090
IssuesEvent
2022-07-29 16:21:06
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
reopened
kubeconfig-default-ttl-minutes setting should be made to accept only integer values
[zube]: To Test kind/bug-qa team/area1
Setup Rancher version: 2.6-head `78017e6` Browser type & version: Chrome **Describe the bug** Currently we can set the value for `kubeconfig-default-ttl-minutes` setting in the UI to any non-integer values. By accidentally setting the value to a non-integer value, when we try to download the kubeconfig button, the file will not be downloaded and fails with an error **To Reproduce** 1. On rancher v2.6-head, navigate to global settings >> Settings >> set `kubeconfig-token-ttl-minutes` to `2/` 2. Navigate to local explorer, click on download kubeconfig button **Result** Download fails with the error `message: "failed to get default token TTL: failed to parse setting 'kubeconfig-default-ttl-minutes': error parsing token ttl: time: unknown unit \";m\" in duration \"2;m\"" ` **Expected Result** UI not allowing to save non integer values. Related: https://github.com/rancher/dashboard/issues/6120
1.0
kubeconfig-default-ttl-minutes setting should be made to accept only integer values - Setup Rancher version: 2.6-head `78017e6` Browser type & version: Chrome **Describe the bug** Currently we can set the value for `kubeconfig-default-ttl-minutes` setting in the UI to any non-integer values. By accidentally setting the value to a non-integer value, when we try to download the kubeconfig button, the file will not be downloaded and fails with an error **To Reproduce** 1. On rancher v2.6-head, navigate to global settings >> Settings >> set `kubeconfig-token-ttl-minutes` to `2/` 2. Navigate to local explorer, click on download kubeconfig button **Result** Download fails with the error `message: "failed to get default token TTL: failed to parse setting 'kubeconfig-default-ttl-minutes': error parsing token ttl: time: unknown unit \";m\" in duration \"2;m\"" ` **Expected Result** UI not allowing to save non integer values. Related: https://github.com/rancher/dashboard/issues/6120
non_infrastructure
kubeconfig default ttl minutes setting should be made to accept only integer values setup rancher version head browser type version chrome describe the bug currently we can set the value for kubeconfig default ttl minutes setting in the ui to any non integer values by accidentally setting the value to a non integer value when we try to download the kubeconfig button the file will not be downloaded and fails with an error to reproduce on rancher head navigate to global settings settings set kubeconfig token ttl minutes to navigate to local explorer click on download kubeconfig button result download fails with the error message failed to get default token ttl failed to parse setting kubeconfig default ttl minutes error parsing token ttl time unknown unit m in duration m expected result ui not allowing to save non integer values related
0
194,339
22,261,958,872
IssuesEvent
2022-06-10 01:54:26
nanopathi/linux-4.19.72_CVE-2021-32399
https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2021-32399
reopened
CVE-2020-35508 (Medium) detected in linuxlinux-4.19.236
security vulnerability
## CVE-2020-35508 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-32399/commit/03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66">03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/fork.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw possibility of race condition and incorrect initialization of the process id was found in the Linux kernel child/parent process identification handling while filtering signal handlers. A local attacker is able to abuse this flaw to bypass checks to send any signal to a privileged process. <p>Publish Date: 2021-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35508>CVE-2020-35508</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-35508">https://www.linuxkernelcves.com/cves/CVE-2020-35508</a></p> <p>Release Date: 2021-03-26</p> <p>Fix Resolution: v4.4.242, v4.9.242, v4.14.205, v4.19.156, v5.4.76, v5.9.7, v5.10-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-35508 (Medium) detected in linuxlinux-4.19.236 - ## CVE-2020-35508 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-32399/commit/03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66">03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/fork.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw possibility of race condition and incorrect initialization of the process id was found in the Linux kernel child/parent process identification handling while filtering signal handlers. A local attacker is able to abuse this flaw to bypass checks to send any signal to a privileged process. <p>Publish Date: 2021-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-35508>CVE-2020-35508</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-35508">https://www.linuxkernelcves.com/cves/CVE-2020-35508</a></p> <p>Release Date: 2021-03-26</p> <p>Fix Resolution: v4.4.242, v4.9.242, v4.14.205, v4.19.156, v5.4.76, v5.9.7, v5.10-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files kernel fork c vulnerability details a flaw possibility of race condition and incorrect initialization of the process id was found in the linux kernel child parent process identification handling while filtering signal handlers a local attacker is able to abuse this flaw to bypass checks to send any signal to a privileged process publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
258,133
22,285,505,720
IssuesEvent
2022-06-11 15:02:24
Nedilko/weekends
https://api.github.com/repos/Nedilko/weekends
closed
[Test case]: Verify that application can use system theme
test case
### Description This TC verifies that if user choose "use system theme" function then application theme adjusts to theme which set in system. Preconditions: - Not first application load - Chosen system theme is light - Chosen application theme is light - "Use system theme" in settings modal set into left position (off) ### Steps to reproduce 1. Go to https://weekends.nac.in.ua/ 2. Click on 'Gear' (Settings) button in the top left corner. 3. Click on 'Use system theme' switcher. 4. Click on 'Apply' button. 5. Move 'Use system theme' switcher to position 'off'. 6. Click on 'Apply' button. 7. Change system theme to dark. 8. Reload page. 9. Move 'Use system theme' switcher to position 'on'. 10. Click on 'Apply' button. ### Expected result 1. Main page is opened. Application theme is set to light. Toggle position is left. 2. Settings modal appears. Theme switcher in position left. 3. Toggle switcher changes position from off (left) to on (right) 4. Theme adjusts to theme which set in system (light). Theme toggle (one in the top right corner) disappears. 5. Theme switcher in position left 6. Application theme is set to light. Theme toggle (one in the top right corner) appears. Toggle position is left. 7. System theme is changed to dark 8. Application theme is set to light. 9. Toggle switcher changes position from off (left) to on (right) 10. Theme adjusts to theme which set in system (dark). Theme toggle (one in the top right corner) disappears.
1.0
[Test case]: Verify that application can use system theme - ### Description This TC verifies that if user choose "use system theme" function then application theme adjusts to theme which set in system. Preconditions: - Not first application load - Chosen system theme is light - Chosen application theme is light - "Use system theme" in settings modal set into left position (off) ### Steps to reproduce 1. Go to https://weekends.nac.in.ua/ 2. Click on 'Gear' (Settings) button in the top left corner. 3. Click on 'Use system theme' switcher. 4. Click on 'Apply' button. 5. Move 'Use system theme' switcher to position 'off'. 6. Click on 'Apply' button. 7. Change system theme to dark. 8. Reload page. 9. Move 'Use system theme' switcher to position 'on'. 10. Click on 'Apply' button. ### Expected result 1. Main page is opened. Application theme is set to light. Toggle position is left. 2. Settings modal appears. Theme switcher in position left. 3. Toggle switcher changes position from off (left) to on (right) 4. Theme adjusts to theme which set in system (light). Theme toggle (one in the top right corner) disappears. 5. Theme switcher in position left 6. Application theme is set to light. Theme toggle (one in the top right corner) appears. Toggle position is left. 7. System theme is changed to dark 8. Application theme is set to light. 9. Toggle switcher changes position from off (left) to on (right) 10. Theme adjusts to theme which set in system (dark). Theme toggle (one in the top right corner) disappears.
non_infrastructure
verify that application can use system theme description this tc verifies that if user choose use system theme function then application theme adjusts to theme which set in system preconditions not first application load chosen system theme is light chosen application theme is light use system theme in settings modal set into left position off steps to reproduce go to click on gear settings button in the top left corner click on use system theme switcher click on apply button move use system theme switcher to position off click on apply button change system theme to dark reload page move use system theme switcher to position on click on apply button expected result main page is opened application theme is set to light toggle position is left settings modal appears theme switcher in position left toggle switcher changes position from off left to on right theme adjusts to theme which set in system light theme toggle one in the top right corner disappears theme switcher in position left application theme is set to light theme toggle one in the top right corner appears toggle position is left system theme is changed to dark application theme is set to light toggle switcher changes position from off left to on right theme adjusts to theme which set in system dark theme toggle one in the top right corner disappears
0
16,424
11,958,021,664
IssuesEvent
2020-04-04 16:28:24
Xilinx/brevitas
https://api.github.com/repos/Xilinx/brevitas
closed
Template tests.yaml
infrastructure tests
Current tests.yaml should be splitted into multiples actions, to test e.g. develop installation separately from the test suite, with separate matrix parametrization. Given that a lot of the logic around cache creation would be shared, a python driven flow for templating it should be setup. noxfile.py has to be able to read the template in order to match the matrix configuration that would be generate for different scenarios.
1.0
Template tests.yaml - Current tests.yaml should be splitted into multiples actions, to test e.g. develop installation separately from the test suite, with separate matrix parametrization. Given that a lot of the logic around cache creation would be shared, a python driven flow for templating it should be setup. noxfile.py has to be able to read the template in order to match the matrix configuration that would be generate for different scenarios.
infrastructure
template tests yaml current tests yaml should be splitted into multiples actions to test e g develop installation separately from the test suite with separate matrix parametrization given that a lot of the logic around cache creation would be shared a python driven flow for templating it should be setup noxfile py has to be able to read the template in order to match the matrix configuration that would be generate for different scenarios
1
27,129
21,190,983,279
IssuesEvent
2022-04-08 17:22:58
bootstrapworld/curriculum
https://api.github.com/repos/bootstrapworld/curriculum
closed
@data-cycle directive
Infrastructure
The Data Cycle has been standardized, and it's already used in more than a dozen places across the DS pathway. It will likely be used more and more, so it would be nice to create a directive and separate the asciidoc from the data. Shown below is an example of what this directive might look like as it's passed as input to the preprocessor, and what output we'd like it to generate. I've used all-caps to show where these values are placed. **Input:** ``` @data-cycle{ <INSTRUCTIONS> #:question <QUESTION> #:show-question? #t #:rows <ROWS> #:show-rows? #t #:cols <COLS> #:show-cols? #t #:filter-fn <FILTER-FN> #:show-filter? #t #:build-fn <BUILD-FN> #:show-build? #t #:expression <EXPRESSION> #:show-expression? #t #:finding <FINDING> #:show-finding? #t #:new-question <NEW-QUESTION> #:show-new-question? #t } ``` **Output:** (Note: the blank newlines are _not_ extraneous!) ``` [.data-cycle-directions] <DIRECTIONS> [.data-cycle, cols="^.^3, .^20", stripes="none"] |=== | @image{../images/AskQuestions.png, "Ask Questions icon"} | @fitbruby{}{<QUESTION>}{What question do you have?} @fitb{}{} What type of question is this? (circle one) @hspace{1.5em} _ookup @hspace{1.5em} Arithmetic @hspace{1.5em} Statistical | @image{../images/ConsiderData.png, "Consider Data icon"} | @fitbruby{}{<ROWS>}{What Rows? (Just the cats, fixed rabbits, etc.)} @fitbruby{}{<COLS>}{Which Column(s)? (age, weight-in-kilograms, weeks, etc.)} | @image{../images/AnalyzeData.png, "Analyze icon"} | @fitbruby{}{<FILTER-FN>}{If you only need *_some_* rows, define your filter function here (Need help? Use the Design Recipe!)} @fitbruby{}{<BUILD-FN>}{If you need to make a *_new_* column, define your builder function here (Need help? Use the Design Recipe!)} @fitbruby{}{<EXPRESSION>}{Write the code that will make the table or display you want.} | @image{../images/InterpretData.png, "Interpret icon"} | @fitbruby{}{<FINDING>}{What did you find out? What can you infer?} @fitb{}{} @fitbruby{}{<NEW-QUESTION>}{What new question(s) does this raise?} @fitb{}{} |=== ```
1.0
@data-cycle directive - The Data Cycle has been standardized, and it's already used in more than a dozen places across the DS pathway. It will likely be used more and more, so it would be nice to create a directive and separate the asciidoc from the data. Shown below is an example of what this directive might look like as it's passed as input to the preprocessor, and what output we'd like it to generate. I've used all-caps to show where these values are placed. **Input:** ``` @data-cycle{ <INSTRUCTIONS> #:question <QUESTION> #:show-question? #t #:rows <ROWS> #:show-rows? #t #:cols <COLS> #:show-cols? #t #:filter-fn <FILTER-FN> #:show-filter? #t #:build-fn <BUILD-FN> #:show-build? #t #:expression <EXPRESSION> #:show-expression? #t #:finding <FINDING> #:show-finding? #t #:new-question <NEW-QUESTION> #:show-new-question? #t } ``` **Output:** (Note: the blank newlines are _not_ extraneous!) ``` [.data-cycle-directions] <DIRECTIONS> [.data-cycle, cols="^.^3, .^20", stripes="none"] |=== | @image{../images/AskQuestions.png, "Ask Questions icon"} | @fitbruby{}{<QUESTION>}{What question do you have?} @fitb{}{} What type of question is this? (circle one) @hspace{1.5em} _ookup @hspace{1.5em} Arithmetic @hspace{1.5em} Statistical | @image{../images/ConsiderData.png, "Consider Data icon"} | @fitbruby{}{<ROWS>}{What Rows? (Just the cats, fixed rabbits, etc.)} @fitbruby{}{<COLS>}{Which Column(s)? (age, weight-in-kilograms, weeks, etc.)} | @image{../images/AnalyzeData.png, "Analyze icon"} | @fitbruby{}{<FILTER-FN>}{If you only need *_some_* rows, define your filter function here (Need help? Use the Design Recipe!)} @fitbruby{}{<BUILD-FN>}{If you need to make a *_new_* column, define your builder function here (Need help? Use the Design Recipe!)} @fitbruby{}{<EXPRESSION>}{Write the code that will make the table or display you want.} | @image{../images/InterpretData.png, "Interpret icon"} | @fitbruby{}{<FINDING>}{What did you find out? What can you infer?} @fitb{}{} @fitbruby{}{<NEW-QUESTION>}{What new question(s) does this raise?} @fitb{}{} |=== ```
infrastructure
data cycle directive the data cycle has been standardized and it s already used in more than a dozen places across the ds pathway it will likely be used more and more so it would be nice to create a directive and separate the asciidoc from the data shown below is an example of what this directive might look like as it s passed as input to the preprocessor and what output we d like it to generate i ve used all caps to show where these values are placed input data cycle question show question t rows show rows t cols show cols t filter fn show filter t build fn show build t expression show expression t finding show finding t new question show new question t output note the blank newlines are not extraneous image images askquestions png ask questions icon fitbruby what question do you have fitb what type of question is this circle one hspace ookup hspace arithmetic hspace statistical image images considerdata png consider data icon fitbruby what rows just the cats fixed rabbits etc fitbruby which column s age weight in kilograms weeks etc image images analyzedata png analyze icon fitbruby if you only need some rows define your filter function here need help use the design recipe fitbruby if you need to make a new column define your builder function here need help use the design recipe fitbruby write the code that will make the table or display you want image images interpretdata png interpret icon fitbruby what did you find out what can you infer fitb fitbruby what new question s does this raise fitb
1
21,185
14,432,318,796
IssuesEvent
2020-12-07 01:33:51
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Facade assemblies contain dead SR strings
area-Infrastructure-libraries untriaged
Eg., System.CodeDom.dll is a pure facade. It should be ~12KB but we build into it SR.cs which adds another 4KB. If GenFacades ends up with an empty assembly, we should not include SR.cs.
1.0
Facade assemblies contain dead SR strings - Eg., System.CodeDom.dll is a pure facade. It should be ~12KB but we build into it SR.cs which adds another 4KB. If GenFacades ends up with an empty assembly, we should not include SR.cs.
infrastructure
facade assemblies contain dead sr strings eg system codedom dll is a pure facade it should be but we build into it sr cs which adds another if genfacades ends up with an empty assembly we should not include sr cs
1
259,826
19,634,314,627
IssuesEvent
2022-01-08 02:15:41
kubernetes/sig-release
https://api.github.com/repos/kubernetes/sig-release
reopened
Create Visual Path To Release
priority/important-soon sig/release kind/documentation lifecycle/rotten area/release-team
During the last SIG-Release meeting, we discussed awareness of the release outside of the Release Team (i.e. contributors/kep authors/etc) and the need to better communicate this and provide refresher training to the community, as well as the need to better highlight things with the release team. One proposal was developing a "visual path" to a release, with all steps and things related documented. This could be useful for: * Onboarding for RT members * Onboarding for general new contributors * Re-onboarding of veterans who might assume what the processes are now but don’t have new info and context This should include things like education about dates (enhancement freeze, code freeze, test freeze), phases of the release, and information about all artifacts generated (release notes, blogs, etc), so that everyone can be familiar with the what, when, and why. This could initially be picked up by the various shadows toward the end of 1.21 (or shadows from 1.20 could contribute what they learned in 1.20). Then, we could have on-going update. periodic updates each release, as things evolve, to ensure that visuals and text do not become stale. Once developed, the release team can communicate this and evangelize this to various SIG meetings and during the monthly community meetings. /sig release /kind documentation /area release-team /priority important-soon /cc @kubernetes/sig-release-leads /milestone v1.21
1.0
Create Visual Path To Release - During the last SIG-Release meeting, we discussed awareness of the release outside of the Release Team (i.e. contributors/kep authors/etc) and the need to better communicate this and provide refresher training to the community, as well as the need to better highlight things with the release team. One proposal was developing a "visual path" to a release, with all steps and things related documented. This could be useful for: * Onboarding for RT members * Onboarding for general new contributors * Re-onboarding of veterans who might assume what the processes are now but don’t have new info and context This should include things like education about dates (enhancement freeze, code freeze, test freeze), phases of the release, and information about all artifacts generated (release notes, blogs, etc), so that everyone can be familiar with the what, when, and why. This could initially be picked up by the various shadows toward the end of 1.21 (or shadows from 1.20 could contribute what they learned in 1.20). Then, we could have on-going update. periodic updates each release, as things evolve, to ensure that visuals and text do not become stale. Once developed, the release team can communicate this and evangelize this to various SIG meetings and during the monthly community meetings. /sig release /kind documentation /area release-team /priority important-soon /cc @kubernetes/sig-release-leads /milestone v1.21
non_infrastructure
create visual path to release during the last sig release meeting we discussed awareness of the release outside of the release team i e contributors kep authors etc and the need to better communicate this and provide refresher training to the community as well as the need to better highlight things with the release team one proposal was developing a visual path to a release with all steps and things related documented this could be useful for onboarding for rt members onboarding for general new contributors re onboarding of veterans who might assume what the processes are now but don’t have new info and context this should include things like education about dates enhancement freeze code freeze test freeze phases of the release and information about all artifacts generated release notes blogs etc so that everyone can be familiar with the what when and why this could initially be picked up by the various shadows toward the end of or shadows from could contribute what they learned in then we could have on going update periodic updates each release as things evolve to ensure that visuals and text do not become stale once developed the release team can communicate this and evangelize this to various sig meetings and during the monthly community meetings sig release kind documentation area release team priority important soon cc kubernetes sig release leads milestone
0
765,018
26,828,598,521
IssuesEvent
2023-02-02 14:34:29
NIAEFEUP/tts-revamp-fe
https://api.github.com/repos/NIAEFEUP/tts-revamp-fe
closed
Professor's filtering
low priority medium effort
For each subject, add a dropdown to filter the professors. Mocks: ![image](https://user-images.githubusercontent.com/63104941/203020842-3474666a-1e17-408f-b9a1-6982e7aa03d7.png) index: 1 - Dropdown with the classes 2 - Multi Select dropdown with the Professors
1.0
Professor's filtering - For each subject, add a dropdown to filter the professors. Mocks: ![image](https://user-images.githubusercontent.com/63104941/203020842-3474666a-1e17-408f-b9a1-6982e7aa03d7.png) index: 1 - Dropdown with the classes 2 - Multi Select dropdown with the Professors
non_infrastructure
professor s filtering for each subject add a dropdown to filter the professors mocks index dropdown with the classes multi select dropdown with the professors
0
111,534
14,106,540,368
IssuesEvent
2020-11-06 15:04:09
jphacks/C_2006
https://api.github.com/repos/jphacks/C_2006
closed
Search画面のUI変更
Design must
## 追加する機能 <!--追加する機能を具体的に箇条書きで--> - v-ifで遷移するのではないようにする - 絞り込みボタンを右上かどこかに設置して、それをタップするとモーダル等が出力して、それで探すように変更する ## 編集するファイル <!--既存するものであればそのファイル名を、新しく作成する場合は新しく作成するファイル名を記述--> -
1.0
Search画面のUI変更 - ## 追加する機能 <!--追加する機能を具体的に箇条書きで--> - v-ifで遷移するのではないようにする - 絞り込みボタンを右上かどこかに設置して、それをタップするとモーダル等が出力して、それで探すように変更する ## 編集するファイル <!--既存するものであればそのファイル名を、新しく作成する場合は新しく作成するファイル名を記述--> -
non_infrastructure
search画面のui変更 追加する機能 v ifで遷移するのではないようにする 絞り込みボタンを右上かどこかに設置して、それをタップするとモーダル等が出力して、それで探すように変更する 編集するファイル
0
10,918
8,805,663,752
IssuesEvent
2018-12-26 21:12:07
edgi-govdata-archiving/overview
https://api.github.com/repos/edgi-govdata-archiving/overview
closed
Perform security audit around development practices and tools
coordination infrastructure stale
While this is a bit vague as the objectives are part of a growing conversation, I wanted to drop this into an issue as a reminder. Resources - https://safetag.org/ - https://www.mozilla.org/en-US/moss/secure-open-source/ - https://www.opentech.fund/lab/engineering-lab
1.0
Perform security audit around development practices and tools - While this is a bit vague as the objectives are part of a growing conversation, I wanted to drop this into an issue as a reminder. Resources - https://safetag.org/ - https://www.mozilla.org/en-US/moss/secure-open-source/ - https://www.opentech.fund/lab/engineering-lab
infrastructure
perform security audit around development practices and tools while this is a bit vague as the objectives are part of a growing conversation i wanted to drop this into an issue as a reminder resources
1
14,797
11,148,568,505
IssuesEvent
2019-12-23 15:54:10
bajo1207/piggymetrics
https://api.github.com/repos/bajo1207/piggymetrics
closed
Basic project structure in CherryPy
Infrastructure
Setup virtual environment and implement underlying MVC structure in CherryPy.
1.0
Basic project structure in CherryPy - Setup virtual environment and implement underlying MVC structure in CherryPy.
infrastructure
basic project structure in cherrypy setup virtual environment and implement underlying mvc structure in cherrypy
1
265,620
20,104,934,365
IssuesEvent
2022-02-07 09:34:04
scikit-mobility/scikit-mobility
https://api.github.com/repos/scikit-mobility/scikit-mobility
closed
change documentation in the ReadMe: stops --> stay_location
documentation issue
`detection.stops` has been renamed into `stay_location`. The ReadMe should be updated accordingly.
1.0
change documentation in the ReadMe: stops --> stay_location - `detection.stops` has been renamed into `stay_location`. The ReadMe should be updated accordingly.
non_infrastructure
change documentation in the readme stops stay location detection stops has been renamed into stay location the readme should be updated accordingly
0
25,691
18,985,706,317
IssuesEvent
2021-11-21 17:27:11
fsprojects/ExcelFinancialFunctions
https://api.github.com/repos/fsprojects/ExcelFinancialFunctions
closed
Disable unstable test: mduration shouldn't be greater than maturity()
infrastructure
The test named "mduration shouldn't be greater than maturity()" fails occasionally. Will disable for now, troubleshoot in the future when I understand FsCheck better.
1.0
Disable unstable test: mduration shouldn't be greater than maturity() - The test named "mduration shouldn't be greater than maturity()" fails occasionally. Will disable for now, troubleshoot in the future when I understand FsCheck better.
infrastructure
disable unstable test mduration shouldn t be greater than maturity the test named mduration shouldn t be greater than maturity fails occasionally will disable for now troubleshoot in the future when i understand fscheck better
1
222,291
7,431,340,980
IssuesEvent
2018-03-25 13:49:19
galletti94/main
https://api.github.com/repos/galletti94/main
opened
Appathon Tasks
PRIORITY activity class function good first issue help wanted
TODO during Hackathon - [ ] Dynamic layout rendering for ViewModules @rhodesrm @natalieschulz - [ ] Dynamic layout rendering for ViewTasks @rhodesrm @natalieschulz - [ ] Dynamic layout rendering for ViewSingleTask @rhodesrm @natalieschulz - [ ] API for Database calls @Latterization @cpchang11 - [ ] Refinement of Database calls (urls + views) @victoriakayola @cpchang11
1.0
Appathon Tasks - TODO during Hackathon - [ ] Dynamic layout rendering for ViewModules @rhodesrm @natalieschulz - [ ] Dynamic layout rendering for ViewTasks @rhodesrm @natalieschulz - [ ] Dynamic layout rendering for ViewSingleTask @rhodesrm @natalieschulz - [ ] API for Database calls @Latterization @cpchang11 - [ ] Refinement of Database calls (urls + views) @victoriakayola @cpchang11
non_infrastructure
appathon tasks todo during hackathon dynamic layout rendering for viewmodules rhodesrm natalieschulz dynamic layout rendering for viewtasks rhodesrm natalieschulz dynamic layout rendering for viewsingletask rhodesrm natalieschulz api for database calls latterization refinement of database calls urls views victoriakayola
0
76,192
21,216,786,170
IssuesEvent
2022-04-11 08:09:33
PaddlePaddle/Paddle
https://api.github.com/repos/PaddlePaddle/Paddle
closed
曙光本地编译Paddle环境要求有哪些?
status/following-up type/build
使用paddlepaddle/paddle:latest-dev-rocm4.0-miopen2.11镜像,本地编译Paddle一直通不过,想问以下本地编译的要求有哪些,必须要有gpu吗,对cpu有没有要求? 本地环境为云服务器(网络环境便于使用github),1核cpu,2g内存,无gpu
1.0
曙光本地编译Paddle环境要求有哪些? - 使用paddlepaddle/paddle:latest-dev-rocm4.0-miopen2.11镜像,本地编译Paddle一直通不过,想问以下本地编译的要求有哪些,必须要有gpu吗,对cpu有没有要求? 本地环境为云服务器(网络环境便于使用github),1核cpu,2g内存,无gpu
non_infrastructure
曙光本地编译paddle环境要求有哪些? 使用paddlepaddle paddle latest dev ,本地编译paddle一直通不过,想问以下本地编译的要求有哪些,必须要有gpu吗,对cpu有没有要求? 本地环境为云服务器(网络环境便于使用github), , ,无gpu
0
11,174
3,471,155,084
IssuesEvent
2015-12-23 13:41:07
CartoDB/cartodb
https://api.github.com/repos/CartoDB/cartodb
closed
password_confirmation is not needed at Enterprise user management API
documentation
`password_confirmation` isn't actually needed, API works. It should be removed from the docs. cc @xavijam @gfiorav
1.0
password_confirmation is not needed at Enterprise user management API - `password_confirmation` isn't actually needed, API works. It should be removed from the docs. cc @xavijam @gfiorav
non_infrastructure
password confirmation is not needed at enterprise user management api password confirmation isn t actually needed api works it should be removed from the docs cc xavijam gfiorav
0
55,130
23,394,513,878
IssuesEvent
2022-08-11 21:25:52
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Remove index pattern selection when creating a filter in dashboard
Feature:Filters Feature:Query Bar loe:hours Team:AppServicesSv impact:low
Many of our users are using multiple index patterns in a single dashboard. While Elastic doesn't have a joins capability, the ability to filter multiple index patterns which have fields with the same name is a common use case. Up until recently when setting a filter it wouldn't take into consideration the index pattern e.g. `geo.country:US` . will filter any index pattern with `geo.country` field. I noticed that index patterns dropdown was added to the dashboard filters, this change can impact a lot of our users and prevent them from filtering multiple index patterns used on the same dashboard ![image](https://user-images.githubusercontent.com/32550316/65289767-e6ca2080-db00-11e9-9817-6450cca50942.png)
1.0
Remove index pattern selection when creating a filter in dashboard - Many of our users are using multiple index patterns in a single dashboard. While Elastic doesn't have a joins capability, the ability to filter multiple index patterns which have fields with the same name is a common use case. Up until recently when setting a filter it wouldn't take into consideration the index pattern e.g. `geo.country:US` . will filter any index pattern with `geo.country` field. I noticed that index patterns dropdown was added to the dashboard filters, this change can impact a lot of our users and prevent them from filtering multiple index patterns used on the same dashboard ![image](https://user-images.githubusercontent.com/32550316/65289767-e6ca2080-db00-11e9-9817-6450cca50942.png)
non_infrastructure
remove index pattern selection when creating a filter in dashboard many of our users are using multiple index patterns in a single dashboard while elastic doesn t have a joins capability the ability to filter multiple index patterns which have fields with the same name is a common use case up until recently when setting a filter it wouldn t take into consideration the index pattern e g geo country us will filter any index pattern with geo country field i noticed that index patterns dropdown was added to the dashboard filters this change can impact a lot of our users and prevent them from filtering multiple index patterns used on the same dashboard
0
12,638
9,885,311,089
IssuesEvent
2019-06-25 01:57:55
danieldeveloper001/FinancrAPI
https://api.github.com/repos/danieldeveloper001/FinancrAPI
closed
Currency Seed Script
infrastructure
## Objective Create a template seed script for the initial release with three fake currencies. ## Requirements - Currency specification
1.0
Currency Seed Script - ## Objective Create a template seed script for the initial release with three fake currencies. ## Requirements - Currency specification
infrastructure
currency seed script objective create a template seed script for the initial release with three fake currencies requirements currency specification
1
43,860
5,575,262,500
IssuesEvent
2017-03-28 01:10:32
infiniteautomation/ma-core-public
https://api.github.com/repos/infiniteautomation/ma-core-public
closed
Data File Data Source - Binary File importer
Enhancement Ready for Testing
I added a "Binary" type importer to the data file importer that takes an input stream
1.0
Data File Data Source - Binary File importer - I added a "Binary" type importer to the data file importer that takes an input stream
non_infrastructure
data file data source binary file importer i added a binary type importer to the data file importer that takes an input stream
0
130,749
12,461,947,640
IssuesEvent
2020-05-28 08:03:39
opendevstack/ods-core
https://api.github.com/repos/opendevstack/ods-core
closed
create-projects: Exchange environment variable with proper command line arguments
documentation enhancement
Both scripts in create-projects use some undocumented environment variables. A huge improvement to those scripts would be to exchange the usage of environment variables, which can lead to unforeseen side effects on other scripts or be effected by already exported values, with proper command line arguments with usage documentation.
1.0
create-projects: Exchange environment variable with proper command line arguments - Both scripts in create-projects use some undocumented environment variables. A huge improvement to those scripts would be to exchange the usage of environment variables, which can lead to unforeseen side effects on other scripts or be effected by already exported values, with proper command line arguments with usage documentation.
non_infrastructure
create projects exchange environment variable with proper command line arguments both scripts in create projects use some undocumented environment variables a huge improvement to those scripts would be to exchange the usage of environment variables which can lead to unforeseen side effects on other scripts or be effected by already exported values with proper command line arguments with usage documentation
0
29,001
23,660,810,739
IssuesEvent
2022-08-26 15:23:40
beamer-bridge/beamer
https://api.github.com/repos/beamer-bridge/beamer
closed
Remove TST token deployment from deployment script
infrastructure :hammer: post mainnet
We should remove TST token deployment from our deployment script because it does not make sense do deploy a new token contract every time we deploy Beamer contracts. This will also simplify our deployment process.
1.0
Remove TST token deployment from deployment script - We should remove TST token deployment from our deployment script because it does not make sense do deploy a new token contract every time we deploy Beamer contracts. This will also simplify our deployment process.
infrastructure
remove tst token deployment from deployment script we should remove tst token deployment from our deployment script because it does not make sense do deploy a new token contract every time we deploy beamer contracts this will also simplify our deployment process
1
19,163
13,196,263,597
IssuesEvent
2020-08-13 20:18:01
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
Test and merge the Developer Simulation Tool
Infrastructure
https://trello.com/c/Aly06hBh/143-test-and-merge-the-developer-simulation-tool Tool to run during upgrading the Lab to generate some load and watch for issues that may affect devs during the upgrade of Prod https://github.com/bcgov-c/platform-ops/pull/42
1.0
Test and merge the Developer Simulation Tool - https://trello.com/c/Aly06hBh/143-test-and-merge-the-developer-simulation-tool Tool to run during upgrading the Lab to generate some load and watch for issues that may affect devs during the upgrade of Prod https://github.com/bcgov-c/platform-ops/pull/42
infrastructure
test and merge the developer simulation tool tool to run during upgrading the lab to generate some load and watch for issues that may affect devs during the upgrade of prod
1
88,381
15,800,790,690
IssuesEvent
2021-04-03 01:17:59
hiucimon/zookeeper
https://api.github.com/repos/hiucimon/zookeeper
opened
CVE-2019-16943 (High) detected in jackson-databind-2.6.3.jar
security vulnerability
## CVE-2019-16943 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: zookeeper/libs/jackson-databind-2.6.3.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the p6spy (3.8.6) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of com.p6spy.engine.spy.P6DataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16943>CVE-2019-16943</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-16943 (High) detected in jackson-databind-2.6.3.jar - ## CVE-2019-16943 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: zookeeper/libs/jackson-databind-2.6.3.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the p6spy (3.8.6) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of com.p6spy.engine.spy.P6DataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16943>CVE-2019-16943</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16943</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library zookeeper libs jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of com engine spy mishandling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
105,609
9,087,669,068
IssuesEvent
2019-02-18 14:19:51
saltstack/salt
https://api.github.com/repos/saltstack/salt
closed
Add eauth tests for Windows
Needs Testcase Windows stale team-windows
### Description of Issue/Question Many eauth tests have been disabled on Windows. Need to test eauth on Windows to see what's working. Create tests that work on Windows. https://github.com/saltstack/salt/pull/44168
1.0
Add eauth tests for Windows - ### Description of Issue/Question Many eauth tests have been disabled on Windows. Need to test eauth on Windows to see what's working. Create tests that work on Windows. https://github.com/saltstack/salt/pull/44168
non_infrastructure
add eauth tests for windows description of issue question many eauth tests have been disabled on windows need to test eauth on windows to see what s working create tests that work on windows
0
506,958
14,677,002,527
IssuesEvent
2020-12-30 21:50:20
JambuOverflow/lepic
https://api.github.com/repos/JambuOverflow/lepic
reopened
11. As a Teacher OR Support Professional, I want to create/read/update a student account with his/her name and class
points: 3 priority: highest
## Acceptance Criteria 1. If I create a student account, the name and class should remain the same ones I put when I created it. 2. I should be able to see the student's data 2.1 See name and class 3. If I modify the information of a student (like name or class), the information in the student account should change. ## Sub-Tasks - [x] Prototype - [x] Front-End Dev - [x] Design - [x] Back-End Dev - [x] Data Bases
1.0
11. As a Teacher OR Support Professional, I want to create/read/update a student account with his/her name and class - ## Acceptance Criteria 1. If I create a student account, the name and class should remain the same ones I put when I created it. 2. I should be able to see the student's data 2.1 See name and class 3. If I modify the information of a student (like name or class), the information in the student account should change. ## Sub-Tasks - [x] Prototype - [x] Front-End Dev - [x] Design - [x] Back-End Dev - [x] Data Bases
non_infrastructure
as a teacher or support professional i want to create read update a student account with his her name and class acceptance criteria if i create a student account the name and class should remain the same ones i put when i created it i should be able to see the student s data see name and class if i modify the information of a student like name or class the information in the student account should change sub tasks prototype front end dev design back end dev data bases
0
31,414
25,729,800,027
IssuesEvent
2022-12-07 19:21:05
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
[Automated] PRs inserted in VS build main-33207.93
Area-Infrastructure untriaged vs-insertion
[View Complete Diff of Changes](https://github.com/dotnet/roslyn/compare/ea0b6351c129c7decd9bf76f4388239509e28e0d...c1237adacef7321e8cd5ccb38537ba982e146f4b?w=1) - [Move IDS_FeatureTuples checks out of the parser. (65679)](https://github.com/dotnet/roslyn/pull/65679) - [Update Language Feature Status for Primary Constructors (65698)](https://github.com/dotnet/roslyn/pull/65698) - [Fix incorrect duplicate conversion (65686)](https://github.com/dotnet/roslyn/pull/65686) - [Project system relayering (65701)](https://github.com/dotnet/roslyn/pull/65701) - [Revert "Revert "Merge pull request #65407 from Youssef1313/scoped-ide"" (65696)](https://github.com/dotnet/roslyn/pull/65696) - [Cleanup keyword recommender tests (65664)](https://github.com/dotnet/roslyn/pull/65664) - [Move IDS_FeatureXXX checks out of the parser (65587)](https://github.com/dotnet/roslyn/pull/65587) - [Simplify and clarify code around diagnostic modes. (65689)](https://github.com/dotnet/roslyn/pull/65689) - [Add feature flag to allow diagnostic tagging to be either push or pull. (65671)](https://github.com/dotnet/roslyn/pull/65671) - [[LSP] Don't throw when changed member is not a method member (65676)](https://github.com/dotnet/roslyn/pull/65676) - [Do all work for the 'fully qualify' fixer out of process. (65665)](https://github.com/dotnet/roslyn/pull/65665) - [[Infra] 17.5 P2 snap (65674)](https://github.com/dotnet/roslyn/pull/65674) - [Fix generation of parameter symbols with explicit default values (65639)](https://github.com/dotnet/roslyn/pull/65639) - [Move SRM to 6.0.1 (65416)](https://github.com/dotnet/roslyn/pull/65416)
1.0
[Automated] PRs inserted in VS build main-33207.93 - [View Complete Diff of Changes](https://github.com/dotnet/roslyn/compare/ea0b6351c129c7decd9bf76f4388239509e28e0d...c1237adacef7321e8cd5ccb38537ba982e146f4b?w=1) - [Move IDS_FeatureTuples checks out of the parser. (65679)](https://github.com/dotnet/roslyn/pull/65679) - [Update Language Feature Status for Primary Constructors (65698)](https://github.com/dotnet/roslyn/pull/65698) - [Fix incorrect duplicate conversion (65686)](https://github.com/dotnet/roslyn/pull/65686) - [Project system relayering (65701)](https://github.com/dotnet/roslyn/pull/65701) - [Revert "Revert "Merge pull request #65407 from Youssef1313/scoped-ide"" (65696)](https://github.com/dotnet/roslyn/pull/65696) - [Cleanup keyword recommender tests (65664)](https://github.com/dotnet/roslyn/pull/65664) - [Move IDS_FeatureXXX checks out of the parser (65587)](https://github.com/dotnet/roslyn/pull/65587) - [Simplify and clarify code around diagnostic modes. (65689)](https://github.com/dotnet/roslyn/pull/65689) - [Add feature flag to allow diagnostic tagging to be either push or pull. (65671)](https://github.com/dotnet/roslyn/pull/65671) - [[LSP] Don't throw when changed member is not a method member (65676)](https://github.com/dotnet/roslyn/pull/65676) - [Do all work for the 'fully qualify' fixer out of process. (65665)](https://github.com/dotnet/roslyn/pull/65665) - [[Infra] 17.5 P2 snap (65674)](https://github.com/dotnet/roslyn/pull/65674) - [Fix generation of parameter symbols with explicit default values (65639)](https://github.com/dotnet/roslyn/pull/65639) - [Move SRM to 6.0.1 (65416)](https://github.com/dotnet/roslyn/pull/65416)
infrastructure
prs inserted in vs build main don t throw when changed member is not a method member snap
1
33,879
27,965,861,877
IssuesEvent
2023-03-24 19:27:46
CDL-Dryad/dryad-product-roadmap
https://api.github.com/repos/CDL-Dryad/dryad-product-roadmap
closed
Renew certificates or update DataCite before Jan 5 or 12th
infrastructure maintenance
These two campus certificates expire at the first of next year. (January 5th and 12th). We either need to update DataCite to use the new datadryad.org URLs for the dataset landing pages or else renew these certificates so that redirects for these continue to work after those dates. ``` end_date: 2020-01-05 15:59:00 instance_name: uc3-dash2-prd location: /apps/dash2/ssl/2020-01-05/dash_lib_uci_edu_cert.cer serial_number: 2C43987054DBC7BAE1AB83995EB13C4B cn: dash.lib.uci.edu alternate_subject_names: DNS:dash.lib.uci.edu, DNS:dash-dev.lib.uci.edu, DNS:dash-stg.lib.uci.edu issuer: /C=US/ST=MI/L=Ann Arbor/O=Internet2/OU=InCommon/CN=InCommon RSA Server CA ``` and ``` end_date: 2020-01-12 15:59:00 instance_name: uc3-dash2-prd <---------------------------- host is still running location: /dash2/ssl/2020-01-12/dash_ucmerced_edu_cert.cer serial_number: 13984A76C66C38A146E4B3869CFF1508 cn: dash.ucmerced.edu alternate_subject_names: DNS:dash.ucmerced.edu issuer: /C=US/ST=MI/L=Ann Arbor/O=Internet2/OU=InCommon/CN=InCommon RSA Server CA ```
1.0
Renew certificates or update DataCite before Jan 5 or 12th - These two campus certificates expire at the first of next year. (January 5th and 12th). We either need to update DataCite to use the new datadryad.org URLs for the dataset landing pages or else renew these certificates so that redirects for these continue to work after those dates. ``` end_date: 2020-01-05 15:59:00 instance_name: uc3-dash2-prd location: /apps/dash2/ssl/2020-01-05/dash_lib_uci_edu_cert.cer serial_number: 2C43987054DBC7BAE1AB83995EB13C4B cn: dash.lib.uci.edu alternate_subject_names: DNS:dash.lib.uci.edu, DNS:dash-dev.lib.uci.edu, DNS:dash-stg.lib.uci.edu issuer: /C=US/ST=MI/L=Ann Arbor/O=Internet2/OU=InCommon/CN=InCommon RSA Server CA ``` and ``` end_date: 2020-01-12 15:59:00 instance_name: uc3-dash2-prd <---------------------------- host is still running location: /dash2/ssl/2020-01-12/dash_ucmerced_edu_cert.cer serial_number: 13984A76C66C38A146E4B3869CFF1508 cn: dash.ucmerced.edu alternate_subject_names: DNS:dash.ucmerced.edu issuer: /C=US/ST=MI/L=Ann Arbor/O=Internet2/OU=InCommon/CN=InCommon RSA Server CA ```
infrastructure
renew certificates or update datacite before jan or these two campus certificates expire at the first of next year january and we either need to update datacite to use the new datadryad org urls for the dataset landing pages or else renew these certificates so that redirects for these continue to work after those dates end date instance name prd location apps ssl dash lib uci edu cert cer serial number cn dash lib uci edu alternate subject names dns dash lib uci edu dns dash dev lib uci edu dns dash stg lib uci edu issuer c us st mi l ann arbor o ou incommon cn incommon rsa server ca and end date instance name prd host is still running location ssl dash ucmerced edu cert cer serial number cn dash ucmerced edu alternate subject names dns dash ucmerced edu issuer c us st mi l ann arbor o ou incommon cn incommon rsa server ca
1
6,810
6,620,460,760
IssuesEvent
2017-09-21 15:36:54
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
closed
Verbose error message for lint failure
Infrastructure
When I have a lint error the error message is extremely verbose. ``` ~\work\TypeScript [master ≡ +1 ~1 -0 !]> gulp runtests-parallel [08:07:23] Requiring external module ts-node/register [08:07:23] Using gulpfile ~\work\TypeScript\gulpfile.ts [08:07:23] Starting 'build-rules'... [08:07:23] Starting 'built\local\lib.dom.d.ts'... [08:07:23] Starting 'built\local\lib.dom.iterable.d.ts'... [08:07:23] Starting 'built\local\lib.webworker.d.ts'... [08:07:23] Starting 'built\local\lib.scripthost.d.ts'... [08:07:23] Starting 'built\local\lib.es5.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.d.ts'... [08:07:23] Starting 'built\local\lib.es2016.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.d.ts'... [08:07:23] Starting 'built\local\lib.esnext.d.ts'... [08:07:23] Starting 'built\local\lib.d.ts'... [08:07:23] Starting 'built\local\lib.es6.d.ts'... [08:07:23] Starting 'built\local\lib.es2016.full.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.full.d.ts'... [08:07:23] Starting 'built\local\lib.esnext.full.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.core.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.collection.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.generator.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.iterable.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.promise.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.proxy.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.reflect.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.symbol.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.symbol.wellknown.d.ts'... [08:07:23] Starting 'built\local\lib.es2016.array.include.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.object.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.sharedmemory.d.ts'... [08:07:24] Starting 'built\local\lib.es2017.string.d.ts'... [08:07:24] Starting 'built\local\lib.es2017.intl.d.ts'... [08:07:24] Starting 'built\local\lib.esnext.asynciterable.d.ts'... [08:07:24] Starting 'scripts\processDiagnosticMessages.js'... [08:07:24] Finished 'scripts\processDiagnosticMessages.js' after 43 ms [08:07:24] Starting 'src\compiler\diagnosticInformationMap.generated.ts'... [08:07:24] Finished 'src\compiler\diagnosticInformationMap.generated.ts' after 482 μs [08:07:24] Starting 'generate-diagnostics'... [08:07:24] Finished 'generate-diagnostics' after 22 μs [08:07:24] Finished 'built\local\lib.dom.iterable.d.ts' after 120 ms [08:07:24] Finished 'built\local\lib.es2016.d.ts' after 113 ms [08:07:24] Finished 'built\local\lib.es2015.d.ts' after 116 ms [08:07:24] Finished 'built\local\lib.esnext.d.ts' after 113 ms [08:07:24] Finished 'built\local\lib.scripthost.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2017.d.ts' after 117 ms [08:07:24] Finished 'built\local\lib.webworker.d.ts' after 132 ms [08:07:24] Finished 'built\local\lib.es5.d.ts' after 138 ms [08:07:24] Finished 'built\local\lib.es2015.collection.d.ts' after 112 ms [08:07:24] Finished 'built\local\lib.es2015.core.d.ts' after 117 ms [08:07:24] Finished 'built\local\lib.es2015.generator.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2015.promise.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2015.reflect.d.ts' after 120 ms [08:07:24] Finished 'built\local\lib.es2015.proxy.d.ts' after 123 ms [08:07:24] Finished 'built\local\lib.es2015.iterable.d.ts' after 127 ms [08:07:24] Finished 'built\local\lib.es2015.symbol.wellknown.d.ts' after 119 ms [08:07:24] Finished 'built\local\lib.es2015.symbol.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2016.array.include.d.ts' after 120 ms [08:07:24] Finished 'built\local\lib.es2017.sharedmemory.d.ts' after 118 ms [08:07:24] Finished 'built\local\lib.es2017.string.d.ts' after 117 ms [08:07:24] Finished 'built\local\lib.es2017.object.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2017.intl.d.ts' after 119 ms [08:07:24] Finished 'built\local\lib.dom.d.ts' after 178 ms [08:07:24] Finished 'built\local\lib.esnext.asynciterable.d.ts' after 113 ms [08:07:24] Finished 'built\local\lib.d.ts' after 172 ms [08:07:24] Finished 'build-rules' after 355 ms [08:07:24] Finished 'built\local\lib.es2016.full.d.ts' after 184 ms [08:07:24] Finished 'built\local\lib.es2017.full.d.ts' after 180 ms [08:07:24] Finished 'built\local\lib.esnext.full.d.ts' after 179 ms [08:07:24] Finished 'built\local\lib.es6.d.ts' after 195 ms [08:07:24] Starting 'lib'... [08:07:24] Finished 'lib' after 3.23 μs [08:07:24] Starting 'built\local\typescriptServices.js'... Compiler option 'compileOnSave' requires a value of type boolean. [08:07:24] Finished 'built\local\typescriptServices.js' after 111 ms [08:07:24] Starting 'built\local\run.js'... [08:07:24] Finished 'built\local\run.js' after 257 ms [08:07:24] Starting 'tests'... [08:07:24] Finished 'tests' after 4.11 μs [08:07:24] Starting 'runtests-parallel'... Running tests with config: {"light":false,"workerCount":8,"taskConfigsFolder":"C:\\Users\\anhans\\AppData\\Local\\Temp/ts-tests142","noColor":false} node built\local\run.js Discovering tests... Discovered 11201 test files in 1048ms. Starting to run tests using 8 threads... Batching initial test lists... Batched into 8 groups with approximate total file sizes of 752084 bytes in each group. [▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬] √ 57024 passing (3m) 57024 passing (3m) [08:09:57] Starting 'build-rules'... [08:09:57] Finished 'build-rules' after 58 ms [08:09:57] Starting 'lint'... Linting: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish src/services/formatting/formattingScanner.ts ERROR: 128:27 comment-format comment must start with a space [08:10:07] 'lint' errored after 9.81 s [08:10:07] Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish at checkExecSyncError (child_process.js:591:13) at Object.execSync (child_process.js:631:13) at Gulp.<anonymous> (C:\Users\anhans\work\TypeScript\gulpfile.ts:1057:19) at module.exports (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:34:7) at Gulp.Orchestrator._runTask (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:273:3) at Gulp.Orchestrator._runStep (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:214:10) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:279:18 at finish (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:21:8) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:52:4 at f (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\node_modules\once\once.js:17:25) at DestroyableTransform.onend (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\index.js:31:18) at emitNone (events.js:110:20) at DestroyableTransform.emit (events.js:207:7) at C:\Users\anhans\work\TypeScript\node_modules\vinyl-fs\node_modules\readable-stream\lib\_stream_readable.js:965:16 at _combinedTickCallback (internal/process/next_tick.js:131:7) at process._tickCallback (internal/process/next_tick.js:180:9) { Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish at checkExecSyncError (child_process.js:591:13) at Object.execSync (child_process.js:631:13) at Gulp.<anonymous> (C:\Users\anhans\work\TypeScript\gulpfile.ts:1057:19) at module.exports (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:34:7) at Gulp.Orchestrator._runTask (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:273:3) at Gulp.Orchestrator._runStep (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:214:10) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:279:18 at finish (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:21:8) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:52:4 at f (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\node_modules\once\once.js:17:25) at DestroyableTransform.onend (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\index.js:31:18) at emitNone (events.js:110:20) at DestroyableTransform.emit (events.js:207:7) at C:\Users\anhans\work\TypeScript\node_modules\vinyl-fs\node_modules\readable-stream\lib\_stream_readable.js:965:16 at _combinedTickCallback (internal/process/next_tick.js:131:7) at process._tickCallback (internal/process/next_tick.js:180:9) error: null, cmd: 'node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish', file: 'C:\\WINDOWS\\system32\\cmd.exe', args: [ 'C:\\WINDOWS\\system32\\cmd.exe', '/d', '/s', '/c', '"node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish"' ], options: { stdio: [ [Object], [Object], [Object] ], shell: true, windowsVerbatimArguments: true, file: 'C:\\WINDOWS\\system32\\cmd.exe', args: [ 'C:\\WINDOWS\\system32\\cmd.exe', '/d', '/s', '/c', '"node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish"' ], envPairs: [ ...censored... ], killSignal: undefined }, envPairs: [ ...censored... ], stderr: null, stdout: null, pid: 12364, output: [ null, null, null ], signal: null, status: 2, name: 'Error', message: 'Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish', stack: 'Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish\n at checkExecSyncError (child_process.js:591:13)\n at Object.execSync (child_process.js:631:13)\n at Gulp.<anonymous> (C:\\Users\\anhans\\work\\TypeScript\\gulpfile.ts:1057:19)\n at module.exports (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\lib\\runTask.js:34:7)\n at Gulp.Orchestrator._runTask (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\index.js:273:3)\n at Gulp.Orchestrator._runStep (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\index.js:214:10)\n at C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\index.js:279:18\n at finish (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\lib\\runTask.js:21:8)\n at C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\lib\\runTask.js:52:4\n at f (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\end-of-stream\\node_modules\\once\\once.js:17:25)\n at DestroyableTransform.onend (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\end-of-stream\\index.js:31:18)\n at emitNone (events.js:110:20)\n at DestroyableTransform.emit (events.js:207:7)\n at C:\\Users\\anhans\\work\\TypeScript\\node_modules\\vinyl-fs\\node_modules\\readable-stream\\lib\\_stream_readable.js:965:16\n at _combinedTickCallback (internal/process/next_tick.js:131:7)\n at process._tickCallback (internal/process/next_tick.js:180:9)', showStack: true, showProperties: true, plugin: 'run-sequence(lint)' } [08:10:08] 'runtests-parallel' errored after 2.72 min [08:10:08] Error in plugin 'run-sequence(lint)' Message: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish Details: error: null cmd: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish file: C:\WINDOWS\system32\cmd.exe args: C:\WINDOWS\system32\cmd.exe,/d,/s,/c,"node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish" options: [object Object] envPairs: ...censored... stderr: null stdout: null pid: 12364 output: ,, signal: null status: 2 Stack: Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish at checkExecSyncError (child_process.js:591:13) at Object.execSync (child_process.js:631:13) at Gulp.<anonymous> (C:\Users\anhans\work\TypeScript\gulpfile.ts:1057:19) at module.exports (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:34:7) at Gulp.Orchestrator._runTask (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:273:3) at Gulp.Orchestrator._runStep (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:214:10) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:279:18 at finish (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:21:8) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:52:4 at f (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\node_modules\once\once.js:17:25) at DestroyableTransform.onend (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\index.js:31:18) at emitNone (events.js:110:20) at DestroyableTransform.emit (events.js:207:7) at C:\Users\anhans\work\TypeScript\node_modules\vinyl-fs\node_modules\readable-stream\lib\_stream_readable.js:965:16 at _combinedTickCallback (internal/process/next_tick.js:131:7) at process._tickCallback (internal/process/next_tick.js:180:9) ``` It's even longer when you include printing out my environment variables three times, which I've omitted here. Probably related to #18462. CC @weswigham
1.0
Verbose error message for lint failure - When I have a lint error the error message is extremely verbose. ``` ~\work\TypeScript [master ≡ +1 ~1 -0 !]> gulp runtests-parallel [08:07:23] Requiring external module ts-node/register [08:07:23] Using gulpfile ~\work\TypeScript\gulpfile.ts [08:07:23] Starting 'build-rules'... [08:07:23] Starting 'built\local\lib.dom.d.ts'... [08:07:23] Starting 'built\local\lib.dom.iterable.d.ts'... [08:07:23] Starting 'built\local\lib.webworker.d.ts'... [08:07:23] Starting 'built\local\lib.scripthost.d.ts'... [08:07:23] Starting 'built\local\lib.es5.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.d.ts'... [08:07:23] Starting 'built\local\lib.es2016.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.d.ts'... [08:07:23] Starting 'built\local\lib.esnext.d.ts'... [08:07:23] Starting 'built\local\lib.d.ts'... [08:07:23] Starting 'built\local\lib.es6.d.ts'... [08:07:23] Starting 'built\local\lib.es2016.full.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.full.d.ts'... [08:07:23] Starting 'built\local\lib.esnext.full.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.core.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.collection.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.generator.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.iterable.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.promise.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.proxy.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.reflect.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.symbol.d.ts'... [08:07:23] Starting 'built\local\lib.es2015.symbol.wellknown.d.ts'... [08:07:23] Starting 'built\local\lib.es2016.array.include.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.object.d.ts'... [08:07:23] Starting 'built\local\lib.es2017.sharedmemory.d.ts'... [08:07:24] Starting 'built\local\lib.es2017.string.d.ts'... [08:07:24] Starting 'built\local\lib.es2017.intl.d.ts'... [08:07:24] Starting 'built\local\lib.esnext.asynciterable.d.ts'... [08:07:24] Starting 'scripts\processDiagnosticMessages.js'... [08:07:24] Finished 'scripts\processDiagnosticMessages.js' after 43 ms [08:07:24] Starting 'src\compiler\diagnosticInformationMap.generated.ts'... [08:07:24] Finished 'src\compiler\diagnosticInformationMap.generated.ts' after 482 μs [08:07:24] Starting 'generate-diagnostics'... [08:07:24] Finished 'generate-diagnostics' after 22 μs [08:07:24] Finished 'built\local\lib.dom.iterable.d.ts' after 120 ms [08:07:24] Finished 'built\local\lib.es2016.d.ts' after 113 ms [08:07:24] Finished 'built\local\lib.es2015.d.ts' after 116 ms [08:07:24] Finished 'built\local\lib.esnext.d.ts' after 113 ms [08:07:24] Finished 'built\local\lib.scripthost.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2017.d.ts' after 117 ms [08:07:24] Finished 'built\local\lib.webworker.d.ts' after 132 ms [08:07:24] Finished 'built\local\lib.es5.d.ts' after 138 ms [08:07:24] Finished 'built\local\lib.es2015.collection.d.ts' after 112 ms [08:07:24] Finished 'built\local\lib.es2015.core.d.ts' after 117 ms [08:07:24] Finished 'built\local\lib.es2015.generator.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2015.promise.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2015.reflect.d.ts' after 120 ms [08:07:24] Finished 'built\local\lib.es2015.proxy.d.ts' after 123 ms [08:07:24] Finished 'built\local\lib.es2015.iterable.d.ts' after 127 ms [08:07:24] Finished 'built\local\lib.es2015.symbol.wellknown.d.ts' after 119 ms [08:07:24] Finished 'built\local\lib.es2015.symbol.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2016.array.include.d.ts' after 120 ms [08:07:24] Finished 'built\local\lib.es2017.sharedmemory.d.ts' after 118 ms [08:07:24] Finished 'built\local\lib.es2017.string.d.ts' after 117 ms [08:07:24] Finished 'built\local\lib.es2017.object.d.ts' after 122 ms [08:07:24] Finished 'built\local\lib.es2017.intl.d.ts' after 119 ms [08:07:24] Finished 'built\local\lib.dom.d.ts' after 178 ms [08:07:24] Finished 'built\local\lib.esnext.asynciterable.d.ts' after 113 ms [08:07:24] Finished 'built\local\lib.d.ts' after 172 ms [08:07:24] Finished 'build-rules' after 355 ms [08:07:24] Finished 'built\local\lib.es2016.full.d.ts' after 184 ms [08:07:24] Finished 'built\local\lib.es2017.full.d.ts' after 180 ms [08:07:24] Finished 'built\local\lib.esnext.full.d.ts' after 179 ms [08:07:24] Finished 'built\local\lib.es6.d.ts' after 195 ms [08:07:24] Starting 'lib'... [08:07:24] Finished 'lib' after 3.23 μs [08:07:24] Starting 'built\local\typescriptServices.js'... Compiler option 'compileOnSave' requires a value of type boolean. [08:07:24] Finished 'built\local\typescriptServices.js' after 111 ms [08:07:24] Starting 'built\local\run.js'... [08:07:24] Finished 'built\local\run.js' after 257 ms [08:07:24] Starting 'tests'... [08:07:24] Finished 'tests' after 4.11 μs [08:07:24] Starting 'runtests-parallel'... Running tests with config: {"light":false,"workerCount":8,"taskConfigsFolder":"C:\\Users\\anhans\\AppData\\Local\\Temp/ts-tests142","noColor":false} node built\local\run.js Discovering tests... Discovered 11201 test files in 1048ms. Starting to run tests using 8 threads... Batching initial test lists... Batched into 8 groups with approximate total file sizes of 752084 bytes in each group. [▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬] √ 57024 passing (3m) 57024 passing (3m) [08:09:57] Starting 'build-rules'... [08:09:57] Finished 'build-rules' after 58 ms [08:09:57] Starting 'lint'... Linting: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish src/services/formatting/formattingScanner.ts ERROR: 128:27 comment-format comment must start with a space [08:10:07] 'lint' errored after 9.81 s [08:10:07] Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish at checkExecSyncError (child_process.js:591:13) at Object.execSync (child_process.js:631:13) at Gulp.<anonymous> (C:\Users\anhans\work\TypeScript\gulpfile.ts:1057:19) at module.exports (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:34:7) at Gulp.Orchestrator._runTask (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:273:3) at Gulp.Orchestrator._runStep (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:214:10) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:279:18 at finish (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:21:8) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:52:4 at f (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\node_modules\once\once.js:17:25) at DestroyableTransform.onend (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\index.js:31:18) at emitNone (events.js:110:20) at DestroyableTransform.emit (events.js:207:7) at C:\Users\anhans\work\TypeScript\node_modules\vinyl-fs\node_modules\readable-stream\lib\_stream_readable.js:965:16 at _combinedTickCallback (internal/process/next_tick.js:131:7) at process._tickCallback (internal/process/next_tick.js:180:9) { Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish at checkExecSyncError (child_process.js:591:13) at Object.execSync (child_process.js:631:13) at Gulp.<anonymous> (C:\Users\anhans\work\TypeScript\gulpfile.ts:1057:19) at module.exports (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:34:7) at Gulp.Orchestrator._runTask (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:273:3) at Gulp.Orchestrator._runStep (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:214:10) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:279:18 at finish (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:21:8) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:52:4 at f (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\node_modules\once\once.js:17:25) at DestroyableTransform.onend (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\index.js:31:18) at emitNone (events.js:110:20) at DestroyableTransform.emit (events.js:207:7) at C:\Users\anhans\work\TypeScript\node_modules\vinyl-fs\node_modules\readable-stream\lib\_stream_readable.js:965:16 at _combinedTickCallback (internal/process/next_tick.js:131:7) at process._tickCallback (internal/process/next_tick.js:180:9) error: null, cmd: 'node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish', file: 'C:\\WINDOWS\\system32\\cmd.exe', args: [ 'C:\\WINDOWS\\system32\\cmd.exe', '/d', '/s', '/c', '"node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish"' ], options: { stdio: [ [Object], [Object], [Object] ], shell: true, windowsVerbatimArguments: true, file: 'C:\\WINDOWS\\system32\\cmd.exe', args: [ 'C:\\WINDOWS\\system32\\cmd.exe', '/d', '/s', '/c', '"node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish"' ], envPairs: [ ...censored... ], killSignal: undefined }, envPairs: [ ...censored... ], stderr: null, stdout: null, pid: 12364, output: [ null, null, null ], signal: null, status: 2, name: 'Error', message: 'Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish', stack: 'Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts \'scripts/tslint/*.ts\' \'src/**/*.ts\' --exclude src/lib/es5.d.ts --exclude \'src/lib/*.generated.d.ts\' --format stylish\n at checkExecSyncError (child_process.js:591:13)\n at Object.execSync (child_process.js:631:13)\n at Gulp.<anonymous> (C:\\Users\\anhans\\work\\TypeScript\\gulpfile.ts:1057:19)\n at module.exports (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\lib\\runTask.js:34:7)\n at Gulp.Orchestrator._runTask (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\index.js:273:3)\n at Gulp.Orchestrator._runStep (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\index.js:214:10)\n at C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\index.js:279:18\n at finish (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\lib\\runTask.js:21:8)\n at C:\\Users\\anhans\\work\\TypeScript\\node_modules\\orchestrator\\lib\\runTask.js:52:4\n at f (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\end-of-stream\\node_modules\\once\\once.js:17:25)\n at DestroyableTransform.onend (C:\\Users\\anhans\\work\\TypeScript\\node_modules\\end-of-stream\\index.js:31:18)\n at emitNone (events.js:110:20)\n at DestroyableTransform.emit (events.js:207:7)\n at C:\\Users\\anhans\\work\\TypeScript\\node_modules\\vinyl-fs\\node_modules\\readable-stream\\lib\\_stream_readable.js:965:16\n at _combinedTickCallback (internal/process/next_tick.js:131:7)\n at process._tickCallback (internal/process/next_tick.js:180:9)', showStack: true, showProperties: true, plugin: 'run-sequence(lint)' } [08:10:08] 'runtests-parallel' errored after 2.72 min [08:10:08] Error in plugin 'run-sequence(lint)' Message: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish Details: error: null cmd: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish file: C:\WINDOWS\system32\cmd.exe args: C:\WINDOWS\system32\cmd.exe,/d,/s,/c,"node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish" options: [object Object] envPairs: ...censored... stderr: null stdout: null pid: 12364 output: ,, signal: null status: 2 Stack: Error: Command failed: node node_modules/tslint/bin/tslint Gulpfile.ts 'scripts/tslint/*.ts' 'src/**/*.ts' --exclude src/lib/es5.d.ts --exclude 'src/lib/*.generated.d.ts' --format stylish at checkExecSyncError (child_process.js:591:13) at Object.execSync (child_process.js:631:13) at Gulp.<anonymous> (C:\Users\anhans\work\TypeScript\gulpfile.ts:1057:19) at module.exports (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:34:7) at Gulp.Orchestrator._runTask (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:273:3) at Gulp.Orchestrator._runStep (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:214:10) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\index.js:279:18 at finish (C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:21:8) at C:\Users\anhans\work\TypeScript\node_modules\orchestrator\lib\runTask.js:52:4 at f (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\node_modules\once\once.js:17:25) at DestroyableTransform.onend (C:\Users\anhans\work\TypeScript\node_modules\end-of-stream\index.js:31:18) at emitNone (events.js:110:20) at DestroyableTransform.emit (events.js:207:7) at C:\Users\anhans\work\TypeScript\node_modules\vinyl-fs\node_modules\readable-stream\lib\_stream_readable.js:965:16 at _combinedTickCallback (internal/process/next_tick.js:131:7) at process._tickCallback (internal/process/next_tick.js:180:9) ``` It's even longer when you include printing out my environment variables three times, which I've omitted here. Probably related to #18462. CC @weswigham
infrastructure
verbose error message for lint failure when i have a lint error the error message is extremely verbose work typescript gulp runtests parallel requiring external module ts node register using gulpfile work typescript gulpfile ts starting build rules starting built local lib dom d ts starting built local lib dom iterable d ts starting built local lib webworker d ts starting built local lib scripthost d ts starting built local lib d ts starting built local lib d ts starting built local lib d ts starting built local lib d ts starting built local lib esnext d ts starting built local lib d ts starting built local lib d ts starting built local lib full d ts starting built local lib full d ts starting built local lib esnext full d ts starting built local lib core d ts starting built local lib collection d ts starting built local lib generator d ts starting built local lib iterable d ts starting built local lib promise d ts starting built local lib proxy d ts starting built local lib reflect d ts starting built local lib symbol d ts starting built local lib symbol wellknown d ts starting built local lib array include d ts starting built local lib object d ts starting built local lib sharedmemory d ts starting built local lib string d ts starting built local lib intl d ts starting built local lib esnext asynciterable d ts starting scripts processdiagnosticmessages js finished scripts processdiagnosticmessages js after ms starting src compiler diagnosticinformationmap generated ts finished src compiler diagnosticinformationmap generated ts after μs starting generate diagnostics finished generate diagnostics after μs finished built local lib dom iterable d ts after ms finished built local lib d ts after ms finished built local lib d ts after ms finished built local lib esnext d ts after ms finished built local lib scripthost d ts after ms finished built local lib d ts after ms finished built local lib webworker d ts after ms finished built local lib d ts after ms finished built local lib collection d ts after ms finished built local lib core d ts after ms finished built local lib generator d ts after ms finished built local lib promise d ts after ms finished built local lib reflect d ts after ms finished built local lib proxy d ts after ms finished built local lib iterable d ts after ms finished built local lib symbol wellknown d ts after ms finished built local lib symbol d ts after ms finished built local lib array include d ts after ms finished built local lib sharedmemory d ts after ms finished built local lib string d ts after ms finished built local lib object d ts after ms finished built local lib intl d ts after ms finished built local lib dom d ts after ms finished built local lib esnext asynciterable d ts after ms finished built local lib d ts after ms finished build rules after ms finished built local lib full d ts after ms finished built local lib full d ts after ms finished built local lib esnext full d ts after ms finished built local lib d ts after ms starting lib finished lib after μs starting built local typescriptservices js compiler option compileonsave requires a value of type boolean finished built local typescriptservices js after ms starting built local run js finished built local run js after ms starting tests finished tests after μs starting runtests parallel running tests with config light false workercount taskconfigsfolder c users anhans appdata local temp ts nocolor false node built local run js discovering tests discovered test files in starting to run tests using threads batching initial test lists batched into groups with approximate total file sizes of bytes in each group √ passing passing starting build rules finished build rules after ms starting lint linting node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish src services formatting formattingscanner ts error comment format comment must start with a space lint errored after s error command failed node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish at checkexecsyncerror child process js at object execsync child process js at gulp c users anhans work typescript gulpfile ts at module exports c users anhans work typescript node modules orchestrator lib runtask js at gulp orchestrator runtask c users anhans work typescript node modules orchestrator index js at gulp orchestrator runstep c users anhans work typescript node modules orchestrator index js at c users anhans work typescript node modules orchestrator index js at finish c users anhans work typescript node modules orchestrator lib runtask js at c users anhans work typescript node modules orchestrator lib runtask js at f c users anhans work typescript node modules end of stream node modules once once js at destroyabletransform onend c users anhans work typescript node modules end of stream index js at emitnone events js at destroyabletransform emit events js at c users anhans work typescript node modules vinyl fs node modules readable stream lib stream readable js at combinedtickcallback internal process next tick js at process tickcallback internal process next tick js error command failed node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish at checkexecsyncerror child process js at object execsync child process js at gulp c users anhans work typescript gulpfile ts at module exports c users anhans work typescript node modules orchestrator lib runtask js at gulp orchestrator runtask c users anhans work typescript node modules orchestrator index js at gulp orchestrator runstep c users anhans work typescript node modules orchestrator index js at c users anhans work typescript node modules orchestrator index js at finish c users anhans work typescript node modules orchestrator lib runtask js at c users anhans work typescript node modules orchestrator lib runtask js at f c users anhans work typescript node modules end of stream node modules once once js at destroyabletransform onend c users anhans work typescript node modules end of stream index js at emitnone events js at destroyabletransform emit events js at c users anhans work typescript node modules vinyl fs node modules readable stream lib stream readable js at combinedtickcallback internal process next tick js at process tickcallback internal process next tick js error null cmd node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish file c windows cmd exe args c windows cmd exe d s c node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish options stdio shell true windowsverbatimarguments true file c windows cmd exe args c windows cmd exe d s c node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish envpairs killsignal undefined envpairs stderr null stdout null pid output signal null status name error message command failed node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish stack error command failed node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish n at checkexecsyncerror child process js n at object execsync child process js n at gulp c users anhans work typescript gulpfile ts n at module exports c users anhans work typescript node modules orchestrator lib runtask js n at gulp orchestrator runtask c users anhans work typescript node modules orchestrator index js n at gulp orchestrator runstep c users anhans work typescript node modules orchestrator index js n at c users anhans work typescript node modules orchestrator index js n at finish c users anhans work typescript node modules orchestrator lib runtask js n at c users anhans work typescript node modules orchestrator lib runtask js n at f c users anhans work typescript node modules end of stream node modules once once js n at destroyabletransform onend c users anhans work typescript node modules end of stream index js n at emitnone events js n at destroyabletransform emit events js n at c users anhans work typescript node modules vinyl fs node modules readable stream lib stream readable js n at combinedtickcallback internal process next tick js n at process tickcallback internal process next tick js showstack true showproperties true plugin run sequence lint runtests parallel errored after min error in plugin run sequence lint message command failed node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish details error null cmd node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish file c windows cmd exe args c windows cmd exe d s c node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish options envpairs censored stderr null stdout null pid output signal null status stack error command failed node node modules tslint bin tslint gulpfile ts scripts tslint ts src ts exclude src lib d ts exclude src lib generated d ts format stylish at checkexecsyncerror child process js at object execsync child process js at gulp c users anhans work typescript gulpfile ts at module exports c users anhans work typescript node modules orchestrator lib runtask js at gulp orchestrator runtask c users anhans work typescript node modules orchestrator index js at gulp orchestrator runstep c users anhans work typescript node modules orchestrator index js at c users anhans work typescript node modules orchestrator index js at finish c users anhans work typescript node modules orchestrator lib runtask js at c users anhans work typescript node modules orchestrator lib runtask js at f c users anhans work typescript node modules end of stream node modules once once js at destroyabletransform onend c users anhans work typescript node modules end of stream index js at emitnone events js at destroyabletransform emit events js at c users anhans work typescript node modules vinyl fs node modules readable stream lib stream readable js at combinedtickcallback internal process next tick js at process tickcallback internal process next tick js it s even longer when you include printing out my environment variables three times which i ve omitted here probably related to cc weswigham
1
71,227
18,531,226,301
IssuesEvent
2021-10-21 06:16:53
orbeon/orbeon-forms
https://api.github.com/repos/orbeon/orbeon-forms
closed
Hint for single checkbox (`<fr:checkbox-input>`) isn't saved in the form definition
Module: Form Builder
To reproduce, create a form, add a single checkbox, enter a hint, press enter. The hint shows. Open the form definition and you'll notice that the hint isn't showing in the resources (while of course it should). Save the form, reload the page, and notice that the hint is lost (well, of course, it wasn't saved in the form definition).
1.0
Hint for single checkbox (`<fr:checkbox-input>`) isn't saved in the form definition - To reproduce, create a form, add a single checkbox, enter a hint, press enter. The hint shows. Open the form definition and you'll notice that the hint isn't showing in the resources (while of course it should). Save the form, reload the page, and notice that the hint is lost (well, of course, it wasn't saved in the form definition).
non_infrastructure
hint for single checkbox isn t saved in the form definition to reproduce create a form add a single checkbox enter a hint press enter the hint shows open the form definition and you ll notice that the hint isn t showing in the resources while of course it should save the form reload the page and notice that the hint is lost well of course it wasn t saved in the form definition
0
172,578
27,300,035,175
IssuesEvent
2023-02-24 00:37:02
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[REMOTO] [C#] [BACK-END] [.NET CORE] Tech Lead - Engenheiro de Software (Backend C#) na [STONE]
HOME OFFICE BACK-END .NET C# DOCKER DEVOPS SOLID REMOTO BACKEND RABBITMQ .NET CORE AZURE CI/CD DESIGN PATTERNS APIs RESTFUL HELP WANTED HTTP TECH LEAD ZEROMQ IBM MQ Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Na Tesouraria Stone Co. temos como missão manter nosso grupo saudável financeiramente, arcando com todas as obrigações de forma eficiente, além de garantir que todos os produtos que necessitam de recursos financeiros em suas operações sejam viáveis, sem comprometer a liquidez do grupo. - Isso é uma senhora responsabilidade, operamos bilhões de reais diariamente para alcançarmos tal objetivo. - Procuramos engenheiros de software criativos, responsáveis, com energia e que tenham foco em resolver o problema do cliente com qualidade. Também valorizamos suas habilidades práticas de desenvolvimento e capacidade de trabalhar bem em equipe. - É importante destacar também que você precisa gostar do “economês”, já que o nosso produto tem grande relação com o mercado financeiro. - Se você reúne essas características essa vaga é pra você! **Atividades:** - Você irá projetar, desenvolver, testar, implantar, manter e melhorar o software. - Você gerenciará as prioridades, prazos e trabalhos individuais do projeto. - Você terá que traduzir requisitos funcionais e técnicos em arquitetura e design detalhados de soluções, direcionando o time na direção correta e com visão de longo prazo. - Você precisará instruir o time em boas práticas de desenvolvimento, padrões de projeto, testes e revisão de código. - Você irá ser um guardião das nossas boas práticas e documentações, cultivando e desenvolvendo-as com o time. - Você irá promover uma cultura do time que vá além do código, observando aspectos não funcionais da aplicação como: performance, observabilidade, resiliência, segurança mantendo o foco do time em relação a estes aspectos. - É esperado de você extrema facilidade de trabalho em equipe, muita franqueza e transparência para que no dia-a-dia possa dar bons feedbacks e transmitir o conhecimento para os demais do time. - É esperado de você respeito e compromisso com o desenvolvimento dos demais colegas do time, ajudando no direcionamento da carreira e na evolução constante. ## Local - Home Office ## Benefícios - Familiar: Convênio SESC (gratuito); plano de saúde e odontológico (Bradesco TNQ2 sem co-participação) com Concierge. - Filhos: Auxílio-creche para crianças até 5 anos. - Bem-estar: Gympass (bora malhar!!!); horário flexível; trabalho remoto; dress code informal; seguro de vida; vale Transporte (se quiser, é opcional); vale refeição e/ou vale Alimentação e bicicletário - Mentoria e Engajamento:Super incentivamos idas a eventos e treinamentos, e se fizer sentido na estratégia, melhor ainda (qualquer um pode pedir); trabalhar com referências técnicas; eventos internos (hackathon, black friday,confraternizações das tribos, datas comemorativas) e aprendizado contínuo. ## Requisitos **Obrigatórios:** - Formação de nível superior em análise de sistemas, ciência da computação, engenharia de software ou pelo menos 8 anos de experiência em desenvolvimento de software. - Sólidos conhecimentos sobre algoritmos e estruturas de dados - Sólidos conhecimentos sobre padrões de projetos (Design Patterns) e SOLID - Sólidos conhecimentos da linguagem C# - Sólidos conhecimentos Restful APIs e o protocolo Http - Pelo menos 6 anos de experiência com desenvolvimento backend usando .Net Framework. - Pelo menos 2 anos de experiência com .Net Core. - Domínio de Docker e Conteinerização - Conhecimentos de testes de unidade, de integração e carga - Sólida experiência com sistemas distribuídos e mensageria e/ou eventos - Conhecimentos sobre algum message broker, tais como: RabbitMQ, ZeroMQ, Kafka, IBM Mq - Conhecimentos de abordagens como: Micro Serviços, Event Sourcing, CQRS, Hexagonal Architecture, Lambda Architecture - Experiência com modelagens Relacionais e não Relacionais - Experiência com análise e otimização de Queries SQL - Conhecimentos em OAuth, OpenID e JWT - Usar git de forma natural e conhecer o gitflow - Conhecimento de ferramentas de CI/CD, Azure DevOps é um diferencial aqui - Conhecimento de algum serviço de nuvem, Azure é um diferencial aqui - Familiaridade com metodologias ágeis e suas aplicações práticas, principalmente Kanban e/ou Scrum ## Contratação - a combinar ## Nossa empresa - PRAZER, SOMOS A STECH! O TIME DE TECNOLOGIA DA STONE: - Aqui na Stech, prezamos pela liberdade e autonomia. Liberdade para escolher seus horários, suas stacks, a forma como você vai trabalhar e mais importante, liberdade de ser quem você é. Você vai ter autonomia para fazer as suas escolhas e será cobrado por elas. Somos todos adultos e gostamos muito de um termo chamado “Accountability”, que podemos traduzir para responsabilidade pelos seus resultados. Queremos que você procure primeiro entender, para depois ser entendido. Mas não se preocupe, você sempre será ouvido, sua opinião conta e conta muito. Por último, e com muita importância para nós: Nós prezamos pela troca de conhecimento. “Be humble and kind”, esteja sempre de coração aberto para aprender novas formas de pensar e conhecimentos diferentes. Queremos achar em vocês inteligência, energia e integridade. E em troca, oferecemos autonomia, liberdade e flexibilidade. ## Como se candidatar - [Clique aqui para se candidatar](https://jornada.stone.com.br/formulario/stonetech?gh_jid=4316271003)
1.0
[REMOTO] [C#] [BACK-END] [.NET CORE] Tech Lead - Engenheiro de Software (Backend C#) na [STONE] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Na Tesouraria Stone Co. temos como missão manter nosso grupo saudável financeiramente, arcando com todas as obrigações de forma eficiente, além de garantir que todos os produtos que necessitam de recursos financeiros em suas operações sejam viáveis, sem comprometer a liquidez do grupo. - Isso é uma senhora responsabilidade, operamos bilhões de reais diariamente para alcançarmos tal objetivo. - Procuramos engenheiros de software criativos, responsáveis, com energia e que tenham foco em resolver o problema do cliente com qualidade. Também valorizamos suas habilidades práticas de desenvolvimento e capacidade de trabalhar bem em equipe. - É importante destacar também que você precisa gostar do “economês”, já que o nosso produto tem grande relação com o mercado financeiro. - Se você reúne essas características essa vaga é pra você! **Atividades:** - Você irá projetar, desenvolver, testar, implantar, manter e melhorar o software. - Você gerenciará as prioridades, prazos e trabalhos individuais do projeto. - Você terá que traduzir requisitos funcionais e técnicos em arquitetura e design detalhados de soluções, direcionando o time na direção correta e com visão de longo prazo. - Você precisará instruir o time em boas práticas de desenvolvimento, padrões de projeto, testes e revisão de código. - Você irá ser um guardião das nossas boas práticas e documentações, cultivando e desenvolvendo-as com o time. - Você irá promover uma cultura do time que vá além do código, observando aspectos não funcionais da aplicação como: performance, observabilidade, resiliência, segurança mantendo o foco do time em relação a estes aspectos. - É esperado de você extrema facilidade de trabalho em equipe, muita franqueza e transparência para que no dia-a-dia possa dar bons feedbacks e transmitir o conhecimento para os demais do time. - É esperado de você respeito e compromisso com o desenvolvimento dos demais colegas do time, ajudando no direcionamento da carreira e na evolução constante. ## Local - Home Office ## Benefícios - Familiar: Convênio SESC (gratuito); plano de saúde e odontológico (Bradesco TNQ2 sem co-participação) com Concierge. - Filhos: Auxílio-creche para crianças até 5 anos. - Bem-estar: Gympass (bora malhar!!!); horário flexível; trabalho remoto; dress code informal; seguro de vida; vale Transporte (se quiser, é opcional); vale refeição e/ou vale Alimentação e bicicletário - Mentoria e Engajamento:Super incentivamos idas a eventos e treinamentos, e se fizer sentido na estratégia, melhor ainda (qualquer um pode pedir); trabalhar com referências técnicas; eventos internos (hackathon, black friday,confraternizações das tribos, datas comemorativas) e aprendizado contínuo. ## Requisitos **Obrigatórios:** - Formação de nível superior em análise de sistemas, ciência da computação, engenharia de software ou pelo menos 8 anos de experiência em desenvolvimento de software. - Sólidos conhecimentos sobre algoritmos e estruturas de dados - Sólidos conhecimentos sobre padrões de projetos (Design Patterns) e SOLID - Sólidos conhecimentos da linguagem C# - Sólidos conhecimentos Restful APIs e o protocolo Http - Pelo menos 6 anos de experiência com desenvolvimento backend usando .Net Framework. - Pelo menos 2 anos de experiência com .Net Core. - Domínio de Docker e Conteinerização - Conhecimentos de testes de unidade, de integração e carga - Sólida experiência com sistemas distribuídos e mensageria e/ou eventos - Conhecimentos sobre algum message broker, tais como: RabbitMQ, ZeroMQ, Kafka, IBM Mq - Conhecimentos de abordagens como: Micro Serviços, Event Sourcing, CQRS, Hexagonal Architecture, Lambda Architecture - Experiência com modelagens Relacionais e não Relacionais - Experiência com análise e otimização de Queries SQL - Conhecimentos em OAuth, OpenID e JWT - Usar git de forma natural e conhecer o gitflow - Conhecimento de ferramentas de CI/CD, Azure DevOps é um diferencial aqui - Conhecimento de algum serviço de nuvem, Azure é um diferencial aqui - Familiaridade com metodologias ágeis e suas aplicações práticas, principalmente Kanban e/ou Scrum ## Contratação - a combinar ## Nossa empresa - PRAZER, SOMOS A STECH! O TIME DE TECNOLOGIA DA STONE: - Aqui na Stech, prezamos pela liberdade e autonomia. Liberdade para escolher seus horários, suas stacks, a forma como você vai trabalhar e mais importante, liberdade de ser quem você é. Você vai ter autonomia para fazer as suas escolhas e será cobrado por elas. Somos todos adultos e gostamos muito de um termo chamado “Accountability”, que podemos traduzir para responsabilidade pelos seus resultados. Queremos que você procure primeiro entender, para depois ser entendido. Mas não se preocupe, você sempre será ouvido, sua opinião conta e conta muito. Por último, e com muita importância para nós: Nós prezamos pela troca de conhecimento. “Be humble and kind”, esteja sempre de coração aberto para aprender novas formas de pensar e conhecimentos diferentes. Queremos achar em vocês inteligência, energia e integridade. E em troca, oferecemos autonomia, liberdade e flexibilidade. ## Como se candidatar - [Clique aqui para se candidatar](https://jornada.stone.com.br/formulario/stonetech?gh_jid=4316271003)
non_infrastructure
tech lead engenheiro de software backend c na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga na tesouraria stone co temos como missão manter nosso grupo saudável financeiramente arcando com todas as obrigações de forma eficiente além de garantir que todos os produtos que necessitam de recursos financeiros em suas operações sejam viáveis sem comprometer a liquidez do grupo isso é uma senhora responsabilidade operamos bilhões de reais diariamente para alcançarmos tal objetivo procuramos engenheiros de software criativos responsáveis com energia e que tenham foco em resolver o problema do cliente com qualidade também valorizamos suas habilidades práticas de desenvolvimento e capacidade de trabalhar bem em equipe é importante destacar também que você precisa gostar do “economês” já que o nosso produto tem grande relação com o mercado financeiro se você reúne essas características essa vaga é pra você atividades você irá projetar desenvolver testar implantar manter e melhorar o software você gerenciará as prioridades prazos e trabalhos individuais do projeto você terá que traduzir requisitos funcionais e técnicos em arquitetura e design detalhados de soluções direcionando o time na direção correta e com visão de longo prazo você precisará instruir o time em boas práticas de desenvolvimento padrões de projeto testes e revisão de código você irá ser um guardião das nossas boas práticas e documentações cultivando e desenvolvendo as com o time você irá promover uma cultura do time que vá além do código observando aspectos não funcionais da aplicação como performance observabilidade resiliência segurança mantendo o foco do time em relação a estes aspectos é esperado de você extrema facilidade de trabalho em equipe muita franqueza e transparência para que no dia a dia possa dar bons feedbacks e transmitir o conhecimento para os demais do time é esperado de você respeito e compromisso com o desenvolvimento dos demais colegas do time ajudando no direcionamento da carreira e na evolução constante local home office benefícios familiar convênio sesc gratuito plano de saúde e odontológico bradesco sem co participação com concierge filhos auxílio creche para crianças até anos bem estar gympass bora malhar horário flexível trabalho remoto dress code informal seguro de vida vale transporte se quiser é opcional vale refeição e ou vale alimentação e bicicletário mentoria e engajamento super incentivamos idas a eventos e treinamentos e se fizer sentido na estratégia melhor ainda qualquer um pode pedir trabalhar com referências técnicas eventos internos hackathon black friday confraternizações das tribos datas comemorativas e aprendizado contínuo requisitos obrigatórios formação de nível superior em análise de sistemas ciência da computação engenharia de software ou pelo menos anos de experiência em desenvolvimento de software sólidos conhecimentos sobre algoritmos e estruturas de dados sólidos conhecimentos sobre padrões de projetos design patterns e solid sólidos conhecimentos da linguagem c sólidos conhecimentos restful apis e o protocolo http pelo menos anos de experiência com desenvolvimento backend usando net framework pelo menos anos de experiência com net core domínio de docker e conteinerização conhecimentos de testes de unidade de integração e carga sólida experiência com sistemas distribuídos e mensageria e ou eventos conhecimentos sobre algum message broker tais como rabbitmq zeromq kafka ibm mq conhecimentos de abordagens como micro serviços event sourcing cqrs hexagonal architecture lambda architecture experiência com modelagens relacionais e não relacionais experiência com análise e otimização de queries sql conhecimentos em oauth openid e jwt usar git de forma natural e conhecer o gitflow conhecimento de ferramentas de ci cd azure devops é um diferencial aqui conhecimento de algum serviço de nuvem azure é um diferencial aqui familiaridade com metodologias ágeis e suas aplicações práticas principalmente kanban e ou scrum contratação a combinar nossa empresa prazer somos a stech o time de tecnologia da stone aqui na stech prezamos pela liberdade e autonomia liberdade para escolher seus horários suas stacks a forma como você vai trabalhar e mais importante liberdade de ser quem você é você vai ter autonomia para fazer as suas escolhas e será cobrado por elas somos todos adultos e gostamos muito de um termo chamado “accountability” que podemos traduzir para responsabilidade pelos seus resultados queremos que você procure primeiro entender para depois ser entendido mas não se preocupe você sempre será ouvido sua opinião conta e conta muito por último e com muita importância para nós nós prezamos pela troca de conhecimento “be humble and kind” esteja sempre de coração aberto para aprender novas formas de pensar e conhecimentos diferentes queremos achar em vocês inteligência energia e integridade e em troca oferecemos autonomia liberdade e flexibilidade como se candidatar
0
11,926
9,526,285,052
IssuesEvent
2019-04-28 18:48:11
BenjaVR/SharepointDirectInvoke-ChromeExtension
https://api.github.com/repos/BenjaVR/SharepointDirectInvoke-ChromeExtension
closed
Webpack watch no terminal output when no files are changed
bug infrastructure
When the webpack watch is triggered without real changes (e.g. just pressing save), the terminal output is empty.
1.0
Webpack watch no terminal output when no files are changed - When the webpack watch is triggered without real changes (e.g. just pressing save), the terminal output is empty.
infrastructure
webpack watch no terminal output when no files are changed when the webpack watch is triggered without real changes e g just pressing save the terminal output is empty
1
72,050
8,698,782,882
IssuesEvent
2018-12-05 00:59:17
chapel-lang/chapel
https://api.github.com/repos/chapel-lang/chapel
opened
Bug or feature: config consts can change type based on `-s` flags
area: Compiler area: Language type: Design
When a `config const` or `config var` has its value set at compile-time, the expression it's set to is essentially dropped into the Chapel code as its initializing expression. This can result in surprises. For example, given: ```chapel config const n = 1.5; ``` if the user compiles with `-sn=100` then `n` will be inferred to be an integer rather than a real. This is part of the power of setting configs at compile-time (general expressions can be put into the code), but also represents a potential point of confusion or errors (the programmer obviously thought of `n` as a real). This could be prevented by coding defensively: ```chapel config const n: real = 1.5; ``` but it seems unfortunate to recommend that programmers use types in this way for safety in a context where they're often left out for conciseness. On the other hand, this behavior could also be considered a powerful feature since to change a 64-bit config real into a 32-bit real, I could compile with `-sn=1.5:real(32)` rather than the more involved: ```chapel config type ntype = real; config const n: ntype = 1.5; ``` and then compiling with -sntype=real(32).
1.0
Bug or feature: config consts can change type based on `-s` flags - When a `config const` or `config var` has its value set at compile-time, the expression it's set to is essentially dropped into the Chapel code as its initializing expression. This can result in surprises. For example, given: ```chapel config const n = 1.5; ``` if the user compiles with `-sn=100` then `n` will be inferred to be an integer rather than a real. This is part of the power of setting configs at compile-time (general expressions can be put into the code), but also represents a potential point of confusion or errors (the programmer obviously thought of `n` as a real). This could be prevented by coding defensively: ```chapel config const n: real = 1.5; ``` but it seems unfortunate to recommend that programmers use types in this way for safety in a context where they're often left out for conciseness. On the other hand, this behavior could also be considered a powerful feature since to change a 64-bit config real into a 32-bit real, I could compile with `-sn=1.5:real(32)` rather than the more involved: ```chapel config type ntype = real; config const n: ntype = 1.5; ``` and then compiling with -sntype=real(32).
non_infrastructure
bug or feature config consts can change type based on s flags when a config const or config var has its value set at compile time the expression it s set to is essentially dropped into the chapel code as its initializing expression this can result in surprises for example given chapel config const n if the user compiles with sn then n will be inferred to be an integer rather than a real this is part of the power of setting configs at compile time general expressions can be put into the code but also represents a potential point of confusion or errors the programmer obviously thought of n as a real this could be prevented by coding defensively chapel config const n real but it seems unfortunate to recommend that programmers use types in this way for safety in a context where they re often left out for conciseness on the other hand this behavior could also be considered a powerful feature since to change a bit config real into a bit real i could compile with sn real rather than the more involved chapel config type ntype real config const n ntype and then compiling with sntype real
0
7,902
4,102,394,750
IssuesEvent
2016-06-04 00:50:43
jeff1evesque/machine-learning
https://api.github.com/repos/jeff1evesque/machine-learning
closed
Move arguments in 'setup_tables.py' into 'settings.yaml'
build enhancement
We will move the arguments used for populating `tbl_model_type` into `settings.yaml`. Then, we will respectively reference the yaml attribute within `setup_tables.py`.
1.0
Move arguments in 'setup_tables.py' into 'settings.yaml' - We will move the arguments used for populating `tbl_model_type` into `settings.yaml`. Then, we will respectively reference the yaml attribute within `setup_tables.py`.
non_infrastructure
move arguments in setup tables py into settings yaml we will move the arguments used for populating tbl model type into settings yaml then we will respectively reference the yaml attribute within setup tables py
0
28,870
23,543,008,130
IssuesEvent
2022-08-20 17:48:36
jrsmith3/ibei
https://api.github.com/repos/jrsmith3/ibei
opened
Write GitHub action to post documentation to readthedocs.org for new releases
development infrastructure
# Overview The scope of this issue is to upload the documentation build by the automation described in #55 to readthedocs.org via GitHub action. Documentation should only be updated in this way for releases. # Related issues * Depends on #55.
1.0
Write GitHub action to post documentation to readthedocs.org for new releases - # Overview The scope of this issue is to upload the documentation build by the automation described in #55 to readthedocs.org via GitHub action. Documentation should only be updated in this way for releases. # Related issues * Depends on #55.
infrastructure
write github action to post documentation to readthedocs org for new releases overview the scope of this issue is to upload the documentation build by the automation described in to readthedocs org via github action documentation should only be updated in this way for releases related issues depends on
1
15,585
11,597,061,873
IssuesEvent
2020-02-24 20:06:19
dotnet/docker-tools
https://api.github.com/repos/dotnet/docker-tools
closed
Create Azure Key Vault for managing .NET Docker secrets
area:infrastructure enhancement triaged
This will allow the Docker team to have the necessary permissions to more easily manage secrets instead of relying on Eng. Services.
1.0
Create Azure Key Vault for managing .NET Docker secrets - This will allow the Docker team to have the necessary permissions to more easily manage secrets instead of relying on Eng. Services.
infrastructure
create azure key vault for managing net docker secrets this will allow the docker team to have the necessary permissions to more easily manage secrets instead of relying on eng services
1
8,916
7,734,097,142
IssuesEvent
2018-05-26 19:54:10
Khan/KaTeX
https://api.github.com/repos/Khan/KaTeX
opened
Update release.sh to update URLs in contrib/*/README.md
infrastructure
https://github.com/Khan/KaTeX/pull/1362 does this manually, but it should be automated.
1.0
Update release.sh to update URLs in contrib/*/README.md - https://github.com/Khan/KaTeX/pull/1362 does this manually, but it should be automated.
infrastructure
update release sh to update urls in contrib readme md does this manually but it should be automated
1
3,592
3,964,700,890
IssuesEvent
2016-05-03 02:49:39
eloquent/phony
https://api.github.com/repos/eloquent/phony
closed
Optimization ideas based on cachegrind analysis
performance
- [x] `MockDefinition::isEqualTo()` is called many times. The main culprit is `MockFactory`. Avoiding method calls in this area improves performance notably. - [x] <del>`Arguments` wrapper is created many times. It may be possible to do without it entirely.</del> too hard - [x] <del>`Invoker::callWith()` is used everywhere. Reverting back to `call_user_func_array()` where possible should improve performance.</del> too hard - [x] Replace all usage of `Arguments::adapt()` with manual checks. - [ ] `Stub` calls other stub methods for most common operations. Reducing method calls where possible should help.
True
Optimization ideas based on cachegrind analysis - - [x] `MockDefinition::isEqualTo()` is called many times. The main culprit is `MockFactory`. Avoiding method calls in this area improves performance notably. - [x] <del>`Arguments` wrapper is created many times. It may be possible to do without it entirely.</del> too hard - [x] <del>`Invoker::callWith()` is used everywhere. Reverting back to `call_user_func_array()` where possible should improve performance.</del> too hard - [x] Replace all usage of `Arguments::adapt()` with manual checks. - [ ] `Stub` calls other stub methods for most common operations. Reducing method calls where possible should help.
non_infrastructure
optimization ideas based on cachegrind analysis mockdefinition isequalto is called many times the main culprit is mockfactory avoiding method calls in this area improves performance notably arguments wrapper is created many times it may be possible to do without it entirely too hard invoker callwith is used everywhere reverting back to call user func array where possible should improve performance too hard replace all usage of arguments adapt with manual checks stub calls other stub methods for most common operations reducing method calls where possible should help
0
21,712
14,737,048,674
IssuesEvent
2021-01-07 00:43:24
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
SA Hosted Billing cycle
anc-core anc-infrastructure anc-process anp-important ant-bug ant-support
In GitLab by @kdjstudios on Apr 10, 2018, 12:20 **Submitted by:** "Michelle McKee" <michelle.mckee@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-10-29782/conversation **Server:** Internal **Client/Site:** Sa Hosted **Account:** NA **Issue:** I am trying to do my 4/1/18 Billing cycle for SA Hosted. I tried to do it once. Got as far as “Apply” and when I reviewed the invoices I found a mistake so I reverted the 4/1/18 billing cycle. I went into the site and edited the “Last billing cycle date:” back to 3/1/18. When I went back in to rerun the 4/1/18 billing cycle, it still says that the billing cycle is 5/1/18. Shouldn’t this have changed?
1.0
SA Hosted Billing cycle - In GitLab by @kdjstudios on Apr 10, 2018, 12:20 **Submitted by:** "Michelle McKee" <michelle.mckee@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-10-29782/conversation **Server:** Internal **Client/Site:** Sa Hosted **Account:** NA **Issue:** I am trying to do my 4/1/18 Billing cycle for SA Hosted. I tried to do it once. Got as far as “Apply” and when I reviewed the invoices I found a mistake so I reverted the 4/1/18 billing cycle. I went into the site and edited the “Last billing cycle date:” back to 3/1/18. When I went back in to rerun the 4/1/18 billing cycle, it still says that the billing cycle is 5/1/18. Shouldn’t this have changed?
infrastructure
sa hosted billing cycle in gitlab by kdjstudios on apr submitted by michelle mckee helpdesk server internal client site sa hosted account na issue i am trying to do my billing cycle for sa hosted i tried to do it once got as far as “apply” and when i reviewed the invoices i found a mistake so i reverted the billing cycle i went into the site and edited the “last billing cycle date ” back to when i went back in to rerun the billing cycle it still says that the billing cycle is shouldn’t this have changed
1
85,515
3,691,158,468
IssuesEvent
2016-02-25 22:45:38
matheusportela/Multiagent-RL
https://api.github.com/repos/matheusportela/Multiagent-RL
closed
Refactor out features model.
priority: major type: refactor
Currently, all features are implemented in `agents` module, which is simply wrong. Let's create a new module for it.
1.0
Refactor out features model. - Currently, all features are implemented in `agents` module, which is simply wrong. Let's create a new module for it.
non_infrastructure
refactor out features model currently all features are implemented in agents module which is simply wrong let s create a new module for it
0
30,522
24,892,240,102
IssuesEvent
2022-10-28 13:02:07
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
opened
WooCommerce legacy assets missing SCSS files after being packaged
tool: monorepo infrastructure
### Prerequisites - [X] I have carried out troubleshooting steps and I believe I have found a bug. - [X] I have searched for similar bugs in both open and closed issues and cannot find a duplicate. ### Describe the bug When running `pnpm run build:zip` for WooCommerce, `SCSS` files are missing from the build package of `assets/css`. ### Expected behavior `SCSS` files should still be there. ### Actual behavior `SCSS` files are missing. ### Steps to reproduce 1. `pnpm install` 2. `cd plugins/woocommerce` 3. `pnpm run build:zip` 4. Unzip the package and look in `assets/css`. `SCSS` files are missing. ### WordPress Environment N/A ### Isolating the problem - [X] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [X] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [X] I can reproduce this bug consistently using the steps above.
1.0
WooCommerce legacy assets missing SCSS files after being packaged - ### Prerequisites - [X] I have carried out troubleshooting steps and I believe I have found a bug. - [X] I have searched for similar bugs in both open and closed issues and cannot find a duplicate. ### Describe the bug When running `pnpm run build:zip` for WooCommerce, `SCSS` files are missing from the build package of `assets/css`. ### Expected behavior `SCSS` files should still be there. ### Actual behavior `SCSS` files are missing. ### Steps to reproduce 1. `pnpm install` 2. `cd plugins/woocommerce` 3. `pnpm run build:zip` 4. Unzip the package and look in `assets/css`. `SCSS` files are missing. ### WordPress Environment N/A ### Isolating the problem - [X] I have deactivated other plugins and confirmed this bug occurs when only WooCommerce plugin is active. - [X] This bug happens with a default WordPress theme active, or [Storefront](https://woocommerce.com/storefront/). - [X] I can reproduce this bug consistently using the steps above.
infrastructure
woocommerce legacy assets missing scss files after being packaged prerequisites i have carried out troubleshooting steps and i believe i have found a bug i have searched for similar bugs in both open and closed issues and cannot find a duplicate describe the bug when running pnpm run build zip for woocommerce scss files are missing from the build package of assets css expected behavior scss files should still be there actual behavior scss files are missing steps to reproduce pnpm install cd plugins woocommerce pnpm run build zip unzip the package and look in assets css scss files are missing wordpress environment n a isolating the problem i have deactivated other plugins and confirmed this bug occurs when only woocommerce plugin is active this bug happens with a default wordpress theme active or i can reproduce this bug consistently using the steps above
1
130,403
18,155,785,215
IssuesEvent
2021-09-27 01:14:33
benlazarine/cas-overlay
https://api.github.com/repos/benlazarine/cas-overlay
opened
CVE-2020-9546 (High) detected in jackson-databind-2.9.5.jar
security vulnerability
## CVE-2020-9546 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: cas-overlay/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.hadoop.shaded.com.zaxxer.hikari.HikariConfig (aka shaded hikari-config). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9546>CVE-2020-9546</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9546">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9546</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-9546 (High) detected in jackson-databind-2.9.5.jar - ## CVE-2020-9546 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: cas-overlay/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.hadoop.shaded.com.zaxxer.hikari.HikariConfig (aka shaded hikari-config). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9546>CVE-2020-9546</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9546">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9546</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file cas overlay pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy cas server support oauth webflow jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache hadoop shaded com zaxxer hikari hikariconfig aka shaded hikari config publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
173,726
13,439,362,901
IssuesEvent
2020-09-07 20:51:05
npdarrington/Whats-Cookin
https://api.github.com/repos/npdarrington/Whats-Cookin
closed
Recipe Get Total Cost of Recipe TDD
iteration: 2️⃣ type: enhancement 🌟 type: test 🔮
As a user, I want to be able to see how much a recipe cost to make. - [ ] Create a method that calculates all ingredients and their quantity for a given recipe and return to user. - [ ] Happy/Sad path.
1.0
Recipe Get Total Cost of Recipe TDD - As a user, I want to be able to see how much a recipe cost to make. - [ ] Create a method that calculates all ingredients and their quantity for a given recipe and return to user. - [ ] Happy/Sad path.
non_infrastructure
recipe get total cost of recipe tdd as a user i want to be able to see how much a recipe cost to make create a method that calculates all ingredients and their quantity for a given recipe and return to user happy sad path
0
155,663
5,958,887,363
IssuesEvent
2017-05-29 09:20:14
FACN1/easyNaz
https://api.github.com/repos/FACN1/easyNaz
closed
add functionality to disability page
awaiting-review Priority-1 T2h technical task
- Toggle selection to local storage - When hit next, go to category page - add/remove selected filters to the screen when clicked
1.0
add functionality to disability page - - Toggle selection to local storage - When hit next, go to category page - add/remove selected filters to the screen when clicked
non_infrastructure
add functionality to disability page toggle selection to local storage when hit next go to category page add remove selected filters to the screen when clicked
0
4,642
5,206,249,736
IssuesEvent
2017-01-24 20:05:32
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Linux SDK is not being built with sysroot
area-infrastructure P1 high
The SDK building bots are not using the new GN options to build with wheezy sysroots, but are using the old method of setting c++ environment variables to --sysroot... and so on. The new SDL builds therefore need glibc 2.16 and will not run on precise builders or other machines without glibc 2.16. The way to fix this is to add the right environment variable to the SDK builder recipe. Since we uploaded a 1.22.0-dev.9.0 SDK to cloud storage to be downloaded to the repo when checking out, this is now hitting our 32-bit dartium builder. I will revert the SDK download change, and land it with a new version without the glibc problem. We need to fix this immediately, before sending out another dev version and before building stable.
1.0
Linux SDK is not being built with sysroot - The SDK building bots are not using the new GN options to build with wheezy sysroots, but are using the old method of setting c++ environment variables to --sysroot... and so on. The new SDL builds therefore need glibc 2.16 and will not run on precise builders or other machines without glibc 2.16. The way to fix this is to add the right environment variable to the SDK builder recipe. Since we uploaded a 1.22.0-dev.9.0 SDK to cloud storage to be downloaded to the repo when checking out, this is now hitting our 32-bit dartium builder. I will revert the SDK download change, and land it with a new version without the glibc problem. We need to fix this immediately, before sending out another dev version and before building stable.
infrastructure
linux sdk is not being built with sysroot the sdk building bots are not using the new gn options to build with wheezy sysroots but are using the old method of setting c environment variables to sysroot and so on the new sdl builds therefore need glibc and will not run on precise builders or other machines without glibc the way to fix this is to add the right environment variable to the sdk builder recipe since we uploaded a dev sdk to cloud storage to be downloaded to the repo when checking out this is now hitting our bit dartium builder i will revert the sdk download change and land it with a new version without the glibc problem we need to fix this immediately before sending out another dev version and before building stable
1
72,354
9,575,555,677
IssuesEvent
2019-05-07 06:48:12
ISS-Security/credence-api
https://api.github.com/repos/ISS-Security/credence-api
closed
Typos in readme.md
documentation
There may have some typos in readme. It should be - GET api/v1/document`s`/: returns all confiugration IDs - GET api/v1/document`s`/[ID]: returns details about a single document with given ID - POST api/v1/document`s`/: creates a new document instead of `document` in all requests.
1.0
Typos in readme.md - There may have some typos in readme. It should be - GET api/v1/document`s`/: returns all confiugration IDs - GET api/v1/document`s`/[ID]: returns details about a single document with given ID - POST api/v1/document`s`/: creates a new document instead of `document` in all requests.
non_infrastructure
typos in readme md there may have some typos in readme it should be get api document s returns all confiugration ids get api document s returns details about a single document with given id post api document s creates a new document instead of document in all requests
0
450,610
31,932,402,986
IssuesEvent
2023-09-19 08:19:37
owid/etl
https://api.github.com/repos/owid/etl
closed
docs: Running `mkdocs serve` fails because of `api/python.ipynb`
documentation priority 3 - nice to have
### One-liner Running `mkdocs serve` fails with an error. The error disappears when removing the file `api/python.ipynb`. ### Context & details Full log when running `mkdocs serve`: ``` INFO - DeprecationWarning: Jupyter is migrating its paths to use standard platformdirs given by the platformdirs library. To remove this warning and see the appropriate new directories, set the environment variable `JUPYTER_PLATFORM_DIRS=1` and then run `jupyter --paths`. The use of platformdirs will be the default in `jupyter_core` v6 File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_core/utils/__init__.py", line 89, in deprecation warnings.warn(message, DeprecationWarning, stacklevel=stacklevel + 1) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_client/connect.py", line 20, in from jupyter_core.paths import jupyter_data_dir, jupyter_runtime_dir, secure_write INFO - Building documentation... INFO - Cleaning site directory INFO - The following pages exist in the docs directory, but are not included in the "nav" configuration: - old-readme.md - architecture/design/index.md - architecture/workflow/publish.md - architecture/workflow/other-steps/index.md INFO - Doc file 'index.md' contains an unrecognized relative link 'api/', it was left as is. Did you mean 'api/index.md'? WARNING - Doc file 'old-readme.md' contains a relative link 'docs/future-architecture.png', but the target is not found among documentation files. INFO - Converting notebook (execute=False): /Users/prosado/Documents/owid/repos/etl/docs/api/python.ipynb INFO - DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead. File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/convert.py", line 16, in add_anchor_lower_id from defusedxml import cElementTree as ElementTree File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/defusedxml/cElementTree.py", line 38, in <module> warnings.warn( ERROR - Error reading page 'api/python.ipynb': No module named 'jupyter_contrib_nbextensions' Traceback (most recent call last): File "/Users/prosado/Documents/owid/repos/etl/.venv/bin/mkdocs", line 8, in <module> sys.exit(cli()) ^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1157, in __call__ return self.main(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1078, in main rv = self.invoke(ctx) ^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1688, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1434, in invoke return ctx.invoke(self.callback, **ctx.params) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 783, in invoke return __callback(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/__main__.py", line 270, in serve_command serve.serve(**kwargs) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 86, in serve builder(config) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 67, in builder build(config, live_server=None if is_clean else server, dirty=is_dirty) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/build.py", line 322, in build _populate_page(file.page, config, files, dirty) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/build.py", line 175, in _populate_page page.render(config, files) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/plugin.py", line 110, in new_render toc, title = get_nb_toc(page.file.abs_src_path) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/plugin.py", line 147, in get_nb_toc body = convert.nb2md(fpath) ^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/nbconvert2.py", line 144, in nb2md exporter = MarkdownExporter( ^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/templateexporter.py", line 353, in __init__ super().__init__(config=config, **kw) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/exporter.py", line 117, in __init__ self._init_preprocessors() File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/templateexporter.py", line 519, in _init_preprocessors super()._init_preprocessors() File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/exporter.py", line 278, in _init_preprocessors self.register_preprocessor(preprocessor, enabled=True) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/exporter.py", line 238, in register_preprocessor preprocessor_cls = import_item(preprocessor) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/traitlets/utils/importstring.py", line 30, in import_item module = __import__(package, fromlist=[obj]) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ModuleNotFoundError: No module named 'jupyter_contrib_nbextensions' ``` After that, I did `poetry add jupyter_contrib_nbextensions` and tried again `mkdocs serve`, and now I get a different error. Full log: ``` INFO - DeprecationWarning: Jupyter is migrating its paths to use standard platformdirs given by the platformdirs library. To remove this warning and see the appropriate new directories, set the environment variable `JUPYTER_PLATFORM_DIRS=1` and then run `jupyter --paths`. The use of platformdirs will be the default in `jupyter_core` v6 File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_core/utils/__init__.py", line 89, in deprecation warnings.warn(message, DeprecationWarning, stacklevel=stacklevel + 1) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_client/connect.py", line 20, in from jupyter_core.paths import jupyter_data_dir, jupyter_runtime_dir, secure_write Traceback (most recent call last): File "/Users/prosado/Documents/owid/repos/etl/.venv/bin/mkdocs", line 8, in <module> sys.exit(cli()) ^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1157, in __call__ return self.main(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1078, in main rv = self.invoke(ctx) ^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1688, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1434, in invoke return ctx.invoke(self.callback, **ctx.params) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 783, in invoke return __callback(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/__main__.py", line 270, in serve_command serve.serve(**kwargs) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 59, in serve config = get_config() ^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 47, in get_config config = load_config( ^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/base.py", line 378, in load_config errors, warnings = cfg.validate() ^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/base.py", line 230, in validate run_failed, run_warnings = self._validate() ^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/base.py", line 188, in _validate self[key] = config_option.validate(value) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 182, in validate return self.run_validation(value) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 1064, in run_validation self.load_plugin_with_namespace(name, cfg) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 1100, in load_plugin_with_namespace return (name, self.load_plugin(name, config)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 1118, in load_plugin plugin_cls = self.installed_plugins[name].load() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/homebrew/Cellar/python@3.11/3.11.2_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/importlib/metadata/__init__.py", line 202, in load module = import_module(match.group('module')) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/homebrew/Cellar/python@3.11/3.11.2_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "<frozen importlib._bootstrap>", line 1206, in _gcd_import File "<frozen importlib._bootstrap>", line 1178, in _find_and_load File "<frozen importlib._bootstrap>", line 1149, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 690, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 940, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/plugin.py", line 12, in <module> from . import convert File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/convert.py", line 51, in <module> from .nbconvert2 import * File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/nbconvert2.py", line 20, in <module> from nbconvert.nbconvertapp import NbConvertApp File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/nbconvertapp.py", line 194, in <module> class NbConvertApp(JupyterApp): File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/nbconvertapp.py", line 287, in NbConvertApp formats=get_export_names() ^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/base.py", line 149, in get_export_names e = get_exporter(exporter_name)(config=config) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/base.py", line 100, in get_exporter exporter = entrypoints.get_single("nbconvert.exporters", name).load() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/entrypoints.py", line 79, in load mod = import_module(self.module_name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/homebrew/Cellar/python@3.11/3.11.2_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_contrib_nbextensions/__init__.py", line 5, in <module> import jupyter_nbextensions_configurator File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_nbextensions_configurator/__init__.py", line 18, in <module> from notebook.base.handlers import APIHandler, IPythonHandler ModuleNotFoundError: No module named 'notebook.base' ``` The workaround is to delete the file `docs/api/python.ipynb`.
1.0
docs: Running `mkdocs serve` fails because of `api/python.ipynb` - ### One-liner Running `mkdocs serve` fails with an error. The error disappears when removing the file `api/python.ipynb`. ### Context & details Full log when running `mkdocs serve`: ``` INFO - DeprecationWarning: Jupyter is migrating its paths to use standard platformdirs given by the platformdirs library. To remove this warning and see the appropriate new directories, set the environment variable `JUPYTER_PLATFORM_DIRS=1` and then run `jupyter --paths`. The use of platformdirs will be the default in `jupyter_core` v6 File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_core/utils/__init__.py", line 89, in deprecation warnings.warn(message, DeprecationWarning, stacklevel=stacklevel + 1) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_client/connect.py", line 20, in from jupyter_core.paths import jupyter_data_dir, jupyter_runtime_dir, secure_write INFO - Building documentation... INFO - Cleaning site directory INFO - The following pages exist in the docs directory, but are not included in the "nav" configuration: - old-readme.md - architecture/design/index.md - architecture/workflow/publish.md - architecture/workflow/other-steps/index.md INFO - Doc file 'index.md' contains an unrecognized relative link 'api/', it was left as is. Did you mean 'api/index.md'? WARNING - Doc file 'old-readme.md' contains a relative link 'docs/future-architecture.png', but the target is not found among documentation files. INFO - Converting notebook (execute=False): /Users/prosado/Documents/owid/repos/etl/docs/api/python.ipynb INFO - DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead. File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/convert.py", line 16, in add_anchor_lower_id from defusedxml import cElementTree as ElementTree File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/defusedxml/cElementTree.py", line 38, in <module> warnings.warn( ERROR - Error reading page 'api/python.ipynb': No module named 'jupyter_contrib_nbextensions' Traceback (most recent call last): File "/Users/prosado/Documents/owid/repos/etl/.venv/bin/mkdocs", line 8, in <module> sys.exit(cli()) ^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1157, in __call__ return self.main(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1078, in main rv = self.invoke(ctx) ^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1688, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1434, in invoke return ctx.invoke(self.callback, **ctx.params) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 783, in invoke return __callback(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/__main__.py", line 270, in serve_command serve.serve(**kwargs) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 86, in serve builder(config) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 67, in builder build(config, live_server=None if is_clean else server, dirty=is_dirty) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/build.py", line 322, in build _populate_page(file.page, config, files, dirty) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/build.py", line 175, in _populate_page page.render(config, files) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/plugin.py", line 110, in new_render toc, title = get_nb_toc(page.file.abs_src_path) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/plugin.py", line 147, in get_nb_toc body = convert.nb2md(fpath) ^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/nbconvert2.py", line 144, in nb2md exporter = MarkdownExporter( ^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/templateexporter.py", line 353, in __init__ super().__init__(config=config, **kw) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/exporter.py", line 117, in __init__ self._init_preprocessors() File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/templateexporter.py", line 519, in _init_preprocessors super()._init_preprocessors() File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/exporter.py", line 278, in _init_preprocessors self.register_preprocessor(preprocessor, enabled=True) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/exporter.py", line 238, in register_preprocessor preprocessor_cls = import_item(preprocessor) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/traitlets/utils/importstring.py", line 30, in import_item module = __import__(package, fromlist=[obj]) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ModuleNotFoundError: No module named 'jupyter_contrib_nbextensions' ``` After that, I did `poetry add jupyter_contrib_nbextensions` and tried again `mkdocs serve`, and now I get a different error. Full log: ``` INFO - DeprecationWarning: Jupyter is migrating its paths to use standard platformdirs given by the platformdirs library. To remove this warning and see the appropriate new directories, set the environment variable `JUPYTER_PLATFORM_DIRS=1` and then run `jupyter --paths`. The use of platformdirs will be the default in `jupyter_core` v6 File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_core/utils/__init__.py", line 89, in deprecation warnings.warn(message, DeprecationWarning, stacklevel=stacklevel + 1) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_client/connect.py", line 20, in from jupyter_core.paths import jupyter_data_dir, jupyter_runtime_dir, secure_write Traceback (most recent call last): File "/Users/prosado/Documents/owid/repos/etl/.venv/bin/mkdocs", line 8, in <module> sys.exit(cli()) ^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1157, in __call__ return self.main(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1078, in main rv = self.invoke(ctx) ^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1688, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 1434, in invoke return ctx.invoke(self.callback, **ctx.params) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/click/core.py", line 783, in invoke return __callback(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/__main__.py", line 270, in serve_command serve.serve(**kwargs) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 59, in serve config = get_config() ^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/commands/serve.py", line 47, in get_config config = load_config( ^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/base.py", line 378, in load_config errors, warnings = cfg.validate() ^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/base.py", line 230, in validate run_failed, run_warnings = self._validate() ^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/base.py", line 188, in _validate self[key] = config_option.validate(value) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 182, in validate return self.run_validation(value) ^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 1064, in run_validation self.load_plugin_with_namespace(name, cfg) File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 1100, in load_plugin_with_namespace return (name, self.load_plugin(name, config)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs/config/config_options.py", line 1118, in load_plugin plugin_cls = self.installed_plugins[name].load() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/homebrew/Cellar/python@3.11/3.11.2_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/importlib/metadata/__init__.py", line 202, in load module = import_module(match.group('module')) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/homebrew/Cellar/python@3.11/3.11.2_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "<frozen importlib._bootstrap>", line 1206, in _gcd_import File "<frozen importlib._bootstrap>", line 1178, in _find_and_load File "<frozen importlib._bootstrap>", line 1149, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 690, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 940, in exec_module File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/plugin.py", line 12, in <module> from . import convert File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/convert.py", line 51, in <module> from .nbconvert2 import * File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/mkdocs_jupyter/nbconvert2.py", line 20, in <module> from nbconvert.nbconvertapp import NbConvertApp File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/nbconvertapp.py", line 194, in <module> class NbConvertApp(JupyterApp): File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/nbconvertapp.py", line 287, in NbConvertApp formats=get_export_names() ^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/base.py", line 149, in get_export_names e = get_exporter(exporter_name)(config=config) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/nbconvert/exporters/base.py", line 100, in get_exporter exporter = entrypoints.get_single("nbconvert.exporters", name).load() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/entrypoints.py", line 79, in load mod = import_module(self.module_name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/homebrew/Cellar/python@3.11/3.11.2_1/Frameworks/Python.framework/Versions/3.11/lib/python3.11/importlib/__init__.py", line 126, in import_module return _bootstrap._gcd_import(name[level:], package, level) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_contrib_nbextensions/__init__.py", line 5, in <module> import jupyter_nbextensions_configurator File "/Users/prosado/Documents/owid/repos/etl/.venv/lib/python3.11/site-packages/jupyter_nbextensions_configurator/__init__.py", line 18, in <module> from notebook.base.handlers import APIHandler, IPythonHandler ModuleNotFoundError: No module named 'notebook.base' ``` The workaround is to delete the file `docs/api/python.ipynb`.
non_infrastructure
docs running mkdocs serve fails because of api python ipynb one liner running mkdocs serve fails with an error the error disappears when removing the file api python ipynb context details full log when running mkdocs serve info deprecationwarning jupyter is migrating its paths to use standard platformdirs given by the platformdirs library to remove this warning and see the appropriate new directories set the environment variable jupyter platform dirs and then run jupyter paths the use of platformdirs will be the default in jupyter core file users prosado documents owid repos etl venv lib site packages jupyter core utils init py line in deprecation warnings warn message deprecationwarning stacklevel stacklevel file users prosado documents owid repos etl venv lib site packages jupyter client connect py line in from jupyter core paths import jupyter data dir jupyter runtime dir secure write info building documentation info cleaning site directory info the following pages exist in the docs directory but are not included in the nav configuration old readme md architecture design index md architecture workflow publish md architecture workflow other steps index md info doc file index md contains an unrecognized relative link api it was left as is did you mean api index md warning doc file old readme md contains a relative link docs future architecture png but the target is not found among documentation files info converting notebook execute false users prosado documents owid repos etl docs api python ipynb info deprecationwarning defusedxml celementtree is deprecated import from defusedxml elementtree instead file users prosado documents owid repos etl venv lib site packages mkdocs jupyter convert py line in add anchor lower id from defusedxml import celementtree as elementtree file users prosado documents owid repos etl venv lib site packages defusedxml celementtree py line in warnings warn error error reading page api python ipynb no module named jupyter contrib nbextensions traceback most recent call last file users prosado documents owid repos etl venv bin mkdocs line in sys exit cli file users prosado documents owid repos etl venv lib site packages click core py line in call return self main args kwargs file users prosado documents owid repos etl venv lib site packages click core py line in main rv self invoke ctx file users prosado documents owid repos etl venv lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file users prosado documents owid repos etl venv lib site packages click core py line in invoke return ctx invoke self callback ctx params file users prosado documents owid repos etl venv lib site packages click core py line in invoke return callback args kwargs file users prosado documents owid repos etl venv lib site packages mkdocs main py line in serve command serve serve kwargs file users prosado documents owid repos etl venv lib site packages mkdocs commands serve py line in serve builder config file users prosado documents owid repos etl venv lib site packages mkdocs commands serve py line in builder build config live server none if is clean else server dirty is dirty file users prosado documents owid repos etl venv lib site packages mkdocs commands build py line in build populate page file page config files dirty file users prosado documents owid repos etl venv lib site packages mkdocs commands build py line in populate page page render config files file users prosado documents owid repos etl venv lib site packages mkdocs jupyter plugin py line in new render toc title get nb toc page file abs src path file users prosado documents owid repos etl venv lib site packages mkdocs jupyter plugin py line in get nb toc body convert fpath file users prosado documents owid repos etl venv lib site packages mkdocs jupyter py line in exporter markdownexporter file users prosado documents owid repos etl venv lib site packages nbconvert exporters templateexporter py line in init super init config config kw file users prosado documents owid repos etl venv lib site packages nbconvert exporters exporter py line in init self init preprocessors file users prosado documents owid repos etl venv lib site packages nbconvert exporters templateexporter py line in init preprocessors super init preprocessors file users prosado documents owid repos etl venv lib site packages nbconvert exporters exporter py line in init preprocessors self register preprocessor preprocessor enabled true file users prosado documents owid repos etl venv lib site packages nbconvert exporters exporter py line in register preprocessor preprocessor cls import item preprocessor file users prosado documents owid repos etl venv lib site packages traitlets utils importstring py line in import item module import package fromlist modulenotfounderror no module named jupyter contrib nbextensions after that i did poetry add jupyter contrib nbextensions and tried again mkdocs serve and now i get a different error full log info deprecationwarning jupyter is migrating its paths to use standard platformdirs given by the platformdirs library to remove this warning and see the appropriate new directories set the environment variable jupyter platform dirs and then run jupyter paths the use of platformdirs will be the default in jupyter core file users prosado documents owid repos etl venv lib site packages jupyter core utils init py line in deprecation warnings warn message deprecationwarning stacklevel stacklevel file users prosado documents owid repos etl venv lib site packages jupyter client connect py line in from jupyter core paths import jupyter data dir jupyter runtime dir secure write traceback most recent call last file users prosado documents owid repos etl venv bin mkdocs line in sys exit cli file users prosado documents owid repos etl venv lib site packages click core py line in call return self main args kwargs file users prosado documents owid repos etl venv lib site packages click core py line in main rv self invoke ctx file users prosado documents owid repos etl venv lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file users prosado documents owid repos etl venv lib site packages click core py line in invoke return ctx invoke self callback ctx params file users prosado documents owid repos etl venv lib site packages click core py line in invoke return callback args kwargs file users prosado documents owid repos etl venv lib site packages mkdocs main py line in serve command serve serve kwargs file users prosado documents owid repos etl venv lib site packages mkdocs commands serve py line in serve config get config file users prosado documents owid repos etl venv lib site packages mkdocs commands serve py line in get config config load config file users prosado documents owid repos etl venv lib site packages mkdocs config base py line in load config errors warnings cfg validate file users prosado documents owid repos etl venv lib site packages mkdocs config base py line in validate run failed run warnings self validate file users prosado documents owid repos etl venv lib site packages mkdocs config base py line in validate self config option validate value file users prosado documents owid repos etl venv lib site packages mkdocs config config options py line in validate return self run validation value file users prosado documents owid repos etl venv lib site packages mkdocs config config options py line in run validation self load plugin with namespace name cfg file users prosado documents owid repos etl venv lib site packages mkdocs config config options py line in load plugin with namespace return name self load plugin name config file users prosado documents owid repos etl venv lib site packages mkdocs config config options py line in load plugin plugin cls self installed plugins load file opt homebrew cellar python frameworks python framework versions lib importlib metadata init py line in load module import module match group module file opt homebrew cellar python frameworks python framework versions lib importlib init py line in import module return bootstrap gcd import name package level file line in gcd import file line in find and load file line in find and load unlocked file line in load unlocked file line in exec module file line in call with frames removed file users prosado documents owid repos etl venv lib site packages mkdocs jupyter plugin py line in from import convert file users prosado documents owid repos etl venv lib site packages mkdocs jupyter convert py line in from import file users prosado documents owid repos etl venv lib site packages mkdocs jupyter py line in from nbconvert nbconvertapp import nbconvertapp file users prosado documents owid repos etl venv lib site packages nbconvert nbconvertapp py line in class nbconvertapp jupyterapp file users prosado documents owid repos etl venv lib site packages nbconvert nbconvertapp py line in nbconvertapp formats get export names file users prosado documents owid repos etl venv lib site packages nbconvert exporters base py line in get export names e get exporter exporter name config config file users prosado documents owid repos etl venv lib site packages nbconvert exporters base py line in get exporter exporter entrypoints get single nbconvert exporters name load file users prosado documents owid repos etl venv lib site packages entrypoints py line in load mod import module self module name file opt homebrew cellar python frameworks python framework versions lib importlib init py line in import module return bootstrap gcd import name package level file users prosado documents owid repos etl venv lib site packages jupyter contrib nbextensions init py line in import jupyter nbextensions configurator file users prosado documents owid repos etl venv lib site packages jupyter nbextensions configurator init py line in from notebook base handlers import apihandler ipythonhandler modulenotfounderror no module named notebook base the workaround is to delete the file docs api python ipynb
0
24,988
18,019,126,692
IssuesEvent
2021-09-16 17:04:20
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
Platform Infrastructure: Migrate forward proxy to EKS
operations VSP-Initiative infrastructure eks network
## Problem Statement The platform is implementing container orchestration as a way for development teams to host and deploy applications and services in a way that is reliable, scalable, and secure. Before offering container-orchestration-as-a-service to other teams using the platform, the Ops team is using container orchestration to host and deploy the applications and services that it manages and maintains. By doing this, Ops will help to ensure that they are both a developer and consumer of the solutions that it offers to others. Once the main Ops-managed utilities have been migrated to EKS and are live in all environments, the final remaining pieces are to migrate the reverse proxy and the fwdproxy. In order to… Migrate fwdproxy to EKS: lift and shift the current fwdproxy (ha proxy) configuration and move it into a container that running in EKS (rather than on an EC2 instance). There are some details to work through in terms of network routing as we translate that service into Kubernetes. In addition, we want to be sure that we can deploy the forward proxy routinely and reliably using CI/CD. We need to be sure that we have full observability and alerting for the forward proxy so that on-call personnel are notified accordingly. Lastly, we need to document the new setup and configuration for forward proxy running in EKS. In a separate, future initiative, we can decomission the existing forward proxy. *Follow your problem description up with a "How might we... _______" statement re-framing that challenge as an opportunity. Don't hint too much at what the solution might be, you should have enough of a focal point here to guide your ideas, but plenty of freedom to think laterally and innovatively as you experiment and prototype later.* ## Hypothesis or Bet *How will this initiative impact the quality of VFS or VSP teams' work?* *How will this initiative be easy for VFS or VSP teams? Or how will it be easier than what they did before?* ## We will know we're done when... ("Definition of Done") *What requirements does this project need to meet for you to finish this initiative?* ## Known Blockers/Dependencies *List any blockers or dependencies for this work to be completed* ## Projected Launch Date End of Q4 2021 ## Launch Checklist ### Is this service / tool / feature... ### ... tested? - [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it. - [ ] ... and issues discovered in usability testing have been addressed. * _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._ - [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching - [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD ### ... documented? - [ ] New documentation is written pursuant to our [documentation style guide](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/documentation/style-guide) - [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0) * _List the existing product that this initiative fits within, or add a new product to this list._ - [ ] Internal-facing: there's a [Product Outline](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-outline-template.md) checked into [`products/platform/PRODUCT_NAME/`](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/platform/) * _Note: the Product Directory Name should match 1:1 with the List of VSP Products_ - [ ] External-facing: a [VFS-facing README](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-readme-template.md) exists for this product/feature tool - [ ] ... and should be located at `platform/PRODUCT_NAME/README.md` - [ ] External-facing: a [User Guide](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/writing-user-guides.md) exists for this product/feature/tool, and is updated for changes from this initiative - [ ] ... and should be linked from the VFS-facing README for your product - [ ] ... and should be located within `platform/PRODUCT_NAME/`, unless you already have another location for it - [ ] _(if applicable)_... and post to [#vsp-content-ia](https://dsva.slack.com/channels/vsp-content-ia) about whether this should be added to the [Documentation homepage](https://department-of-veterans-affairs.github.io/va.gov-team/) - [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings) ### ... measurable - [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it. * For help, see: [Analytics team](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/analytics) - [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose). ## Required Artifacts ### Documentation * **`PRODUCT_NAME`**: _directory name used for your product documentation_ * **Product Outline**: _link to Product Outline_ * **README**: _link to VFS-facing README for your product_ * **User Guide**: _link to User Guide_ ### Testing * **Usability test**: _link to GitHub issue, or provide reason for skipping_ * **Manual QA**: _link to GitHub issue or documented results_ * **Automated tests**: _link to tests, or "N/A"_ ### Measurement * **Success metrics**: _link to where success metrics are measured, or where they're defined (Product Outline is OK), or provide reason for skipping_ * **Release plan**: _link to Release Plan ticket_
1.0
Platform Infrastructure: Migrate forward proxy to EKS - ## Problem Statement The platform is implementing container orchestration as a way for development teams to host and deploy applications and services in a way that is reliable, scalable, and secure. Before offering container-orchestration-as-a-service to other teams using the platform, the Ops team is using container orchestration to host and deploy the applications and services that it manages and maintains. By doing this, Ops will help to ensure that they are both a developer and consumer of the solutions that it offers to others. Once the main Ops-managed utilities have been migrated to EKS and are live in all environments, the final remaining pieces are to migrate the reverse proxy and the fwdproxy. In order to… Migrate fwdproxy to EKS: lift and shift the current fwdproxy (ha proxy) configuration and move it into a container that running in EKS (rather than on an EC2 instance). There are some details to work through in terms of network routing as we translate that service into Kubernetes. In addition, we want to be sure that we can deploy the forward proxy routinely and reliably using CI/CD. We need to be sure that we have full observability and alerting for the forward proxy so that on-call personnel are notified accordingly. Lastly, we need to document the new setup and configuration for forward proxy running in EKS. In a separate, future initiative, we can decomission the existing forward proxy. *Follow your problem description up with a "How might we... _______" statement re-framing that challenge as an opportunity. Don't hint too much at what the solution might be, you should have enough of a focal point here to guide your ideas, but plenty of freedom to think laterally and innovatively as you experiment and prototype later.* ## Hypothesis or Bet *How will this initiative impact the quality of VFS or VSP teams' work?* *How will this initiative be easy for VFS or VSP teams? Or how will it be easier than what they did before?* ## We will know we're done when... ("Definition of Done") *What requirements does this project need to meet for you to finish this initiative?* ## Known Blockers/Dependencies *List any blockers or dependencies for this work to be completed* ## Projected Launch Date End of Q4 2021 ## Launch Checklist ### Is this service / tool / feature... ### ... tested? - [ ] Usability test (_TODO: link_) has been performed, to validate that new changes enable users to do what was intended and that these changes don't worsen quality elsewhere. If usability test isn't relevant for this change, document the reason for skipping it. - [ ] ... and issues discovered in usability testing have been addressed. * _Note on skipping: metrics that show the impact of before/after can be a substitute for usability testing._ - [ ] End-to-end [manual QA](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/quality-assurance/README.md) or [UAT](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/research/planning/what-is-uat.md) is complete, to validate there are no high-severity issues before launching - [ ] _(if applicable)_ New functionality has thorough, automated tests running in CI/CD ### ... documented? - [ ] New documentation is written pursuant to our [documentation style guide](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/documentation/style-guide) - [ ] Product is included in the [List of VSP Products](https://docs.google.com/spreadsheets/d/1Fn2lD419WE3sTZJtN2Ensrjqaz0jH3WvLaBtn812Wjo/edit#gid=0) * _List the existing product that this initiative fits within, or add a new product to this list._ - [ ] Internal-facing: there's a [Product Outline](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-outline-template.md) checked into [`products/platform/PRODUCT_NAME/`](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/platform/) * _Note: the Product Directory Name should match 1:1 with the List of VSP Products_ - [ ] External-facing: a [VFS-facing README](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/product-readme-template.md) exists for this product/feature tool - [ ] ... and should be located at `platform/PRODUCT_NAME/README.md` - [ ] External-facing: a [User Guide](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/teams/vsp/product-management/writing-user-guides.md) exists for this product/feature/tool, and is updated for changes from this initiative - [ ] ... and should be linked from the VFS-facing README for your product - [ ] ... and should be located within `platform/PRODUCT_NAME/`, unless you already have another location for it - [ ] _(if applicable)_... and post to [#vsp-content-ia](https://dsva.slack.com/channels/vsp-content-ia) about whether this should be added to the [Documentation homepage](https://department-of-veterans-affairs.github.io/va.gov-team/) - [ ] _(if applicable)_ Post to [#vsp-service-design](https://dsva.slack.com/channels/vsp-service-design) for external communication about this change (e.g. VSP Newsletter, customer-facing meetings) ### ... measurable - [ ] _(if applicable)_ This change has clearly-defined success metrics, with instrumentation of those analytics where possible, or a reason documented for skipping it. * For help, see: [Analytics team](https://github.com/department-of-veterans-affairs/va.gov-team/tree/master/platform/analytics) - [ ] This change has an accompanying [VSP Initiative Release Plan](https://github.com/department-of-veterans-affairs/va.gov-team/issues/new/choose). ## Required Artifacts ### Documentation * **`PRODUCT_NAME`**: _directory name used for your product documentation_ * **Product Outline**: _link to Product Outline_ * **README**: _link to VFS-facing README for your product_ * **User Guide**: _link to User Guide_ ### Testing * **Usability test**: _link to GitHub issue, or provide reason for skipping_ * **Manual QA**: _link to GitHub issue or documented results_ * **Automated tests**: _link to tests, or "N/A"_ ### Measurement * **Success metrics**: _link to where success metrics are measured, or where they're defined (Product Outline is OK), or provide reason for skipping_ * **Release plan**: _link to Release Plan ticket_
infrastructure
platform infrastructure migrate forward proxy to eks problem statement the platform is implementing container orchestration as a way for development teams to host and deploy applications and services in a way that is reliable scalable and secure before offering container orchestration as a service to other teams using the platform the ops team is using container orchestration to host and deploy the applications and services that it manages and maintains by doing this ops will help to ensure that they are both a developer and consumer of the solutions that it offers to others once the main ops managed utilities have been migrated to eks and are live in all environments the final remaining pieces are to migrate the reverse proxy and the fwdproxy in order to… migrate fwdproxy to eks lift and shift the current fwdproxy ha proxy configuration and move it into a container that running in eks rather than on an instance there are some details to work through in terms of network routing as we translate that service into kubernetes in addition we want to be sure that we can deploy the forward proxy routinely and reliably using ci cd we need to be sure that we have full observability and alerting for the forward proxy so that on call personnel are notified accordingly lastly we need to document the new setup and configuration for forward proxy running in eks in a separate future initiative we can decomission the existing forward proxy follow your problem description up with a how might we statement re framing that challenge as an opportunity don t hint too much at what the solution might be you should have enough of a focal point here to guide your ideas but plenty of freedom to think laterally and innovatively as you experiment and prototype later hypothesis or bet how will this initiative impact the quality of vfs or vsp teams work how will this initiative be easy for vfs or vsp teams or how will it be easier than what they did before we will know we re done when definition of done what requirements does this project need to meet for you to finish this initiative known blockers dependencies list any blockers or dependencies for this work to be completed projected launch date end of launch checklist is this service tool feature tested usability test todo link has been performed to validate that new changes enable users to do what was intended and that these changes don t worsen quality elsewhere if usability test isn t relevant for this change document the reason for skipping it and issues discovered in usability testing have been addressed note on skipping metrics that show the impact of before after can be a substitute for usability testing end to end or is complete to validate there are no high severity issues before launching if applicable new functionality has thorough automated tests running in ci cd documented new documentation is written pursuant to our product is included in the list the existing product that this initiative fits within or add a new product to this list internal facing there s a checked into note the product directory name should match with the list of vsp products external facing a exists for this product feature tool and should be located at platform product name readme md external facing a exists for this product feature tool and is updated for changes from this initiative and should be linked from the vfs facing readme for your product and should be located within platform product name unless you already have another location for it if applicable and post to about whether this should be added to the if applicable post to for external communication about this change e g vsp newsletter customer facing meetings measurable if applicable this change has clearly defined success metrics with instrumentation of those analytics where possible or a reason documented for skipping it for help see this change has an accompanying required artifacts documentation product name directory name used for your product documentation product outline link to product outline readme link to vfs facing readme for your product user guide link to user guide testing usability test link to github issue or provide reason for skipping manual qa link to github issue or documented results automated tests link to tests or n a measurement success metrics link to where success metrics are measured or where they re defined product outline is ok or provide reason for skipping release plan link to release plan ticket
1
69,527
13,262,106,480
IssuesEvent
2020-08-20 21:06:58
abraunegg/onedrive
https://api.github.com/repos/abraunegg/onedrive
closed
Option to just download a single directory or file
Enhancement Implemented Invalid - Old Code Version
**Is your feature request related to a problem? Please describe.** Some times, it's just useful to download some files (or a directory) from OneDrive to the local disk. No need to sync changes or the file/directory to previously exist on the local drive, just be able to download a list of files. I found about "sync_list" but that requires modifying the config and doing a full synchronization, which is not ideal (IMO). **Describe the solution you'd like** Just using this command: "onedrive --synchronize --single-directory --download-only my-dir" should download the directory "my-dir" (which obviously does exist on my OneDrive) from OneDrive to the local disk on my PC. Currently, if I run this command it complains because the local directory ~/OneDrive/my-dir does not exist: ``` [user@laptop ~]$ onedrive --synchronize --single-directory --download-only my-dir Initializing the Synchronization Engine ... ERROR: The requested local directory does not exist. Please check ~/OneDrive/ for requested path [user@laptop ~]$ ``` If I create an empty directory "~/OneDrive/my-dir" and run the same command with "--verbose" then I get this: ``` [user@laptop ~]$ onedrive --synchronize --single-directory --download-only --verbose my-dir Using Config Dir: /home/gpulido/.config/onedrive No config file found, using application defaults Initializing the OneDrive API ... Opening the item database ... All operations will be performed in: /home/gpulido/OneDrive Application version: v2.4.2 Account Type: personal Default Drive ID: 198b1db90d811d51 Default Root ID: 198B1DB90D811D51!121 Remaining Free Space: 31129946763 Fetching details for OneDrive Root OneDrive Root exists in the database Initializing the Synchronization Engine ... Syncing changes from this selected path: bootdisks Syncing changes from selected OneDrive path ... Getting path details from OneDrive ... Applying changes of Path ID: 198B1DB90D811D51!100387 Processing 234 OneDrive items to ensure consistent local state Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope ... [user@laptop ~]$ ``` The last message gets repeated "ad infinitum" apparently (I waited about 20 minutes and continued to appear, had to stop the sync with Ctrl+C). Thanks!
1.0
Option to just download a single directory or file - **Is your feature request related to a problem? Please describe.** Some times, it's just useful to download some files (or a directory) from OneDrive to the local disk. No need to sync changes or the file/directory to previously exist on the local drive, just be able to download a list of files. I found about "sync_list" but that requires modifying the config and doing a full synchronization, which is not ideal (IMO). **Describe the solution you'd like** Just using this command: "onedrive --synchronize --single-directory --download-only my-dir" should download the directory "my-dir" (which obviously does exist on my OneDrive) from OneDrive to the local disk on my PC. Currently, if I run this command it complains because the local directory ~/OneDrive/my-dir does not exist: ``` [user@laptop ~]$ onedrive --synchronize --single-directory --download-only my-dir Initializing the Synchronization Engine ... ERROR: The requested local directory does not exist. Please check ~/OneDrive/ for requested path [user@laptop ~]$ ``` If I create an empty directory "~/OneDrive/my-dir" and run the same command with "--verbose" then I get this: ``` [user@laptop ~]$ onedrive --synchronize --single-directory --download-only --verbose my-dir Using Config Dir: /home/gpulido/.config/onedrive No config file found, using application defaults Initializing the OneDrive API ... Opening the item database ... All operations will be performed in: /home/gpulido/OneDrive Application version: v2.4.2 Account Type: personal Default Drive ID: 198b1db90d811d51 Default Root ID: 198B1DB90D811D51!121 Remaining Free Space: 31129946763 Fetching details for OneDrive Root OneDrive Root exists in the database Initializing the Synchronization Engine ... Syncing changes from this selected path: bootdisks Syncing changes from selected OneDrive path ... Getting path details from OneDrive ... Applying changes of Path ID: 198B1DB90D811D51!100387 Processing 234 OneDrive items to ensure consistent local state Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope Remote change discarded - not in --single-directory sync scope ... [user@laptop ~]$ ``` The last message gets repeated "ad infinitum" apparently (I waited about 20 minutes and continued to appear, had to stop the sync with Ctrl+C). Thanks!
non_infrastructure
option to just download a single directory or file is your feature request related to a problem please describe some times it s just useful to download some files or a directory from onedrive to the local disk no need to sync changes or the file directory to previously exist on the local drive just be able to download a list of files i found about sync list but that requires modifying the config and doing a full synchronization which is not ideal imo describe the solution you d like just using this command onedrive synchronize single directory download only my dir should download the directory my dir which obviously does exist on my onedrive from onedrive to the local disk on my pc currently if i run this command it complains because the local directory onedrive my dir does not exist onedrive synchronize single directory download only my dir initializing the synchronization engine error the requested local directory does not exist please check onedrive for requested path if i create an empty directory onedrive my dir and run the same command with verbose then i get this onedrive synchronize single directory download only verbose my dir using config dir home gpulido config onedrive no config file found using application defaults initializing the onedrive api opening the item database all operations will be performed in home gpulido onedrive application version account type personal default drive id default root id remaining free space fetching details for onedrive root onedrive root exists in the database initializing the synchronization engine syncing changes from this selected path bootdisks syncing changes from selected onedrive path getting path details from onedrive applying changes of path id processing onedrive items to ensure consistent local state remote change discarded not in single directory sync scope remote change discarded not in single directory sync scope remote change discarded not in single directory sync scope remote change discarded not in single directory sync scope remote change discarded not in single directory sync scope remote change discarded not in single directory sync scope the last message gets repeated ad infinitum apparently i waited about minutes and continued to appear had to stop the sync with ctrl c thanks
0
9,192
7,859,752,309
IssuesEvent
2018-06-21 17:37:32
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Cannot build with latest master.
Area-IDE Area-Infrastructure
Probably related to: https://github.com/dotnet/roslyn/issues/28003 I can now build normal projects. However, test projets fail in VS with: ``` 23>------ Build started: Project: VisualBasicServicesTest, Configuration: Release Any CPU ------ 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: The "TranslateSource" task failed unexpectedly. 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: System.IO.FileNotFoundException: File not found: Formatting\xlf\XmlLiterals.cs.xlf 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: File name: 'Formatting\xlf\XmlLiterals.cs.xlf' 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at XliffTasks.Tasks.XlfTask.LoadXlfDocument(String path, String language, Boolean createIfNonExistent) 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at XliffTasks.Tasks.TranslateSource.ExecuteCore() 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at XliffTasks.Tasks.XlfTask.Execute() 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at Microsoft.Build.BackEnd.TaskExecutionHost.Microsoft.Build.BackEnd.ITaskExecutionHost.Execute() 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at Microsoft.Build.BackEnd.TaskBuilder.<ExecuteInstantiatedTask>d__26.MoveNext() 23>Done building project "VisualBasicServicesTest.vbproj" -- FAILED. ``` Tagging @tmat @jaredpar for ideas.
1.0
Cannot build with latest master. - Probably related to: https://github.com/dotnet/roslyn/issues/28003 I can now build normal projects. However, test projets fail in VS with: ``` 23>------ Build started: Project: VisualBasicServicesTest, Configuration: Release Any CPU ------ 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: The "TranslateSource" task failed unexpectedly. 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: System.IO.FileNotFoundException: File not found: Formatting\xlf\XmlLiterals.cs.xlf 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: File name: 'Formatting\xlf\XmlLiterals.cs.xlf' 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at XliffTasks.Tasks.XlfTask.LoadXlfDocument(String path, String language, Boolean createIfNonExistent) 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at XliffTasks.Tasks.TranslateSource.ExecuteCore() 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at XliffTasks.Tasks.XlfTask.Execute() 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at Microsoft.Build.BackEnd.TaskExecutionHost.Microsoft.Build.BackEnd.ITaskExecutionHost.Execute() 23>C:\Users\cyrus\.nuget\packages\xlifftasks\0.2.0-beta-63004-01\build\XliffTasks.targets(167,5): error MSB4018: at Microsoft.Build.BackEnd.TaskBuilder.<ExecuteInstantiatedTask>d__26.MoveNext() 23>Done building project "VisualBasicServicesTest.vbproj" -- FAILED. ``` Tagging @tmat @jaredpar for ideas.
infrastructure
cannot build with latest master probably related to i can now build normal projects however test projets fail in vs with build started project visualbasicservicestest configuration release any cpu c users cyrus nuget packages xlifftasks beta build xlifftasks targets error the translatesource task failed unexpectedly c users cyrus nuget packages xlifftasks beta build xlifftasks targets error system io filenotfoundexception file not found formatting xlf xmlliterals cs xlf c users cyrus nuget packages xlifftasks beta build xlifftasks targets error file name formatting xlf xmlliterals cs xlf c users cyrus nuget packages xlifftasks beta build xlifftasks targets error at xlifftasks tasks xlftask loadxlfdocument string path string language boolean createifnonexistent c users cyrus nuget packages xlifftasks beta build xlifftasks targets error at xlifftasks tasks translatesource executecore c users cyrus nuget packages xlifftasks beta build xlifftasks targets error at xlifftasks tasks xlftask execute c users cyrus nuget packages xlifftasks beta build xlifftasks targets error at microsoft build backend taskexecutionhost microsoft build backend itaskexecutionhost execute c users cyrus nuget packages xlifftasks beta build xlifftasks targets error at microsoft build backend taskbuilder d movenext done building project visualbasicservicestest vbproj failed tagging tmat jaredpar for ideas
1
327,552
9,977,330,187
IssuesEvent
2019-07-09 16:59:46
MasonMcGill/artisan
https://api.github.com/repos/MasonMcGill/artisan
closed
Concurrent reading and writing of artifacts in python
high priority
Hi Mason, I have trained a network in one jupyter notebook and simultaneously read out activity data using watchdog in another jupyter notebook to plot the data in real-time. For a few iterations this works fine but after 20 or so it breaks raising `ValueError: Unable to set extend dataset (no write intent on file)`. Apparently there is some interference between one kernel opening the files for writing and one kernel opening the files for reading? Full traceback below: ```python --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-46-71419ef1d62b> in <module> 5 λ_dec = 1e-5, 6 n_samples = n_samples, ----> 7 n_iter = 10_000) ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in __new__(cls, *args, **kwargs) 67 return _artifact_from_path(cls, _resolve_path(path)) 68 elif path is None and conf is not None: ---> 69 return _artifact_from_conf(cls, conf) 70 elif path is not None and conf is not None: 71 return _artifact_from_path_and_conf(cls, _resolve_path(path), conf) ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _artifact_from_conf(cls, conf) 331 else: 332 object.__setattr__(artifact, 'path', _new_artifact_path(type(artifact))) --> 333 _build(artifact) 334 return artifact 335 ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _build(artifact) 383 except BaseException as e: 384 write_meta(spec=spec, status='stopped') --> 385 raise e 386 387 ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _build(artifact) 379 n_build_args = artifact.build.__code__.co_argcount 380 build_args = [artifact.conf] if n_build_args > 1 else [] --> 381 artifact.build(*build_args) 382 write_meta(spec=spec, status='done') 383 except BaseException as e: ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/dvs/network_training.py in build(self, conf) 97 # log aggregated activity 98 self.extend('activity', [np.stack([network_states[i].nodes["activity"].detach().cpu() ---> 99 for i in range(len(network_states))], axis=1).mean()]) 100 if t == 0: 101 self.ctome = ctome ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in extend(self, key, val) 199 else: 200 assert path.suffix == '' --> 201 _extend_h5(path.with_suffix('.h5'), val) 202 203 #-- Attribute-style element access -------------------- ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _extend_h5(path, val) 442 else: 443 dset = f['data'] --> 444 dset.resize(dset.len() + len(val), 0) 445 dset[-len(val):] = val 446 dset.flush() ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/h5py/_hl/dataset.py in resize(self, size, axis) 436 437 size = tuple(size) --> 438 self.id.set_extent(size) 439 #h5f.flush(self.id) # THG recommends 440 h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/h5d.pyx in h5py.h5d.DatasetID.set_extent() ValueError: Unable to set extend dataset (no write intent on file) ```
1.0
Concurrent reading and writing of artifacts in python - Hi Mason, I have trained a network in one jupyter notebook and simultaneously read out activity data using watchdog in another jupyter notebook to plot the data in real-time. For a few iterations this works fine but after 20 or so it breaks raising `ValueError: Unable to set extend dataset (no write intent on file)`. Apparently there is some interference between one kernel opening the files for writing and one kernel opening the files for reading? Full traceback below: ```python --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-46-71419ef1d62b> in <module> 5 λ_dec = 1e-5, 6 n_samples = n_samples, ----> 7 n_iter = 10_000) ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in __new__(cls, *args, **kwargs) 67 return _artifact_from_path(cls, _resolve_path(path)) 68 elif path is None and conf is not None: ---> 69 return _artifact_from_conf(cls, conf) 70 elif path is not None and conf is not None: 71 return _artifact_from_path_and_conf(cls, _resolve_path(path), conf) ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _artifact_from_conf(cls, conf) 331 else: 332 object.__setattr__(artifact, 'path', _new_artifact_path(type(artifact))) --> 333 _build(artifact) 334 return artifact 335 ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _build(artifact) 383 except BaseException as e: 384 write_meta(spec=spec, status='stopped') --> 385 raise e 386 387 ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _build(artifact) 379 n_build_args = artifact.build.__code__.co_argcount 380 build_args = [artifact.conf] if n_build_args > 1 else [] --> 381 artifact.build(*build_args) 382 write_meta(spec=spec, status='done') 383 except BaseException as e: ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/dvs/network_training.py in build(self, conf) 97 # log aggregated activity 98 self.extend('activity', [np.stack([network_states[i].nodes["activity"].detach().cpu() ---> 99 for i in range(len(network_states))], axis=1).mean()]) 100 if t == 0: 101 self.ctome = ctome ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in extend(self, key, val) 199 else: 200 assert path.suffix == '' --> 201 _extend_h5(path.with_suffix('.h5'), val) 202 203 #-- Attribute-style element access -------------------- ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/artisan/_artifacts.py in _extend_h5(path, val) 442 else: 443 dset = f['data'] --> 444 dset.resize(dset.len() + len(val), 0) 445 dset[-len(val):] = val 446 dset.flush() ~/miniconda3/envs/dvs-sim/lib/python3.7/site-packages/h5py/_hl/dataset.py in resize(self, size, axis) 436 437 size = tuple(size) --> 438 self.id.set_extent(size) 439 #h5f.flush(self.id) # THG recommends 440 h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/_objects.pyx in h5py._objects.with_phil.wrapper() h5py/h5d.pyx in h5py.h5d.DatasetID.set_extent() ValueError: Unable to set extend dataset (no write intent on file) ```
non_infrastructure
concurrent reading and writing of artifacts in python hi mason i have trained a network in one jupyter notebook and simultaneously read out activity data using watchdog in another jupyter notebook to plot the data in real time for a few iterations this works fine but after or so it breaks raising valueerror unable to set extend dataset no write intent on file apparently there is some interference between one kernel opening the files for writing and one kernel opening the files for reading full traceback below python valueerror traceback most recent call last in λ dec n samples n samples n iter envs dvs sim lib site packages artisan artifacts py in new cls args kwargs return artifact from path cls resolve path path elif path is none and conf is not none return artifact from conf cls conf elif path is not none and conf is not none return artifact from path and conf cls resolve path path conf envs dvs sim lib site packages artisan artifacts py in artifact from conf cls conf else object setattr artifact path new artifact path type artifact build artifact return artifact envs dvs sim lib site packages artisan artifacts py in build artifact except baseexception as e write meta spec spec status stopped raise e envs dvs sim lib site packages artisan artifacts py in build artifact n build args artifact build code co argcount build args if n build args else artifact build build args write meta spec spec status done except baseexception as e envs dvs sim lib site packages dvs network training py in build self conf log aggregated activity self extend activity nodes detach cpu for i in range len network states axis mean if t self ctome ctome envs dvs sim lib site packages artisan artifacts py in extend self key val else assert path suffix extend path with suffix val attribute style element access envs dvs sim lib site packages artisan artifacts py in extend path val else dset f dset resize dset len len val dset val dset flush envs dvs sim lib site packages hl dataset py in resize self size axis size tuple size self id set extent size flush self id thg recommends objects pyx in objects with phil wrapper objects pyx in objects with phil wrapper pyx in datasetid set extent valueerror unable to set extend dataset no write intent on file
0
8,725
7,595,332,773
IssuesEvent
2018-04-27 04:45:07
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Links to model documentation missing for Next Gen Web site
bug interface/infrastructure
Currently there is just a blank page
1.0
Links to model documentation missing for Next Gen Web site - Currently there is just a blank page
infrastructure
links to model documentation missing for next gen web site currently there is just a blank page
1
9,822
8,188,441,294
IssuesEvent
2018-08-30 01:56:22
DanielCaspers/DigitalInspection
https://api.github.com/repos/DanielCaspers/DigitalInspection
closed
Switch to Newtonsoft.Json for all JSON serialization/deserialization
App Infrastructure
*Motivation* [Performance has been benchmarked to be significantly faster](https://www.newtonsoft.com/json/help/html/Performance.htm#Benchmarks). This will result in even faster request/response times when changing pages, getting customers data for CA-Portal and will lead to better resource usage on the server.
1.0
Switch to Newtonsoft.Json for all JSON serialization/deserialization - *Motivation* [Performance has been benchmarked to be significantly faster](https://www.newtonsoft.com/json/help/html/Performance.htm#Benchmarks). This will result in even faster request/response times when changing pages, getting customers data for CA-Portal and will lead to better resource usage on the server.
infrastructure
switch to newtonsoft json for all json serialization deserialization motivation this will result in even faster request response times when changing pages getting customers data for ca portal and will lead to better resource usage on the server
1
52,744
7,782,862,581
IssuesEvent
2018-06-06 08:07:33
famuvie/breedR
https://api.github.com/repos/famuvie/breedR
opened
breedR guide on the interaction between Genotype and Environment
documentation priority:low
Use the [workshop materials](https://famuvie.github.io/breedR/workshop_IBL/) as a base for the guide. But update the relevant aspects of the interface that have changed since, as it became apparent in issue #94.
1.0
breedR guide on the interaction between Genotype and Environment - Use the [workshop materials](https://famuvie.github.io/breedR/workshop_IBL/) as a base for the guide. But update the relevant aspects of the interface that have changed since, as it became apparent in issue #94.
non_infrastructure
breedr guide on the interaction between genotype and environment use the as a base for the guide but update the relevant aspects of the interface that have changed since as it became apparent in issue
0